Symbol: SCHED_CAPACITY_SCALE
arch/arm64/kernel/topology.c
152
scale = min_t(unsigned long, scale, SCHED_CAPACITY_SCALE);
arch/s390/include/asm/topology.h
76
#define CPU_CAPACITY_HIGH SCHED_CAPACITY_SCALE
arch/s390/include/asm/topology.h
77
#define CPU_CAPACITY_LOW (SCHED_CAPACITY_SCALE >> 3)
arch/x86/kernel/acpi/cppc.c
108
perf_ratio = (div_u64(numerator * SCHED_CAPACITY_SCALE, nominal_perf) + SCHED_CAPACITY_SCALE) >> 1;
arch/x86/kernel/cpu/aperfmperf.c
279
turbo_ratio = div_u64(turbo_freq * SCHED_CAPACITY_SCALE, base_freq);
arch/x86/kernel/cpu/aperfmperf.c
348
per_cpu(arch_freq_scale, cpu) = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
354
DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
388
per_cpu_ptr(arch_cpu_scale, cpu)->capacity = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
433
return SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
459
if (freq_scale > SCHED_CAPACITY_SCALE)
arch/x86/kernel/cpu/aperfmperf.c
460
freq_scale = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
88
static u64 arch_turbo_freq_ratio = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
89
static u64 arch_max_freq_ratio = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
93
arch_max_freq_ratio = turbo_disabled ? SCHED_CAPACITY_SCALE :
drivers/base/arch_topology.c
138
DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;
drivers/base/topology.c
212
DEFINE_PER_CPU(unsigned long, cpu_scale) = SCHED_CAPACITY_SCALE;
drivers/cpufreq/cppc_cpufreq.c
392
#define CPPC_EM_COST_GAP (4 * SCHED_CAPACITY_SCALE * CPPC_EM_COST_STEP \
drivers/cpufreq/virtual-cpufreq.c
82
scale = min(scale, SCHED_CAPACITY_SCALE);
include/linux/sched.h
779
unsigned int value : bits_per(SCHED_CAPACITY_SCALE);
include/linux/sched/topology.h
215
return SCHED_CAPACITY_SCALE;
init/init_task.c
112
.max_allowed_capacity = SCHED_CAPACITY_SCALE,
kernel/sched/core.c
1484
static unsigned int __maybe_unused sysctl_sched_uclamp_util_min = SCHED_CAPACITY_SCALE;
kernel/sched/core.c
1487
static unsigned int __maybe_unused sysctl_sched_uclamp_util_max = SCHED_CAPACITY_SCALE;
kernel/sched/core.c
1504
unsigned int sysctl_sched_uclamp_util_min_rt_default = SCHED_CAPACITY_SCALE;
kernel/sched/core.c
1956
sysctl_sched_uclamp_util_max > SCHED_CAPACITY_SCALE ||
kernel/sched/core.c
1957
sysctl_sched_uclamp_util_min_rt_default > SCHED_CAPACITY_SCALE) {
kernel/sched/core.c
8713
rq->cpu_capacity = SCHED_CAPACITY_SCALE;
kernel/sched/core.c
9364
.util = SCHED_CAPACITY_SCALE,
kernel/sched/core.c
9445
if (util_clamp == SCHED_CAPACITY_SCALE) {
kernel/sched/cpufreq_schedutil.c
11
#define IOWAIT_BOOST_MIN (SCHED_CAPACITY_SCALE / 8)
kernel/sched/cpufreq_schedutil.c
300
min_t(unsigned int, sg_cpu->iowait_boost << 1, SCHED_CAPACITY_SCALE);
kernel/sched/deadline.c
153
arch_scale_cpu_capacity(i) == SCHED_CAPACITY_SCALE) {
kernel/sched/ext_internal.h
19
SCX_CPUPERF_ONE = SCHED_CAPACITY_SCALE,
kernel/sched/fair.c
10545
sgs->avg_load = (sgs->group_load * SCHED_CAPACITY_SCALE) /
kernel/sched/fair.c
10835
sgs->avg_load = (sgs->group_load * SCHED_CAPACITY_SCALE) /
kernel/sched/fair.c
11122
do_div(tmp, 10000 * SCHED_CAPACITY_SCALE);
kernel/sched/fair.c
11123
tmp = min_t(long, tmp, SCHED_CAPACITY_SCALE);
kernel/sched/fair.c
11124
y = SCHED_CAPACITY_SCALE - tmp;
kernel/sched/fair.c
11128
do_div(y, SCHED_CAPACITY_SCALE);
kernel/sched/fair.c
11336
local->avg_load = (local->group_load * SCHED_CAPACITY_SCALE) /
kernel/sched/fair.c
11348
sds->avg_load = (sds->total_load * SCHED_CAPACITY_SCALE) /
kernel/sched/fair.c
11374
) / SCHED_CAPACITY_SCALE;
kernel/sched/fair.c
11469
sds.avg_load = (sds.total_load * SCHED_CAPACITY_SCALE) /
kernel/sched/fair.c
4900
#define UTIL_EST_MARGIN (SCHED_CAPACITY_SCALE / 100)
kernel/sched/fair.c
5067
uclamp_max_fits = (capacity_orig == SCHED_CAPACITY_SCALE) && (uclamp_max == SCHED_CAPACITY_SCALE);
kernel/sched/sched.h
1083
unsigned long value : bits_per(SCHED_CAPACITY_SCALE);
kernel/sched/sched.h
1084
unsigned long tasks : BITS_PER_LONG - bits_per(SCHED_CAPACITY_SCALE);
kernel/sched/sched.h
3096
return SCHED_CAPACITY_SCALE;
kernel/sched/sched.h
3517
return (rq->dl.running_bw * SCHED_CAPACITY_SCALE) >> BW_SHIFT;
kernel/sched/sched.h
3592
return max_util != SCHED_CAPACITY_SCALE && rq_util >= max_util;
kernel/sched/sched.h
3605
return SCHED_CAPACITY_SCALE;
kernel/sched/sched.h
3609
#define UCLAMP_BUCKET_DELTA DIV_ROUND_CLOSEST(SCHED_CAPACITY_SCALE, UCLAMP_BUCKETS)
kernel/sched/sched.h
3632
return SCHED_CAPACITY_SCALE;
kernel/sched/sched.h
3650
return SCHED_CAPACITY_SCALE;
kernel/sched/syscalls.c
327
if (util_min + 1 > SCHED_CAPACITY_SCALE + 1)
kernel/sched/syscalls.c
334
if (util_max + 1 > SCHED_CAPACITY_SCALE + 1)
kernel/sched/topology.c
1000
sg->sgc->min_capacity = SCHED_CAPACITY_SCALE;
kernel/sched/topology.c
1001
sg->sgc->max_capacity = SCHED_CAPACITY_SCALE;
kernel/sched/topology.c
111
if (group->sgc->capacity != SCHED_CAPACITY_SCALE)
kernel/sched/topology.c
1226
sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sched_group_span(sg));
kernel/sched/topology.c
1227
sg->sgc->min_capacity = SCHED_CAPACITY_SCALE;
kernel/sched/topology.c
1228
sg->sgc->max_capacity = SCHED_CAPACITY_SCALE;
kernel/sched/topology.c
999
sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span);