Home
last modified time | relevance | path

Searched refs:SCHED_CAPACITY_SCALE (Results 1 – 17 of 17) sorted by relevance

/linux/arch/x86/kernel/cpu/
H A Daperfmperf.c87 static u64 arch_turbo_freq_ratio = SCHED_CAPACITY_SCALE;
88 static u64 arch_max_freq_ratio = SCHED_CAPACITY_SCALE;
92 arch_max_freq_ratio = turbo_disabled ? SCHED_CAPACITY_SCALE : in arch_set_max_freq_ratio()
278 turbo_ratio = div_u64(turbo_freq * SCHED_CAPACITY_SCALE, base_freq); in intel_set_max_freq_ratio()
343 per_cpu(arch_freq_scale, cpu) = SCHED_CAPACITY_SCALE; in disable_freq_invariance_workfn()
349 DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;
383 per_cpu_ptr(arch_cpu_scale, cpu)->capacity = SCHED_CAPACITY_SCALE; in arch_enable_hybrid_capacity_scale()
428 return SCHED_CAPACITY_SCALE; in arch_scale_cpu_capacity()
454 if (freq_scale > SCHED_CAPACITY_SCALE) in scale_freq_tick()
455 freq_scale = SCHED_CAPACITY_SCALE; in scale_freq_tick()
/linux/arch/s390/include/asm/
H A Dtopology.h70 #define CPU_CAPACITY_HIGH SCHED_CAPACITY_SCALE
71 #define CPU_CAPACITY_LOW (SCHED_CAPACITY_SCALE >> 3)
/linux/arch/x86/kernel/acpi/
H A Dcppc.c106 perf_ratio = (div_u64(numerator * SCHED_CAPACITY_SCALE, nominal_perf) + SCHED_CAPACITY_SCALE) >> 1; in amd_set_max_freq_ratio()
/linux/kernel/sched/
H A Dsched.h1064 unsigned long value : bits_per(SCHED_CAPACITY_SCALE);
1065 unsigned long tasks : BITS_PER_LONG - bits_per(SCHED_CAPACITY_SCALE);
2914 return SCHED_CAPACITY_SCALE; in arch_scale_freq_capacity()
3342 return (rq->dl.running_bw * SCHED_CAPACITY_SCALE) >> BW_SHIFT; in cpu_bw_dl()
3396 return max_util != SCHED_CAPACITY_SCALE && rq_util >= max_util; in uclamp_rq_is_capped()
3422 return SCHED_CAPACITY_SCALE; in uclamp_none()
3426 #define UCLAMP_BUCKET_DELTA DIV_ROUND_CLOSEST(SCHED_CAPACITY_SCALE, UCLAMP_BUCKETS)
3449 return SCHED_CAPACITY_SCALE; in uclamp_eff_value()
3465 return SCHED_CAPACITY_SCALE; in uclamp_rq_get()
H A Dtopology.c103 if (group->sgc->capacity != SCHED_CAPACITY_SCALE) in sched_domain_debug_one()
1010 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); in init_overlap_sched_group()
1011 sg->sgc->min_capacity = SCHED_CAPACITY_SCALE; in init_overlap_sched_group()
1012 sg->sgc->max_capacity = SCHED_CAPACITY_SCALE; in init_overlap_sched_group()
1237 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sched_group_span(sg)); in get_group()
1238 sg->sgc->min_capacity = SCHED_CAPACITY_SCALE; in get_group()
1239 sg->sgc->max_capacity = SCHED_CAPACITY_SCALE; in get_group()
H A Dcpufreq_schedutil.c9 #define IOWAIT_BOOST_MIN (SCHED_CAPACITY_SCALE / 8)
272 min_t(unsigned int, sg_cpu->iowait_boost << 1, SCHED_CAPACITY_SCALE); in sugov_iowait_boost()
H A Dsyscalls.c360 if (util_min + 1 > SCHED_CAPACITY_SCALE + 1) in uclamp_validate()
367 if (util_max + 1 > SCHED_CAPACITY_SCALE + 1) in uclamp_validate()
H A Dfair.c4977 #define UTIL_EST_MARGIN (SCHED_CAPACITY_SCALE / 100)
5151 uclamp_max_fits = (capacity_orig == SCHED_CAPACITY_SCALE) && (uclamp_max == SCHED_CAPACITY_SCALE); in util_fits_cpu()
10427 sgs->avg_load = (sgs->group_load * SCHED_CAPACITY_SCALE) / in update_sg_lb_stats()
10716 sgs->avg_load = (sgs->group_load * SCHED_CAPACITY_SCALE) / in update_sg_wakeup_stats()
11004 do_div(tmp, 10000 * SCHED_CAPACITY_SCALE); in update_idle_cpu_scan()
11005 tmp = min_t(long, tmp, SCHED_CAPACITY_SCALE); in update_idle_cpu_scan()
11006 y = SCHED_CAPACITY_SCALE - tmp; in update_idle_cpu_scan()
11010 do_div(y, SCHED_CAPACITY_SCALE); in update_idle_cpu_scan()
11218 local->avg_load = (local->group_load * SCHED_CAPACITY_SCALE) / in calculate_imbalance()
11230 sds->avg_load = (sds->total_load * SCHED_CAPACITY_SCALE) / in calculate_imbalance()
[all …]
H A Dcore.c1446 static unsigned int __maybe_unused sysctl_sched_uclamp_util_min = SCHED_CAPACITY_SCALE;
1449 static unsigned int __maybe_unused sysctl_sched_uclamp_util_max = SCHED_CAPACITY_SCALE;
1466 unsigned int sysctl_sched_uclamp_util_min_rt_default = SCHED_CAPACITY_SCALE;
1917 sysctl_sched_uclamp_util_max > SCHED_CAPACITY_SCALE || in sysctl_sched_uclamp_handler()
1918 sysctl_sched_uclamp_util_min_rt_default > SCHED_CAPACITY_SCALE) { in sysctl_sched_uclamp_handler()
8584 rq->cpu_capacity = SCHED_CAPACITY_SCALE; in sched_init()
9255 .util = SCHED_CAPACITY_SCALE, in capacity_from_percent()
9336 if (util_clamp == SCHED_CAPACITY_SCALE) { in cpu_uclamp_print()
H A Ddeadline.c159 arch_scale_cpu_capacity(i) == SCHED_CAPACITY_SCALE) { in dl_bw_capacity()
207 return SCHED_CAPACITY_SCALE; in dl_bw_capacity()
H A Dext.c20 SCX_CPUPERF_ONE = SCHED_CAPACITY_SCALE,
/linux/include/linux/sched/
H A Dtopology.h269 return SCHED_CAPACITY_SCALE; in arch_scale_cpu_capacity()
/linux/arch/arm64/kernel/
H A Dtopology.c183 scale = min_t(unsigned long, scale, SCHED_CAPACITY_SCALE); in amu_scale_freq_tick()
/linux/init/
H A Dinit_task.c82 .max_allowed_capacity = SCHED_CAPACITY_SCALE,
H A DKconfig851 will be SCHED_CAPACITY_SCALE/UCLAMP_BUCKETS_COUNT. The higher the
/linux/drivers/base/
H A Darch_topology.c130 DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;
156 DEFINE_PER_CPU(unsigned long, cpu_scale) = SCHED_CAPACITY_SCALE;
/linux/include/linux/
H A Dsched.h424 # define SCHED_CAPACITY_SCALE (1L << SCHED_CAPACITY_SHIFT) macro
744 unsigned int value : bits_per(SCHED_CAPACITY_SCALE);