Searched refs:SCHED_CAPACITY_SHIFT (Results 1 – 12 of 12) sorted by relevance
44 …EFINE_PER_CPU_READ_MOSTLY(unsigned long, arch_max_freq_scale) = 1UL << (2 * SCHED_CAPACITY_SHIFT);107 ratio = ref_rate << (2 * SCHED_CAPACITY_SHIFT); in freq_inv_set_max_ratio()149 scale = div64_u64(scale >> SCHED_CAPACITY_SHIFT, in amu_scale_freq_tick()249 freq >>= SCHED_CAPACITY_SHIFT; in arch_freq_get_on_cpu()
150 >> (SCHED_CAPACITY_SHIFT+1); in parse_dt_topology()153 >> (SCHED_CAPACITY_SHIFT-1)) + 1; in parse_dt_topology()
420 div_u64(cap << SCHED_CAPACITY_SHIFT, max_cap)); in arch_set_cpu_capacity()422 div_u64(cap_freq << SCHED_CAPACITY_SHIFT, base_freq)); in arch_set_cpu_capacity()444 if (check_shl_overflow(acnt, 2*SCHED_CAPACITY_SHIFT, &acnt)) in scale_freq_tick()
145 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()147 sa->util_sum += contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()
140 u32 divider = ((LOAD_AVG_MAX - 1024) << SCHED_CAPACITY_SHIFT) - LOAD_AVG_MAX; in update_idle_rq_clock_pelt()
239 #define cap_scale(v, s) ((v)*(s) >> SCHED_CAPACITY_SHIFT)3539 return cap >= p->dl.dl_density >> (BW_SHIFT - SCHED_CAPACITY_SHIFT); in dl_task_fits_capacity()
157 return dl_bw_cpus(i) << SCHED_CAPACITY_SHIFT; in dl_bw_capacity()
4648 running_sum = se->avg.util_sum >> SCHED_CAPACITY_SHIFT; in update_tg_cfs_load()4874 -(long)(removed_runnable * divider) >> SCHED_CAPACITY_SHIFT); in update_cfs_rq_load_avg()
9654 req.util = req.percent << SCHED_CAPACITY_SHIFT; in capacity_from_percent()
158 scale = (cur_freq << SCHED_CAPACITY_SHIFT) / max_freq; in topology_set_freq_scale()261 capacity = div64_u64(capacity << SCHED_CAPACITY_SHIFT, in topology_normalize_cpu_scale()365 capacity = div64_u64(capacity << SCHED_CAPACITY_SHIFT, in topology_init_cpu_capacity_cppc()
80 cur_freq <<= SCHED_CAPACITY_SHIFT; in virt_scale_freq_tick()
457 # define SCHED_CAPACITY_SHIFT SCHED_FIXEDPOINT_SHIFT macro 458 # define SCHED_CAPACITY_SCALE (1L << SCHED_CAPACITY_SHIFT)