Lines Matching defs:load_avg
1162 sa->load_avg = scale_load_down(se->load.weight);
1164 /* when this task is enqueued, it will contribute to its cfs_rq's load_avg */
1171 * util_avg = cfs_rq->avg.util_avg / (cfs_rq->avg.load_avg + 1)
1220 sa->util_avg /= (cfs_rq->avg.load_avg + 1);
3833 __update_sa(&cfs_rq->avg, load, se->avg.load_avg,
3840 __update_sa(&cfs_rq->avg, load, -se->avg.load_avg,
3887 se->avg.load_avg = div_u64(se_weight(se) * se->avg.load_sum, divider);
3933 * grq->load.weight -> grq->avg.load_avg (2)
3937 * tg->weight * grq->avg.load_avg
3939 * tg->load_avg
3941 * Where: tg->load_avg ~= \Sum grq->avg.load_avg
3948 * one task. It takes time for our CPU's grq->avg.load_avg to build up,
3966 * tg->load_avg - grq->avg.load_avg + grq->load.weight
3969 * we need to use grq->avg.load_avg as its lower bound, which then gives:
3978 * tg_load_avg' = tg->load_avg - grq->avg.load_avg +
3979 * max(grq->load.weight, grq->avg.load_avg)
3996 load = max(scale_load_down(cfs_rq->load.weight), cfs_rq->avg.load_avg);
3998 tg_weight = atomic_long_read(&tg->load_avg);
4089 WARN_ON_ONCE(sa->load_avg ||
4151 * This function 'ensures': tg->load_avg := \Sum tg->cfs_rq[]->avg.load.
4152 * However, because tg->load_avg is a global value there are performance
4159 * Updating tg's load_avg is necessary before update_cfs_share().
4167 * No need to update load_avg for root_task_group as it is not used.
4177 * For migration heavy workloads, access to tg->load_avg can be
4184 delta = cfs_rq->avg.load_avg - cfs_rq->tg_load_avg_contrib;
4186 atomic_long_add(delta, &cfs_rq->tg->load_avg);
4187 cfs_rq->tg_load_avg_contrib = cfs_rq->avg.load_avg;
4198 * No need to update load_avg for root_task_group, as it is not used.
4205 atomic_long_add(delta, &cfs_rq->tg->load_avg);
4282 * ge->avg.load_avg = ge->load.weight * ge->avg.runnable_avg (2)
4287 * grq->avg.load_avg = grq->load.weight * grq->avg.runnable_avg (3)
4295 * ge->load.weight * grq->avg.load_avg
4296 * ge->avg.load_avg = ----------------------------------- (4)
4305 * We specifically want runqueues to have a load_avg that includes
4389 unsigned long load_avg;
4436 load_avg = div_u64(load_sum, divider);
4438 delta_avg = load_avg - se->avg.load_avg;
4445 se->avg.load_avg = load_avg;
4495 if (se->avg.load_avg || se->avg.util_avg)
4639 swap(cfs_rq->removed.load_avg, removed_load);
4706 se->avg.load_sum = se->avg.load_avg * divider;
4830 cfs_rq->removed.load_avg += se->avg.load_avg;
4842 return cfs_rq->avg.load_avg;
7250 load = READ_ONCE(cfs_rq->avg.load_avg);
9835 if (cfs_rq->avg.load_avg)
9969 load = div64_ul(load * se->avg.load_avg,
9982 return div64_ul(p->se.avg.load_avg * cfs_rq->h_load,
10000 return p->se.avg.load_avg;