Home
last modified time | relevance | path

Searched refs:busiest (Results 1 – 2 of 2) sorted by relevance

/linux/kernel/sched/
H A Dfair.c10000 struct sched_group *busiest; /* Busiest group in this sd */ member
10021 .busiest = NULL, in init_sd_lb_stats()
10352 struct sg_lb_stats *busiest, in sibling_imbalance() argument
10358 if (!env->idle || !busiest->sum_nr_running) in sibling_imbalance()
10361 ncores_busiest = sds->busiest->cores; in sibling_imbalance()
10365 imbalance = busiest->sum_nr_running; in sibling_imbalance()
10371 imbalance = ncores_local * busiest->sum_nr_running; in sibling_imbalance()
10379 busiest->sum_nr_running > 1) in sibling_imbalance()
10511 struct sg_lb_stats *busiest = &sds->busiest_stat; in update_sd_pick_busiest() local
10529 if (sgs->group_type > busiest->group_type) in update_sd_pick_busiest()
[all …]
H A Dsched.h3094 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument
3096 __acquires(busiest->lock) in _double_lock_balance()
3100 double_rq_lock(this_rq, busiest); in _double_lock_balance()
3113 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument
3115 __acquires(busiest->lock) in _double_lock_balance()
3118 if (__rq_lockp(this_rq) == __rq_lockp(busiest) || in _double_lock_balance()
3119 likely(raw_spin_rq_trylock(busiest))) { in _double_lock_balance()
3120 double_rq_clock_clear_update(this_rq, busiest); in _double_lock_balance()
3124 if (rq_order_less(this_rq, busiest)) { in _double_lock_balance()
3125 raw_spin_rq_lock_nested(busiest, SINGLE_DEPTH_NESTING); in _double_lock_balance()
[all …]