Searched refs:busiest (Results 1 – 2 of 2) sorted by relevance
10000 struct sched_group *busiest; /* Busiest group in this sd */ member10021 .busiest = NULL, in init_sd_lb_stats()10352 struct sg_lb_stats *busiest, in sibling_imbalance() argument10358 if (!env->idle || !busiest->sum_nr_running) in sibling_imbalance()10361 ncores_busiest = sds->busiest->cores; in sibling_imbalance()10365 imbalance = busiest->sum_nr_running; in sibling_imbalance()10371 imbalance = ncores_local * busiest->sum_nr_running; in sibling_imbalance()10379 busiest->sum_nr_running > 1) in sibling_imbalance()10511 struct sg_lb_stats *busiest = &sds->busiest_stat; in update_sd_pick_busiest() local10529 if (sgs->group_type > busiest->group_type) in update_sd_pick_busiest()[all …]
3094 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument3096 __acquires(busiest->lock) in _double_lock_balance()3100 double_rq_lock(this_rq, busiest); in _double_lock_balance()3113 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument3115 __acquires(busiest->lock) in _double_lock_balance()3118 if (__rq_lockp(this_rq) == __rq_lockp(busiest) || in _double_lock_balance()3119 likely(raw_spin_rq_trylock(busiest))) { in _double_lock_balance()3120 double_rq_clock_clear_update(this_rq, busiest); in _double_lock_balance()3124 if (rq_order_less(this_rq, busiest)) { in _double_lock_balance()3125 raw_spin_rq_lock_nested(busiest, SINGLE_DEPTH_NESTING); in _double_lock_balance()[all …]