Searched refs:src_rq (Results 1 – 4 of 4) sorted by relevance
/linux/kernel/sched/ |
H A D | deadline.c | 2781 struct rq *src_rq; in prio_changed_dl() 2797 src_rq = cpu_rq(cpu); in prio_changed_dl() 2805 src_rq->dl.earliest_dl.next)) in prio_changed_dl() 2810 double_lock_balance(this_rq, src_rq); in prio_changed_dl() 2816 if (src_rq->dl.dl_nr_running <= 1) in task_is_throttled_dl() 2819 p = pick_earliest_pushable_dl_task(src_rq, this_cpu); 2828 WARN_ON(p == src_rq->curr); 2836 src_rq->curr->dl.deadline)) 2840 push_task = get_push_task(src_rq); 2842 deactivate_task(src_rq, 2477 struct rq *src_rq; pull_dl_task() local [all...] |
H A D | rt.c | 2287 struct rq *src_rq; in rto_push_irq_work_func() 2315 src_rq = cpu_rq(cpu); in pull_rt_task() 2318 * Don't bother taking the src_rq->lock if the next highest in pull_rt_task() 2321 * logically higher, the src_rq will push this task away. in pull_rt_task() 2324 if (src_rq->rt.highest_prio.next >= in pull_rt_task() 2334 double_lock_balance(this_rq, src_rq); in pull_rt_task() 2340 p = pick_highest_pushable_task(src_rq, this_cpu); in pull_rt_task() 2347 WARN_ON(p == src_rq->curr); in pull_rt_task() 2358 if (p->prio < src_rq->curr->prio) in pull_rt_task() 2362 push_task = get_push_task(src_rq); in pull_rt_task() 2304 struct rq *src_rq; pull_rt_task() local [all...] |
H A D | core.c | 3297 struct rq *src_rq, *dst_rq; in __migrate_swap_task() local 3300 src_rq = task_rq(p); in __migrate_swap_task() 3303 rq_pin_lock(src_rq, &srf); in __migrate_swap_task() 3306 deactivate_task(src_rq, p, 0); in __migrate_swap_task() 3312 rq_unpin_lock(src_rq, &srf); in __migrate_swap_task() 3332 struct rq *src_rq, *dst_rq; in migrate_swap_stop() local 3337 src_rq = cpu_rq(arg->src_cpu); in migrate_swap_stop() 3341 guard(double_rq_lock)(src_rq, dst_rq); in migrate_swap_stop() 10137 int __sched_mm_cid_migrate_from_fetch_cid(struct rq *src_rq, in __sched_mm_cid_migrate_from_fetch_cid() argument 10166 src_task = rcu_dereference(src_rq->curr); in __sched_mm_cid_migrate_from_fetch_cid() [all …]
|
H A D | fair.c | 9223 struct rq *src_rq; in detach_tasks() 9254 lockdep_assert_rq_held(env->src_rq); 9286 delta = rq_clock_task(env->src_rq) - p->se.exec_start; in attach_tasks() 9317 if (env->src_rq->nr_running > env->src_rq->nr_preferred_running) in others_have_blocked() 9359 lockdep_assert_rq_held(env->src_rq); in __update_blocked_others() 9411 if (task_on_cpu(env->src_rq, p)) { in __update_blocked_fair() 9448 lockdep_assert_rq_held(env->src_rq); in update_cfs_rq_h_load() 9450 deactivate_task(env->src_rq, p, DEQUEUE_NOCLOCK); in update_cfs_rq_h_load() 9455 * detach_one_task() -- tries to dequeue exactly one task from env->src_rq, a in update_cfs_rq_h_load() 8853 struct rq *src_rq; global() member [all...] |