Home
last modified time | relevance | path

Searched refs:task_rq (Results 1 – 8 of 8) sorted by relevance

/linux/kernel/sched/
H A Dext.c691 if (WARN_ON_ONCE(rq != task_rq(p))) in nldsq_cursor_lost_task()
1684 lockdep_assert_rq_held(task_rq(p)); in dispatch_dequeue_locked()
1777 struct rq *rq = task_rq(p); in direct_dispatch()
2366 !WARN_ON_ONCE(src_rq != task_rq(p)); in unlink_dsq_and_lock_src_rq()
2406 struct rq *src_rq = task_rq(p), *dst_rq; in move_task_between_dsqs()
2471 struct rq *task_rq = task_rq(p); in consume_dispatch_q() local
2484 if (rq == task_rq) { in consume_dispatch_q()
2492 if (likely(consume_remote_task(rq, p, enq_flags, dsq, task_rq))) in consume_dispatch_q()
2528 struct rq *src_rq = task_rq(p); in dispatch_to_local_dsq()
2576 !WARN_ON_ONCE(src_rq != task_rq(p))) { in dispatch_to_local_dsq()
[all …]
H A Dcore.c280 if (prio_less(b, a, !!task_rq(a)->core->core_forceidle_count)) in __sched_core_less()
724 rq = task_rq(p); in ___task_rq_lock()
726 if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) { in ___task_rq_lock()
746 rq = task_rq(p); in _task_rq_lock()
765 if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) { in _task_rq_lock()
1523 p->sched_class->reweight_task(task_rq(p), p, &lw); in set_load_weight()
2317 rq = task_rq(p); in wait_task_inactive()
2612 if (task_rq(p) == rq) { in migration_cpu_stop()
2693 if (task_rq(p) != rq) in push_cpu_stop()
2712 if (task_rq(p) == rq) { in push_cpu_stop()
[all …]
H A Dext_internal.h1398 lockdep_is_held(__rq_lockp(task_rq(p)))); in scx_task_sched()
1465 lockdep_is_held(__rq_lockp(task_rq(p)))); in scx_task_sched()
H A Dsyscalls.c1130 if (!cpumask_subset(task_rq(p)->rd->span, mask)) in dl_task_check_affinity()
1421 p_rq = task_rq(p); in yield_to()
1430 if (task_rq(p) != p_rq) in yield_to()
H A Ddeadline.c71 rq = task_rq(dl_task_of(dl_se)); in rq_of_dl_se()
343 dl_rq_change_utilization(task_rq(p), &p->dl, new_bw); in dl_change_utilization()
2464 rq = task_rq(p); in migrate_task_rq_dl()
2734 if (!cpudl_find(&task_rq(task)->rd->cpudl, task, later_mask)) in find_later_rq()
2887 (task_rq(task) != rq || in find_lock_later_rq()
3115 rq = task_rq(p); in set_cpus_allowed_dl()
3634 struct rq *rq = task_rq(p); in __getparam_dl()
H A Dsched.h1401 #define task_rq(p) cpu_rq(task_cpu(p)) macro
1714 return &task_rq(p)->cfs; in task_cfs_rq()
1720 struct rq *rq = task_rq(p); in cfs_rq_of()
3033 WARN_ON_ONCE(task_rq(p) != rq); in attach_task()
H A Dext_idle.c945 if (task_rq(p) != scx_locked_rq()) in select_cpu_from_kfunc()
H A Dfair.c1675 (lockdep_is_held(__rq_lockp(task_rq(p))) && !READ_ONCE(p->on_cpu))); in deref_task_numa_group()
7051 WARN_ON_ONCE(task_rq(p) != rq); in hrtick_start_fair()
13619 struct rq *rq = task_rq(a); in cfs_prio_less()
13626 WARN_ON_ONCE(task_rq(b)->core != rq->core); in cfs_prio_less()
13649 cfs_rqa = &task_rq(a)->cfs; in cfs_prio_less()
13650 cfs_rqb = &task_rq(b)->cfs; in cfs_prio_less()
13704 check_update_overutilized_status(task_rq(curr)); in task_tick_fair()