Home
last modified time | relevance | path

Searched refs:this_rq (Results 1 – 10 of 10) sorted by relevance

/linux/kernel/sched/
H A Dloadavg.c80 long calc_load_fold_active(struct rq *this_rq, long adjust) in calc_load_fold_active() argument
84 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active()
85 nr_active += (long)this_rq->nr_uninterruptible; in calc_load_fold_active()
87 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active()
88 delta = nr_active - this_rq->calc_load_active; in calc_load_fold_active()
89 this_rq->calc_load_active = nr_active; in calc_load_fold_active()
253 calc_load_nohz_fold(this_rq()); in calc_load_nohz_start()
267 struct rq *this_rq = this_rq(); in calc_load_nohz_stop() local
272 this_rq->calc_load_update = READ_ONCE(calc_load_update); in calc_load_nohz_stop()
273 if (time_before(jiffies, this_rq->calc_load_update)) in calc_load_nohz_stop()
[all …]
H A Dsched.h107 extern void calc_global_load_tick(struct rq *this_rq);
108 extern long calc_load_fold_active(struct rq *this_rq, long adjust);
1391 #define this_rq() __this_rq() macro
2004 rq = this_rq(); in _this_rq_lock_irq()
2585 void (*task_woken)(struct rq *this_rq, struct task_struct *task);
2621 void (*switching_from)(struct rq *this_rq, struct task_struct *task);
2622 void (*switched_from) (struct rq *this_rq, struct task_struct *task);
2623 void (*switching_to) (struct rq *this_rq, struct task_struct *task);
2624 void (*switched_to) (struct rq *this_rq, struct task_struct *task);
2625 u64 (*get_prio) (struct rq *this_rq, struct task_struct *task);
[all …]
H A Drt.c583 return this_rq()->rd->span; in sched_rt_period_mask()
2210 rq = this_rq(); in rto_push_irq_work_func()
2240 static void pull_rt_task(struct rq *this_rq) in pull_rt_task() argument
2242 int this_cpu = this_rq->cpu, cpu; in pull_rt_task()
2246 int rt_overload_count = rt_overloaded(this_rq); in pull_rt_task()
2259 cpumask_test_cpu(this_rq->cpu, this_rq->rd->rto_mask)) in pull_rt_task()
2264 tell_cpu_to_push(this_rq); in pull_rt_task()
2269 for_each_cpu(cpu, this_rq->rd->rto_mask) { in pull_rt_task()
2283 this_rq->rt.highest_prio.curr) in pull_rt_task()
2292 double_lock_balance(this_rq, src_rq); in pull_rt_task()
[all …]
H A Dext.c1830 static bool consume_remote_task(struct rq *this_rq, struct task_struct *p,
1833 raw_spin_rq_unlock(this_rq);
1836 move_remote_task_to_local_dsq(p, 0, src_rq, this_rq);
1840 raw_spin_rq_lock(this_rq);
2137 dsq = find_dsq_for_dispatch(sch, this_rq(), dsq_id, p); in flush_dispatch_buf()
5663 static bool kick_one_cpu(s32 cpu, struct rq *this_rq, unsigned long *ksyncs) in kick_one_cpu()
5666 struct scx_rq *this_scx = &this_rq->scx; in kick_one_cpu()
5680 if ((cpu_online(cpu) || cpu == cpu_of(this_rq)) && in kick_one_cpu()
5708 static void kick_one_cpu_if_idle(s32 cpu, struct rq *this_rq) in kick_cpus_irq_workfn()
5716 (cpu_online(cpu) || cpu == cpu_of(this_rq))) in kick_cpus_irq_workfn()
1813 consume_remote_task(struct rq * this_rq,struct task_struct * p,struct scx_dispatch_q * dsq,struct rq * src_rq) consume_remote_task() argument
5642 kick_one_cpu(s32 cpu,struct rq * this_rq,unsigned long * ksyncs) kick_one_cpu() argument
5687 kick_one_cpu_if_idle(s32 cpu,struct rq * this_rq) kick_one_cpu_if_idle() argument
5703 struct rq *this_rq = this_rq(); kick_cpus_irq_workfn() local
6085 struct rq *this_rq, *src_rq, *locked_rq; scx_dsq_move() local
6554 struct rq *this_rq; scx_kick_cpu() local
[all...]
H A Dfair.c4844 static int sched_balance_newidle(struct rq *this_rq, struct rq_flags *rf)
4845 __must_hold(__rq_lockp(this_rq));
6143 if (rq == this_rq()) { in __unthrottle_cfs_rq_async()
7892 this_rq()->nr_running <= 1 && in select_idle_sibling()
8394 struct root_domain *rd = this_rq()->rd; in find_energy_efficient_cpu()
8605 if (!is_rd_overutilized(this_rq()->rd)) { in select_task_rq_fair()
11869 static int sched_balance_rq(int this_cpu, struct rq *this_rq, in sched_balance_rq() argument
11882 .dst_rq = this_rq, in sched_balance_rq()
12631 WARN_ON_ONCE(rq != this_rq()); in nohz_balance_exit_idle()
12737 static void _nohz_idle_balance(struct rq *this_rq, unsigned int flags) in _nohz_idle_balance() argument
[all …]
H A Ddeadline.c2986 static void pull_dl_task(struct rq *this_rq) in pull_dl_task() argument
2988 int this_cpu = this_rq->cpu, cpu; in pull_dl_task()
2994 if (likely(!dl_overloaded(this_rq))) in pull_dl_task()
3003 for_each_cpu(cpu, this_rq->rd->dlo_mask) { in pull_dl_task()
3013 if (this_rq->dl.dl_nr_running && in pull_dl_task()
3014 dl_time_before(this_rq->dl.earliest_dl.curr, in pull_dl_task()
3020 double_lock_balance(this_rq, src_rq); in pull_dl_task()
3037 dl_task_is_earliest_deadline(p, this_rq)) { in pull_dl_task()
3052 move_queued_task_locked(src_rq, this_rq, p); in pull_dl_task()
3060 double_unlock_balance(this_rq, src_rq); in pull_dl_task()
[all …]
H A Didle.c24 idle_set_state(this_rq(), idle_state); in sched_idle_set_state()
H A Dcore.c938 if (rq == this_rq()) in hrtick_start()
2518 struct rq *rq = this_rq(); in migration_cpu_stop()
2629 struct rq *lowest_rq = NULL, *rq = this_rq(); in push_cpu_stop()
3596 rq = this_rq(); in ttwu_stat()
3737 struct rq *rq = this_rq(); in sched_ttwu_pending()
5004 __acquires(__rq_lockp(this_rq())) in prepare_lock_switch()
5022 __acquire(__rq_lockp(this_rq())); in prepare_lock_switch()
5113 __releases(__rq_lockp(this_rq())) in finish_task_switch()
5115 struct rq *rq = this_rq(); in finish_task_switch()
5209 __releases(__rq_lockp(this_rq())) in schedule_tail()
[all …]
H A Dsyscalls.c1411 rq = this_rq(); in yield_to()
/linux/tools/testing/selftests/bpf/progs/
H A Dtest_access_variable_array.c11 int BPF_PROG(fentry_fentry, int this_cpu, struct rq *this_rq, in BPF_PROG() argument