Home
last modified time | relevance | path

Searched refs:task_cpu (Results 1 – 25 of 27) sorted by relevance

12

/linux/Documentation/translations/zh_CN/scheduler/
H A Dsched-capacity.rst302 task_util(p) < capacity(task_cpu(p))
358 则任务可能变为CPU受限的,也就是说 ``task_util(p) > capacity(task_cpu(p))`` ;CPU算力
374 task_uclamp_min(p) <= capacity(task_cpu(cpu))
387 task_bandwidth(p) < capacity(task_cpu(p))
/linux/kernel/sched/
H A Dstop_task.c15 return task_cpu(p); /* stop tasks as never migrate */ in select_task_rq_stop()
H A Dcore.c348 int cpu = task_cpu(p); in sched_core_next()
2182 return cpu_curr(task_cpu(p)) == p; in task_curr()
2461 WARN_ON_ONCE(task_cpu(p) != new_cpu); in move_queued_task()
2562 if (cpumask_test_cpu(task_cpu(p), &p->cpus_mask)) in migration_cpu_stop()
2595 if (cpumask_test_cpu(task_cpu(p), p->cpus_ptr)) { in migration_cpu_stop()
2610 stop_one_cpu_nowait(task_cpu(p), migration_cpu_stop, in migration_cpu_stop()
2875 if (cpumask_test_cpu(task_cpu(p), &p->cpus_mask) || in affine_move_task()
3063 !cpumask_test_cpu(task_cpu(p), ctx->new_mask))) { in __set_cpus_allowed_ptr_locked()
3285 if (task_cpu(p) != new_cpu) { in set_task_cpu()
3344 if (task_cpu(arg->dst_task) != arg->dst_cpu) in migrate_swap_stop()
[all …]
H A Ddeadline.c442 struct dl_bw *dl_b = dl_bw_of(task_cpu(p)); in task_non_contending()
447 __dl_sub(dl_b, dl_se->dl_bw, dl_bw_cpus(task_cpu(p))); in task_non_contending()
1984 struct dl_bw *dl_b = dl_bw_of(task_cpu(p)); in inactive_task_timer()
1993 __dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in inactive_task_timer()
2720 int cpu = task_cpu(task); in find_later_rq()
2811 WARN_ON_ONCE(rq->cpu != task_cpu(p)); in pick_next_pushable_dl_task()
3118 __dl_sub(src_dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in set_cpus_allowed_dl()
3544 int cpus, err = -1, cpu = task_cpu(p); in sched_dl_overflow()
H A Dpsi.c900 task->pid, task->comm, task_cpu(task), in psi_flags_change()
911 int cpu = task_cpu(task); in psi_task_change()
930 int cpu = task_cpu(prev); in psi_task_switch()
1008 int cpu = task_cpu(curr); in psi_account_irqtime()
H A Dcpudeadline.c138 (cpu == task_cpu(p) && cap == max_cap)) { in cpudl_find()
H A Didle.c440 return task_cpu(p); /* IDLE tasks as never migrated */ in select_task_rq_idle()
H A Dcpuacct.c338 unsigned int cpu = task_cpu(tsk); in cpuacct_charge()
H A Dfair.c2510 .src_cpu = task_cpu(p), in task_numa_migrate()
3645 int src_nid = cpu_to_node(task_cpu(p)); in update_scan_period()
7176 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_load_without()
7199 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_runnable_without()
7952 if (p && task_cpu(p) == cpu && dst_cpu != cpu) in cpu_util()
7954 else if (p && task_cpu(p) != cpu && dst_cpu == cpu) in cpu_util()
8025 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_util_without()
10671 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in task_running_on_cpu()
13600 set_task_rq(p, task_cpu(p)); in task_change_group_fair()
H A Drt.c1768 int cpu = task_cpu(task); in find_lowest_rq()
1864 BUG_ON(rq->cpu != task_cpu(p)); in pick_next_pushable_task()
H A Dext.c251 return sch->global_dsqs[cpu_to_node(task_cpu(p))]; in find_global_dsq()
1718 WARN_ON_ONCE(task_cpu(p) == cpu); in task_can_run_on_remote_rq()
1735 p->comm, p->pid, task_cpu(p), cpu); in task_can_run_on_remote_rq()
7071 return task_cpu(p); in scx_bpf_task_cpu()
H A Dsched.h1389 #define task_rq(p) cpu_rq(task_cpu(p))
2571 int (*select_task_rq)(struct task_struct *p, int task_cpu, int flags);
/linux/kernel/trace/
H A Dtrace_sched_wakeup.c408 entry->next_cpu = task_cpu(next); in tracing_sched_switch_trace()
434 entry->next_cpu = task_cpu(wakee); in tracing_sched_wakeup_trace()
579 wakeup_cpu = task_cpu(p); in probe_wakeup()
H A Dtrace_osnoise.c1200 long cpu = task_cpu(p); in trace_sched_migrate_callback()
/linux/Documentation/scheduler/
H A Dsched-capacity.rst342 task_util(p) < capacity(task_cpu(p))
405 then it might become CPU-bound, IOW ``task_util(p) > capacity(task_cpu(p))``;
424 task_uclamp_min(p) <= capacity(task_cpu(cpu))
438 task_bandwidth(p) < capacity(task_cpu(p))
/linux/include/linux/
H A Dsched.h2256 static inline unsigned int task_cpu(const struct task_struct *p) in task_cpu() function
2265 static inline unsigned int task_cpu(const struct task_struct *p) in task_cpu() function
2313 return READ_ONCE(owner->on_cpu) && !vcpu_is_preempted(task_cpu(owner)); in owner_on_cpu()
2374 return task_cpu(t); in task_mm_cid()
/linux/include/trace/events/
H A Dsched.h158 __entry->target_cpu = task_cpu(p);
290 __entry->orig_cpu = task_cpu(p);
/linux/kernel/
H A Dstop_machine.c58 struct cpu_stopper *stopper = per_cpu_ptr(&cpu_stopper, task_cpu(task)); in print_stop_info()
/linux/kernel/rcu/
H A Dtasks.h999 cpu = task_cpu(t); in rcu_tasks_is_holdout()
1109 cpu = task_cpu(t); in check_holdout_task()
H A Dtree_nocb.h1520 rdp->nocb_gp_kthread ? (int)task_cpu(rdp->nocb_gp_kthread) : -1, in show_rcu_should_be_on_cpu()
1577 rdp->nocb_cb_kthread ? (int)task_cpu(rdp->nocb_cb_kthread) : -1, in show_rcu_nocb_state()
H A Dtree.c770 cpu = task_cpu(t); in rcu_request_urgent_qs_task()
/linux/arch/parisc/kernel/
H A Dtraps.c154 level, task_cpu(current), cr30, cr31); in show_regs()
/linux/kernel/time/
H A Dtick-sched.c477 cpu = task_cpu(tsk); in tick_nohz_kick_task()
/linux/fs/fuse/
H A Ddev_uring.c1244 qid = task_cpu(current); in fuse_uring_task_to_queue()
/linux/fs/resctrl/
H A Drdtgroup.c613 smp_call_function_single(task_cpu(t), _update_task_closid_rmid, t, 1); in update_task_closid_rmid()
3018 cpumask_set_cpu(task_cpu(t), mask); in rdt_move_group_tasks()

12