Lines Matching refs:src_cpu

2050 	int src_cpu, src_nid;
2281 if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr))
2494 .src_cpu = task_cpu(p),
2519 sd = rcu_dereference(per_cpu(sd_numa, env.src_cpu));
2602 trace_sched_stick_numa(p, env.src_cpu, NULL, -1);
2611 trace_sched_stick_numa(p, env.src_cpu, NULL, env.best_cpu);
2615 ret = migrate_swap(p, env.best_task, env.best_cpu, env.src_cpu);
2619 trace_sched_stick_numa(p, env.src_cpu, env.best_task, env.best_cpu);
5850 * Ensure that neither of the group entities corresponding to src_cpu or
5855 int src_cpu, int dest_cpu)
5859 src_cfs_rq = tg->cfs_rq[src_cpu];
6775 int src_cpu, int dest_cpu)
9224 int src_cpu;
9309 src_nid = cpu_to_node(env->src_cpu);
9368 if (throttled_lb_pair(task_group(p), env->src_cpu, env->dst_cpu))
9385 * meet load balance goals by pulling other tasks on src_cpu.
9579 if (task_fits_cpu(p, env->src_cpu))
10176 static inline bool sched_asym(struct sched_domain *sd, int dst_cpu, int src_cpu)
10180 * if it has higher priority than @src_cpu.
10183 sched_asym_prefer(dst_cpu, src_cpu);
11532 * If @env::src_cpu is an SMT core with busy siblings, let
11537 (sched_asym_prefer(env->dst_cpu, env->src_cpu) ||
11538 !sched_use_asym_prio(env->sd, env->src_cpu));
11569 * The dst_cpu is idle and the src_cpu CPU has only 1 CFS task.
11570 * It's worth migrating the task if the src_cpu's capacity is reduced
11577 (capacity_of(env->src_cpu)*sd->imbalance_pct < capacity_of(env->dst_cpu)*100))
11707 env.src_cpu = busiest->cpu;
11755 * Revisit (affine) tasks on src_cpu that couldn't be moved to
11758 * iterate on same src_cpu is dependent on number of CPUs in our
11786 * need to continue with same src_cpu.
12024 .src_cpu = busiest_rq->cpu,