Home
last modified time | relevance | path

Searched refs:cpus_allowed (Results 1 – 13 of 13) sorted by relevance

/linux/kernel/sched/
H A Dext_idle.c115 static s32 pick_idle_cpu_in_node(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_in_node() argument
121 cpu = cpumask_any_and_distribute(idle_cpumask(node)->smt, cpus_allowed); in pick_idle_cpu_in_node()
129 cpu = cpumask_any_and_distribute(idle_cpumask(node)->cpu, cpus_allowed); in pick_idle_cpu_in_node()
150 static s32 pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_from_online_nodes() argument
181 cpu = pick_idle_cpu_in_node(cpus_allowed, node, flags); in pick_idle_cpu_from_online_nodes()
191 pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_from_online_nodes() argument
200 static s32 scx_pick_idle_cpu(const struct cpumask *cpus_allowed, int node, u64 flags) in scx_pick_idle_cpu() argument
209 cpu = pick_idle_cpu_in_node(cpus_allowed, node, flags); in scx_pick_idle_cpu()
224 return pick_idle_cpu_from_online_nodes(cpus_allowed, node, flags); in scx_pick_idle_cpu()
458 const struct cpumask *cpus_allowed, u64 flags) in scx_select_cpu_dfl() argument
[all …]
H A Dsyscalls.c1139 cpumask_var_t cpus_allowed, new_mask; in __sched_setaffinity() local
1141 if (!alloc_cpumask_var(&cpus_allowed, GFP_KERNEL)) in __sched_setaffinity()
1149 cpuset_cpus_allowed(p, cpus_allowed); in __sched_setaffinity()
1150 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity()
1163 cpuset_cpus_allowed(p, cpus_allowed); in __sched_setaffinity()
1164 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity()
1169 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
1184 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
1193 free_cpumask_var(cpus_allowed); in __sched_setaffinity()
H A Dext_idle.h22 const struct cpumask *cpus_allowed, u64 flags);
/linux/kernel/cgroup/
H A Dcpuset-v1.c157 if (!cpumask_empty(cs->cpus_allowed) && in update_relax_domain_level()
264 while (cpumask_empty(parent->cpus_allowed) || in remove_tasks_in_empty_cpuset()
292 cpumask_copy(cs->cpus_allowed, new_cpus); in cpuset1_hotplug_update_tasks()
302 if (cpus_updated && !cpumask_empty(cs->cpus_allowed)) in cpuset1_hotplug_update_tasks()
307 is_empty = cpumask_empty(cs->cpus_allowed) || in cpuset1_hotplug_update_tasks()
341 return cpumask_subset(p->cpus_allowed, q->cpus_allowed) && in is_cpuset_subset()
377 if (!cpumask_empty(cur->cpus_allowed) && in cpuset1_validate_change()
378 cpumask_empty(trial->cpus_allowed)) in cpuset1_validate_change()
403 return cpumask_intersects(cs1->cpus_allowed, in cpuset1_cpus_excl_conflict()
404 cs2->cpus_allowed); in cpuset1_cpus_excl_conflict()
[all …]
H A Dcpuset.c587 &trial->cpus_allowed, in dup_or_alloc_cpuset()
600 cpumask_copy(trial->cpus_allowed, cs->cpus_allowed); in dup_or_alloc_cpuset()
615 free_cpumask_var(cs->cpus_allowed); in free_cpuset()
625 return cpumask_empty(cs->exclusive_cpus) ? cs->cpus_allowed in user_xcpus()
631 return cpumask_empty(cs->cpus_allowed) && in xcpus_empty()
672 if (xcpus_changed && !cpumask_empty(sibling->cpus_allowed) && in cpus_excl_conflict()
673 cpumask_subset(sibling->cpus_allowed, trial->exclusive_cpus)) in cpus_excl_conflict()
1094 cpumask_and(new_cpus, cs->cpus_allowed, parent->effective_cpus); in compute_effective_cpumask()
1193 if (!cpumask_and(cs->effective_cpus, parent->effective_cpus, cs->cpus_allowed)) in reset_partition_data()
2223 !cpumask_equal(cp->cpus_allowed, cp->effective_cpus)); in update_cpumasks_hier()
[all …]
H A Dcpuset-internal.h101 cpumask_var_t cpus_allowed; member
/linux/tools/sched_ext/include/scx/
H A Dcompat.bpf.h214 #define __COMPAT_scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) \ in __COMPAT_scx_bpf_cpu_curr()
216 scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) : \ in __COMPAT_scx_bpf_cpu_curr()
217 scx_bpf_pick_idle_cpu(cpus_allowed, flags)) in __COMPAT_scx_bpf_cpu_curr()
219 #define __COMPAT_scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) \ in __COMPAT_scx_bpf_cpu_curr()
221 scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) : \ in __COMPAT_scx_bpf_cpu_curr()
222 scx_bpf_pick_any_cpu(cpus_allowed, flags))
254 const struct cpumask *cpus_allowed, u64 flags) __ksym __weak; in scx_bpf_select_cpu_and()
263 * @cpus_allowed: cpumask of allowed CPUs in scx_bpf_select_cpu_and()
271 const struct cpumask *cpus_allowed, u64 flags)
280 return __scx_bpf_select_cpu_and(p, cpus_allowed,
196 __COMPAT_scx_bpf_pick_idle_cpu_node(cpus_allowed,node,flags) global() argument
201 __COMPAT_scx_bpf_pick_any_cpu_node(cpus_allowed,node,flags) global() argument
253 scx_bpf_select_cpu_and(struct task_struct * p,s32 prev_cpu,u64 wake_flags,const struct cpumask * cpus_allowed,u64 flags) scx_bpf_select_cpu_and() argument
[all...]
H A Dcommon.bpf.h63 s32 __scx_bpf_select_cpu_and(struct task_struct *p, const struct cpumask *cpus_allowed,
93 s32 scx_bpf_pick_idle_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
94 s32 scx_bpf_pick_idle_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
95 s32 scx_bpf_pick_any_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
96 s32 scx_bpf_pick_any_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
/linux/arch/mips/kernel/
H A Dmips-mt-fpaff.c69 cpumask_var_t cpus_allowed, new_mask, effective_mask; in mipsmt_sys_sched_setaffinity() local
94 if (!alloc_cpumask_var(&cpus_allowed, GFP_KERNEL)) { in mipsmt_sys_sched_setaffinity()
132 cpuset_cpus_allowed(p, cpus_allowed); in mipsmt_sys_sched_setaffinity()
133 if (!cpumask_subset(effective_mask, cpus_allowed)) { in mipsmt_sys_sched_setaffinity()
139 cpumask_copy(new_mask, cpus_allowed); in mipsmt_sys_sched_setaffinity()
148 free_cpumask_var(cpus_allowed); in mipsmt_sys_sched_setaffinity()
/linux/tools/testing/selftests/ublk/
H A Dtest_generic_12.sh41 --cpus_allowed=0 > /dev/null 2>&1
/linux/arch/powerpc/platforms/cell/spufs/
H A Dspufs.h111 cpumask_t cpus_allowed; member
H A Dsched.c131 cpumask_copy(&ctx->cpus_allowed, current->cpus_ptr); in __spu_update_sched_info()
160 if (cpumask_intersects(mask, &ctx->cpus_allowed)) in __node_allowed()
/linux/Documentation/admin-guide/cgroup-v1/
H A Dcpusets.rst61 schedule a task on a CPU that is not allowed in its cpus_allowed
161 displaying the task's cpus_allowed (on which CPUs it may be scheduled)
497 (makes sure that all the CPUs in the cpus_allowed of that cpuset are
680 their cpus_allowed to allow all online CPUs. When memory hotplug