| /linux/kernel/sched/ |
| H A D | ext_idle.c | 115 static s32 pick_idle_cpu_in_node(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_in_node() argument 121 cpu = cpumask_any_and_distribute(idle_cpumask(node)->smt, cpus_allowed); in pick_idle_cpu_in_node() 129 cpu = cpumask_any_and_distribute(idle_cpumask(node)->cpu, cpus_allowed); in pick_idle_cpu_in_node() 150 static s32 pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_from_online_nodes() argument 181 cpu = pick_idle_cpu_in_node(cpus_allowed, node, flags); in pick_idle_cpu_from_online_nodes() 191 pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_from_online_nodes() argument 200 static s32 scx_pick_idle_cpu(const struct cpumask *cpus_allowed, int node, u64 flags) in scx_pick_idle_cpu() argument 209 cpu = pick_idle_cpu_in_node(cpus_allowed, node, flags); in scx_pick_idle_cpu() 224 return pick_idle_cpu_from_online_nodes(cpus_allowed, node, flags); in scx_pick_idle_cpu() 458 const struct cpumask *cpus_allowed, u64 flags) in scx_select_cpu_dfl() argument [all …]
|
| H A D | syscalls.c | 1139 cpumask_var_t cpus_allowed, new_mask; in __sched_setaffinity() local 1141 if (!alloc_cpumask_var(&cpus_allowed, GFP_KERNEL)) in __sched_setaffinity() 1149 cpuset_cpus_allowed(p, cpus_allowed); in __sched_setaffinity() 1150 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity() 1163 cpuset_cpus_allowed(p, cpus_allowed); in __sched_setaffinity() 1164 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity() 1169 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity() 1184 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity() 1193 free_cpumask_var(cpus_allowed); in __sched_setaffinity()
|
| H A D | ext_idle.h | 22 const struct cpumask *cpus_allowed, u64 flags);
|
| /linux/kernel/cgroup/ |
| H A D | cpuset-v1.c | 157 if (!cpumask_empty(cs->cpus_allowed) && in update_relax_domain_level() 264 while (cpumask_empty(parent->cpus_allowed) || in remove_tasks_in_empty_cpuset() 292 cpumask_copy(cs->cpus_allowed, new_cpus); in cpuset1_hotplug_update_tasks() 302 if (cpus_updated && !cpumask_empty(cs->cpus_allowed)) in cpuset1_hotplug_update_tasks() 307 is_empty = cpumask_empty(cs->cpus_allowed) || in cpuset1_hotplug_update_tasks() 341 return cpumask_subset(p->cpus_allowed, q->cpus_allowed) && in is_cpuset_subset() 377 if (!cpumask_empty(cur->cpus_allowed) && in cpuset1_validate_change() 378 cpumask_empty(trial->cpus_allowed)) in cpuset1_validate_change() 403 return cpumask_intersects(cs1->cpus_allowed, in cpuset1_cpus_excl_conflict() 404 cs2->cpus_allowed); in cpuset1_cpus_excl_conflict() [all …]
|
| H A D | cpuset.c | 587 &trial->cpus_allowed, in dup_or_alloc_cpuset() 600 cpumask_copy(trial->cpus_allowed, cs->cpus_allowed); in dup_or_alloc_cpuset() 615 free_cpumask_var(cs->cpus_allowed); in free_cpuset() 625 return cpumask_empty(cs->exclusive_cpus) ? cs->cpus_allowed in user_xcpus() 631 return cpumask_empty(cs->cpus_allowed) && in xcpus_empty() 672 if (xcpus_changed && !cpumask_empty(sibling->cpus_allowed) && in cpus_excl_conflict() 673 cpumask_subset(sibling->cpus_allowed, trial->exclusive_cpus)) in cpus_excl_conflict() 1094 cpumask_and(new_cpus, cs->cpus_allowed, parent->effective_cpus); in compute_effective_cpumask() 1193 if (!cpumask_and(cs->effective_cpus, parent->effective_cpus, cs->cpus_allowed)) in reset_partition_data() 2223 !cpumask_equal(cp->cpus_allowed, cp->effective_cpus)); in update_cpumasks_hier() [all …]
|
| H A D | cpuset-internal.h | 101 cpumask_var_t cpus_allowed; member
|
| /linux/tools/sched_ext/include/scx/ |
| H A D | compat.bpf.h | 214 #define __COMPAT_scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) \ in __COMPAT_scx_bpf_cpu_curr() 216 scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) : \ in __COMPAT_scx_bpf_cpu_curr() 217 scx_bpf_pick_idle_cpu(cpus_allowed, flags)) in __COMPAT_scx_bpf_cpu_curr() 219 #define __COMPAT_scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) \ in __COMPAT_scx_bpf_cpu_curr() 221 scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) : \ in __COMPAT_scx_bpf_cpu_curr() 222 scx_bpf_pick_any_cpu(cpus_allowed, flags)) 254 const struct cpumask *cpus_allowed, u64 flags) __ksym __weak; in scx_bpf_select_cpu_and() 263 * @cpus_allowed: cpumask of allowed CPUs in scx_bpf_select_cpu_and() 271 const struct cpumask *cpus_allowed, u64 flags) 280 return __scx_bpf_select_cpu_and(p, cpus_allowed, 196 __COMPAT_scx_bpf_pick_idle_cpu_node(cpus_allowed,node,flags) global() argument 201 __COMPAT_scx_bpf_pick_any_cpu_node(cpus_allowed,node,flags) global() argument 253 scx_bpf_select_cpu_and(struct task_struct * p,s32 prev_cpu,u64 wake_flags,const struct cpumask * cpus_allowed,u64 flags) scx_bpf_select_cpu_and() argument [all...] |
| H A D | common.bpf.h | 63 s32 __scx_bpf_select_cpu_and(struct task_struct *p, const struct cpumask *cpus_allowed, 93 s32 scx_bpf_pick_idle_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak; 94 s32 scx_bpf_pick_idle_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym; 95 s32 scx_bpf_pick_any_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak; 96 s32 scx_bpf_pick_any_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
|
| /linux/arch/mips/kernel/ |
| H A D | mips-mt-fpaff.c | 69 cpumask_var_t cpus_allowed, new_mask, effective_mask; in mipsmt_sys_sched_setaffinity() local 94 if (!alloc_cpumask_var(&cpus_allowed, GFP_KERNEL)) { in mipsmt_sys_sched_setaffinity() 132 cpuset_cpus_allowed(p, cpus_allowed); in mipsmt_sys_sched_setaffinity() 133 if (!cpumask_subset(effective_mask, cpus_allowed)) { in mipsmt_sys_sched_setaffinity() 139 cpumask_copy(new_mask, cpus_allowed); in mipsmt_sys_sched_setaffinity() 148 free_cpumask_var(cpus_allowed); in mipsmt_sys_sched_setaffinity()
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | test_generic_12.sh | 41 --cpus_allowed=0 > /dev/null 2>&1
|
| /linux/arch/powerpc/platforms/cell/spufs/ |
| H A D | spufs.h | 111 cpumask_t cpus_allowed; member
|
| H A D | sched.c | 131 cpumask_copy(&ctx->cpus_allowed, current->cpus_ptr); in __spu_update_sched_info() 160 if (cpumask_intersects(mask, &ctx->cpus_allowed)) in __node_allowed()
|
| /linux/Documentation/admin-guide/cgroup-v1/ |
| H A D | cpusets.rst | 61 schedule a task on a CPU that is not allowed in its cpus_allowed 161 displaying the task's cpus_allowed (on which CPUs it may be scheduled) 497 (makes sure that all the CPUs in the cpus_allowed of that cpuset are 680 their cpus_allowed to allow all online CPUs. When memory hotplug
|