Home
last modified time | relevance | path

Searched refs:cpus_allowed (Results 1 – 12 of 12) sorted by relevance

/linux/kernel/sched/
H A Dext_idle.c115 static s32 pick_idle_cpu_in_node(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_in_node() argument
121 cpu = cpumask_any_and_distribute(idle_cpumask(node)->smt, cpus_allowed); in pick_idle_cpu_in_node()
129 cpu = cpumask_any_and_distribute(idle_cpumask(node)->cpu, cpus_allowed); in pick_idle_cpu_in_node()
150 static s32 pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_from_online_nodes() argument
181 cpu = pick_idle_cpu_in_node(cpus_allowed, node, flags); in pick_idle_cpu_from_online_nodes()
191 pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags) in pick_idle_cpu_from_online_nodes() argument
200 static s32 scx_pick_idle_cpu(const struct cpumask *cpus_allowed, int node, u64 flags) in scx_pick_idle_cpu() argument
209 cpu = pick_idle_cpu_in_node(cpus_allowed, node, flags); in scx_pick_idle_cpu()
224 return pick_idle_cpu_from_online_nodes(cpus_allowed, node, flags); in scx_pick_idle_cpu()
452 const struct cpumask *cpus_allowed, u64 flags) in scx_select_cpu_dfl() argument
[all …]
H A Dsyscalls.c1131 cpumask_var_t cpus_allowed, new_mask; in __sched_setaffinity() local
1133 if (!alloc_cpumask_var(&cpus_allowed, GFP_KERNEL)) in __sched_setaffinity()
1141 cpuset_cpus_allowed(p, cpus_allowed); in __sched_setaffinity()
1142 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity()
1155 cpuset_cpus_allowed(p, cpus_allowed); in __sched_setaffinity()
1156 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity()
1161 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
1176 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
1185 free_cpumask_var(cpus_allowed); in __sched_setaffinity()
/linux/tools/sched_ext/include/scx/
H A Dcompat.bpf.h196 #define __COMPAT_scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) \ argument
198 scx_bpf_pick_idle_cpu_node(cpus_allowed, node, flags) : \
199 scx_bpf_pick_idle_cpu(cpus_allowed, flags))
201 #define __COMPAT_scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) \ argument
203 scx_bpf_pick_any_cpu_node(cpus_allowed, node, flags) : \
204 scx_bpf_pick_any_cpu(cpus_allowed, flags))
236 const struct cpumask *cpus_allowed, u64 flags) __ksym __weak;
253 const struct cpumask *cpus_allowed, u64 flags) in scx_bpf_select_cpu_and() argument
262 return __scx_bpf_select_cpu_and(p, cpus_allowed, &args); in scx_bpf_select_cpu_and()
265 cpus_allowed, flags); in scx_bpf_select_cpu_and()
H A Dcommon.bpf.h63 s32 __scx_bpf_select_cpu_and(struct task_struct *p, const struct cpumask *cpus_allowed,
93 s32 scx_bpf_pick_idle_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
94 s32 scx_bpf_pick_idle_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
95 s32 scx_bpf_pick_any_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
96 s32 scx_bpf_pick_any_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
/linux/arch/mips/kernel/
H A Dmips-mt-fpaff.c69 cpumask_var_t cpus_allowed, new_mask, effective_mask; in mipsmt_sys_sched_setaffinity() local
94 if (!alloc_cpumask_var(&cpus_allowed, GFP_KERNEL)) { in mipsmt_sys_sched_setaffinity()
132 cpuset_cpus_allowed(p, cpus_allowed); in mipsmt_sys_sched_setaffinity()
133 if (!cpumask_subset(effective_mask, cpus_allowed)) { in mipsmt_sys_sched_setaffinity()
139 cpumask_copy(new_mask, cpus_allowed); in mipsmt_sys_sched_setaffinity()
148 free_cpumask_var(cpus_allowed); in mipsmt_sys_sched_setaffinity()
/linux/kernel/cgroup/
H A Dcpuset-v1.c157 if (!cpumask_empty(cs->cpus_allowed) && in update_relax_domain_level()
264 while (cpumask_empty(parent->cpus_allowed) || in remove_tasks_in_empty_cpuset()
292 cpumask_copy(cs->cpus_allowed, new_cpus); in cpuset1_hotplug_update_tasks()
302 if (cpus_updated && !cpumask_empty(cs->cpus_allowed)) in cpuset1_hotplug_update_tasks()
307 is_empty = cpumask_empty(cs->cpus_allowed) || in cpuset1_hotplug_update_tasks()
341 return cpumask_subset(p->cpus_allowed, q->cpus_allowed) && in is_cpuset_subset()
H A Dcpuset.c544 &trial->cpus_allowed, in dup_or_alloc_cpuset()
557 cpumask_copy(trial->cpus_allowed, cs->cpus_allowed); in dup_or_alloc_cpuset()
572 free_cpumask_var(cs->cpus_allowed); in free_cpuset()
582 return cpumask_empty(cs->exclusive_cpus) ? cs->cpus_allowed in user_xcpus()
588 return cpumask_empty(cs->cpus_allowed) && in xcpus_empty()
630 if (!cpumask_empty(cs1->cpus_allowed) && in cpus_excl_conflict()
631 cpumask_subset(cs1->cpus_allowed, cs2->exclusive_cpus)) in cpus_excl_conflict()
634 if (!cpumask_empty(cs2->cpus_allowed) && in cpus_excl_conflict()
635 cpumask_subset(cs2->cpus_allowed, cs1->exclusive_cpus)) in cpus_excl_conflict()
693 if (!cpumask_empty(cur->cpus_allowed) && in validate_change()
[all …]
H A Dcpuset-internal.h100 cpumask_var_t cpus_allowed; member
/linux/tools/testing/selftests/ublk/
H A Dtest_generic_12.sh42 --cpus_allowed=0 > /dev/null 2>&1
/linux/arch/powerpc/platforms/cell/spufs/
H A Dspufs.h111 cpumask_t cpus_allowed; member
H A Dsched.c131 cpumask_copy(&ctx->cpus_allowed, current->cpus_ptr); in __spu_update_sched_info()
160 if (cpumask_intersects(mask, &ctx->cpus_allowed)) in __node_allowed()
/linux/Documentation/admin-guide/cgroup-v1/
H A Dcpusets.rst61 schedule a task on a CPU that is not allowed in its cpus_allowed
161 displaying the task's cpus_allowed (on which CPUs it may be scheduled)
497 (makes sure that all the CPUs in the cpus_allowed of that cpuset are
680 their cpus_allowed to allow all online CPUs. When memory hotplug