Searched refs:cpu_active_mask (Results 1 – 16 of 16) sorted by relevance
| /linux/kernel/cgroup/ |
| H A D | cpuset.c | 474 if (WARN_ON(!cpumask_and(pmask, possible_mask, cpu_active_mask))) in guarantee_active_cpus() 475 cpumask_copy(pmask, cpu_active_mask); in guarantee_active_cpus() 1004 if (WARN_ON_ONCE(!cpumask_subset(doms[i], cpu_active_mask))) in rebuild_sched_domains_locked() 1172 (!cpumask_intersects(xcpus, cpu_active_mask) && in tasks_nocpu_error() 1268 cpumask_and(parent->effective_cpus, parent->effective_cpus, cpu_active_mask); in partition_xcpus_del() 1302 cpumask_and(full_hk_cpus, full_hk_cpus, cpu_active_mask); in isolated_cpus_can_update() 1508 if (!cpumask_intersects(tmp->new_cpus, cpu_active_mask) || in remote_partition_enable() 1776 cpumask_and(tmp->new_cpus, xcpus, cpu_active_mask); in update_parent_effective_cpumask() 1854 cpumask_and(tmp->new_cpus, tmp->delmask, cpu_active_mask); in update_parent_effective_cpumask() 1863 !cpumask_intersects(tmp->addmask, cpu_active_mask))) { in update_parent_effective_cpumask() [all …]
|
| /linux/arch/mips/kernel/ |
| H A D | mips-mt-fpaff.c | 182 cpumask_and(&mask, &allowed, cpu_active_mask); in mipsmt_sys_sched_getaffinity()
|
| /linux/kernel/sched/ |
| H A D | cpupri.c | 105 cpumask_and(lowest_mask, lowest_mask, cpu_active_mask); in __cpupri_find()
|
| H A D | deadline.c | 132 return cpumask_weight_and(rd->span, cpu_active_mask); in dl_bw_cpus() 140 for_each_cpu_and(i, mask, cpu_active_mask) in __dl_bw_capacity() 182 for_each_cpu_and(i, rd->span, cpu_active_mask) { in __dl_update() 658 cpu = cpumask_any_and(cpu_active_mask, p->cpus_ptr); in dl_task_offline_migration() 671 cpu = cpumask_any(cpu_active_mask); in dl_task_offline_migration() 3171 cpumask_andnot(cpus, cpu_active_mask, hk_msk); in dl_get_task_effective_cpus() 3201 cpu = cpumask_first_and(cpu_active_mask, msk); in dl_add_task_root_domain()
|
| H A D | topology.c | 280 if (!sched_is_eas_possible(cpu_active_mask)) { in sched_energy_aware_handler() 500 if (cpumask_test_cpu(rq->cpu, cpu_active_mask)) in rq_attach_root() 2866 cpumask_and(doms_new[0], cpu_active_mask, in partition_sched_domains_locked() 2890 cpumask_and(doms_new[0], cpu_active_mask, in partition_sched_domains_locked()
|
| H A D | syscalls.c | 1287 cpumask_and(mask, &p->cpus_mask, cpu_active_mask); in sched_getaffinity()
|
| H A D | core.c | 3021 const struct cpumask *cpu_valid_mask = cpu_active_mask; in __set_cpus_allowed_ptr_locked() 8556 sched_init_domains(cpu_active_mask); in sched_init_smp()
|
| H A D | fair.c | 6669 if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask)) in unthrottle_offline_cfs_rqs() 11892 cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask); in sched_balance_rq()
|
| /linux/include/linux/ |
| H A D | cpumask.h | 127 #define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask) macro 1227 #define num_active_cpus() cpumask_weight(cpu_active_mask) 1251 return cpumask_test_cpu(cpu, cpu_active_mask); in cpu_active()
|
| /linux/kernel/ |
| H A D | stop_machine.c | 697 queue_stop_cpus_work(cpu_active_mask, multi_cpu_stop, &msdata, in stop_machine_from_inactive_cpu()
|
| H A D | workqueue.c | 2719 if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask)) in unbind_worker() 5682 if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask)) in init_rescuer()
|
| /linux/arch/powerpc/perf/ |
| H A D | hv-gpci.c | 862 target = cpumask_last(cpu_active_mask); in ppc_hv_gpci_cpu_offline()
|
| H A D | hv-24x7.c | 1670 target = cpumask_last(cpu_active_mask); in ppc_hv_24x7_cpu_offline()
|
| /linux/kernel/power/ |
| H A D | energy_model.c | 498 cpu = cpumask_first_and(em_span_cpus(pd), cpu_active_mask); in em_cpufreq_update_efficiencies()
|
| /linux/kernel/time/ |
| H A D | hrtimer.c | 2304 int i, ncpu = cpumask_any_and(cpu_active_mask, housekeeping_cpumask(HK_TYPE_TIMER)); in hrtimers_cpu_dying()
|
| /linux/arch/arm64/kernel/ |
| H A D | cpufeature.c | 4044 cpu_active_mask); in enable_mismatched_32bit_el0()
|