/linux/kernel/cgroup/ |
H A D | cpuset.c | 363 if (WARN_ON(!cpumask_and(pmask, possible_mask, cpu_online_mask))) in guarantee_online_cpus() 372 cpumask_and(pmask, pmask, cs->effective_cpus); in guarantee_online_cpus() 774 cpumask_and(doms[0], top_cpuset.effective_cpus, in generate_sched_domains() 893 cpumask_and(doms[i], csa[i]->effective_cpus, in generate_sched_domains() 916 cpumask_and(dp, dp, housekeeping_cpumask(HK_TYPE_DOMAIN)); in generate_sched_domains() 1115 cpumask_and(new_cpus, possible_mask, cs->effective_cpus); in cpuset_update_tasks_cpumask() 1133 cpumask_and(new_cpus, cs->cpus_allowed, parent->effective_cpus); in compute_effective_cpumask() 1233 if (!cpumask_and(cs->effective_cpus, parent->effective_cpus, cs->cpus_allowed)) in reset_partition_data() 1311 cpumask_and(xcpus, xcpus, cpu_active_mask); in partition_xcpus_del() 1357 return cpumask_and(xcpus, user_xcpus(cs), parent->effective_xcpus); in compute_effective_exclusive_cpumask() [all …]
|
/linux/lib/ |
H A D | group_cpus.c | 147 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); in alloc_nodes_groups() 272 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); in __group_cpus_evenly() 297 cpumask_and(nmsk, cpu_mask, node_to_cpumask[nv->id]); in __group_cpus_evenly()
|
/linux/arch/mips/kernel/ |
H A D | mips-mt-fpaff.c | 123 cpumask_and(effective_mask, new_mask, &mt_fpu_cpumask); in mipsmt_sys_sched_setaffinity() 182 cpumask_and(&mask, &allowed, cpu_active_mask); in mipsmt_sys_sched_getaffinity()
|
H A D | pm-cps.c | 124 cpumask_and(coupled_mask, cpu_online_mask, in cps_pm_enter_state()
|
/linux/kernel/sched/ |
H A D | cpupri.c | 103 cpumask_and(lowest_mask, &p->cpus_mask, vec->mask); in __cpupri_find() 104 cpumask_and(lowest_mask, lowest_mask, cpu_active_mask); in __cpupri_find()
|
H A D | syscalls.c | 1167 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity() 1197 bool empty = !cpumask_and(new_mask, new_mask, in __sched_setaffinity() 1310 cpumask_and(mask, &p->cpus_mask, cpu_active_mask); in sched_getaffinity()
|
H A D | cpudeadline.c | 122 cpumask_and(later_mask, cp->free_cpus, &p->cpus_mask)) { in cpudl_find()
|
/linux/arch/riscv/kernel/ |
H A D | unaligned_access_speed.c | 154 cpumask_and(&fast_except_me, &fast_misaligned_access, cpu_online_mask); in set_unaligned_access_static_branches_except_cpu() 173 cpumask_and(&fast_and_online, &fast_misaligned_access, cpu_online_mask); in set_unaligned_access_static_branches()
|
/linux/kernel/trace/ |
H A D | trace_hwlat.c | 327 cpumask_and(current_mask, cpu_online_mask, tr->tracing_cpumask); in move_to_next_cpu() 438 cpumask_and(current_mask, cpu_online_mask, tr->tracing_cpumask); in start_single_kthread() 589 cpumask_and(current_mask, cpu_online_mask, tr->tracing_cpumask); in start_per_cpu_kthreads()
|
/linux/arch/sparc/kernel/ |
H A D | sun4m_smp.c | 195 cpumask_and(&mask, cpu_online_mask, &mask); in sun4m_cross_call()
|
H A D | sun4d_smp.c | 319 cpumask_and(&mask, cpu_online_mask, &mask); in sun4d_cross_call()
|
H A D | leon_smp.c | 407 cpumask_and(&mask, cpu_online_mask, &mask); in leon_cross_call()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | affinity.c | 624 cpumask_and(&entry->def_intr.mask, &node_affinity.real_cpu_mask, in hfi1_dev_affinity_init() 675 cpumask_and(&entry->comp_vect_mask, in hfi1_dev_affinity_init() 1115 cpumask_and(available_mask, hw_thread_mask, node_mask); in hfi1_get_proc_affinity()
|
H A D | netdev_rx.c | 176 cpumask_and(node_cpu_mask, cpu_mask, cpumask_of_node(dd->node)); in hfi1_num_netdev_contexts()
|
/linux/drivers/parisc/ |
H A D | gsc.c | 146 if (!cpumask_and(&tmask, dest, cpu_online_mask)) in gsc_set_affinity_irq()
|
H A D | dino.c | 354 if (!cpumask_and(&tmask, dest, cpu_online_mask)) in dino_set_affinity_irq()
|
/linux/kernel/time/ |
H A D | tick-broadcast.c | 393 cpumask_and(tmpmask, cpu_online_mask, tick_broadcast_mask); in tick_do_periodic_broadcast() 742 cpumask_and(tmpmask, tmpmask, cpu_online_mask); in tick_handle_oneshot_broadcast()
|
/linux/kernel/bpf/ |
H A D | cpumask.c | 255 return cpumask_and((struct cpumask *)dst, src1, src2); in bpf_cpumask_and()
|
/linux/arch/powerpc/kernel/ |
H A D | smp.c | 1451 cpumask_and(*mask, cpu_online_mask, cpu_cpu_mask(cpu)); in update_mask_by_l2() 1541 cpumask_and(*mask, cpu_online_mask, cpu_cpu_mask(cpu)); in update_coregroup_mask() 1604 cpumask_and(mask, mask, cpu_cpu_mask(cpu)); in add_cpu_to_masks()
|
/linux/arch/mips/lantiq/ |
H A D | irq.c | 248 if (!cpumask_and(&tmask, cpumask, cpu_online_mask)) in ltq_icu_irq_set_affinity()
|
/linux/arch/powerpc/perf/ |
H A D | imc-pmu.c | 434 if (cpumask_and(&tmp_mask, l_cpumask, &nest_imc_cpumask)) in ppc_nest_imc_cpu_online() 657 if (cpumask_and(&tmp_mask, l_cpumask, &core_imc_cpumask)) in ppc_core_imc_cpu_online()
|
/linux/kernel/ |
H A D | workqueue.c | 4718 cpumask_and(attrs->cpumask, attrs->cpumask, unbound_cpumask); in wqattrs_actualize_cpumask() 5207 cpumask_and(attrs->__pod_cpumask, pt->pod_cpus[pod], attrs->cpumask); in wq_calc_pod_cpumask() 5306 cpumask_and(new_attrs->cpumask, new_attrs->cpumask, cpu_possible_mask); in apply_wqattrs_prepare() 6623 cpumask_and(&cpumask, pool->attrs->cpumask, cpu_online_mask); in restore_unbound_workers_cpumask() 7281 cpumask_and(cpumask, cpumask, cpu_possible_mask); in workqueue_set_unbound_cpumask() 7705 cpumask_and(wq_unbound_cpumask, wq_unbound_cpumask, mask); in restrict_unbound_cpumask()
|
H A D | kthread.c | 374 cpumask_and(cpumask, pref, housekeeping_cpumask(HK_TYPE_KTHREAD)); in kthread_fetch_affinity()
|
/linux/include/linux/ |
H A D | cpumask.h | 634 bool cpumask_and(struct cpumask *dstp, const struct cpumask *src1p, in cpumask_and() function
|
/linux/drivers/irqchip/ |
H A D | irq-gic-v3-its.c | 1711 cpumask_and(tmpmask, cpumask_of_node(node), aff_mask); in its_select_cpu() 1712 cpumask_and(tmpmask, tmpmask, cpu_online_mask); in its_select_cpu() 1738 cpumask_and(tmpmask, aff_mask, cpu_online_mask); in its_select_cpu() 1751 cpumask_and(tmpmask, tmpmask, cpumask_of_node(node)); in its_select_cpu()
|