/linux/kernel/irq/ |
H A D | ipi.c | 35 if (!cpumask_subset(dest, cpu_possible_mask)) { in irq_reserve_ipi() 135 if (!ipimask || WARN_ON(!cpumask_subset(dest, ipimask))) in irq_destroy_ipi() 207 if (!cpumask_subset(dest, ipimask)) in ipi_send_verify()
|
H A D | cpuhotplug.c | 199 if (cpumask_subset(irq_data_get_effective_affinity_mask(data), hk_mask)) in hk_should_isolate()
|
/linux/drivers/base/ |
H A D | arch_topology.c | 36 return cpumask_subset(cpus, &scale_freq_counters_mask); in supports_scale_freq_counters() 731 if (cpumask_subset(&cpu_topology[cpu].core_sibling, core_mask)) { in cpu_coregroup_mask() 737 if (cpumask_subset(&cpu_topology[cpu].llc_sibling, core_mask)) in cpu_coregroup_mask() 747 cpumask_subset(core_mask, &cpu_topology[cpu].cluster_sibling)) in cpu_coregroup_mask() 759 if (cpumask_subset(cpu_coregroup_mask(cpu), in cpu_clustergroup_mask()
|
/linux/arch/powerpc/include/asm/ |
H A D | tlb.h | 55 return cpumask_subset(mm_cpumask(mm), in mm_is_core_local()
|
/linux/kernel/cgroup/ |
H A D | cpuset.c | 609 if (!cpumask_empty(acpus) && cpumask_subset(acpus, xcpus)) in cpuset_update_task_spread_flags() 1019 if (!cpumask_subset(cs->effective_cpus, in generate_sched_domains() 1180 return (cpumask_subset(parent->effective_cpus, xcpus) && in generate_sched_domains() 1370 cpumask_subset(top_cpuset.effective_cpus, tmp->new_cpus)) 1402 WARN_ON_ONCE(!cpumask_subset(tmp->new_cpus, subpartitions_cpus)); in update_tasks_cpumask() 1441 WARN_ON_ONCE(!cpumask_subset(cs->effective_xcpus, subpartitions_cpus)); in update_partition_exclusive() 1456 cpumask_subset(top_cpuset.effective_cpus, tmp->addmask))) 1533 if ((prstate != PRS_ISOLATED) && !cpumask_subset(new_cpus, boot_hk_cpus)) in partition_xcpus_newstate() 1737 cpumask_subset(xcpus, parent->effective_xcpus)) { in remote_partition_disable() 1898 if (!cpumask_subset(chil [all...] |
/linux/arch/mips/kernel/ |
H A D | mips-mt-fpaff.c | 133 if (!cpumask_subset(effective_mask, cpus_allowed)) { in mipsmt_sys_sched_setaffinity()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | irq_affinity.c | 100 if (!cpumask_subset(iter_mask, req_mask)) in irq_pool_find_least_loaded()
|
/linux/arch/arm64/kernel/ |
H A D | topology.c | 197 if (unlikely(cpumask_subset(cpus, amu_fie_cpus))) in amu_fie_setup()
|
/linux/kernel/bpf/ |
H A D | cpumask.c | 335 return cpumask_subset(src1, src2); in bpf_cpumask_subset()
|
/linux/kernel/sched/ |
H A D | syscalls.c | 660 if (!cpumask_subset(span, p->cpus_ptr) || in __sched_setscheduler() 1177 if (!cpumask_subset(task_rq(p)->rd->span, mask)) in sched_attr_copy_to_user() 1212 if (!cpumask_subset(new_mask, cpus_allowed)) { in SYSCALL_DEFINE4()
|
H A D | topology.c | 126 !cpumask_subset(groupmask, sched_domain_span(sd->parent))) in sched_domain_debug_one() 1023 !cpumask_subset(sched_domain_span(sibling->child), in find_descended_sibling() 1105 !cpumask_subset(sched_domain_span(sibling->child), span)) in build_overlap_sched_groups() 2338 if (!cpumask_subset(sched_domain_span(child), in build_sched_domain()
|
H A D | deadline.c | 130 if (cpumask_subset(rd->span, cpu_active_mask)) in dl_bw_cpus()
|
H A D | core.c | 3014 if (!kthread && !cpumask_subset(ctx->new_mask, cpu_allowed_mask)) { in __set_cpus_allowed_ptr()
|
/linux/drivers/iommu/ |
H A D | hyperv-iommu.c | 47 if (!cpumask_subset(mask, &ioapic_max_cpumask)) in hyperv_ir_set_affinity()
|
/linux/arch/powerpc/sysdev/xics/ |
H A D | xics-common.c | 285 if (!cpumask_subset(cpu_possible_mask, cpumask)) { in xics_get_irq_server()
|
/linux/kernel/ |
H A D | taskstats.c | 302 if (!cpumask_subset(mask, cpu_possible_mask)) in add_del_listener()
|
H A D | workqueue.c | 5001 if (cpumask_subset(attrs->__pod_cpumask, pt->pod_cpus[pod])) { in get_unbound_pool()
|
/linux/kernel/time/ |
H A D | tick-broadcast.c | 741 if (WARN_ON_ONCE(!cpumask_subset(tmpmask, cpu_online_mask))) in tick_handle_oneshot_broadcast()
|
/linux/include/linux/ |
H A D | cpumask.h | 727 static inline bool cpumask_subset(const struct cpumask *src1p, in cpumask_subset() function
|
/linux/arch/x86/kernel/cpu/resctrl/ |
H A D | pseudo_lock.c | 1529 if (!cpumask_subset(current->cpus_ptr, &plr->d->hdr.cpu_mask)) { in pseudo_lock_dev_mmap()
|
/linux/arch/powerpc/kernel/ |
H A D | smp.c | 1433 !cpumask_subset(submask_fn(cpu), cpu_l2_cache_mask(cpu))) { in update_mask_by_l2()
|
/linux/kernel/rcu/ |
H A D | tree_nocb.h | 1321 if (!cpumask_subset(rcu_nocb_mask, cpu_possible_mask)) { in lazy_rcu_shrink_scan()
|
/linux/io_uring/ |
H A D | io-wq.c | 1338 if (cpumask_subset(mask, allowed_mask)) in io_wq_cpu_affinity()
|
/linux/arch/mips/cavium-octeon/ |
H A D | octeon-irq.c | 2522 if (!cpumask_subset(dest, cpumask_of_node(cd->ciu_node))) in octeon_irq_ciu3_set_affinity()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | sdma.c | 918 if (!cpumask_subset(mask, cpu_online_mask)) { in sdma_set_cpu_to_sde_map()
|