/linux/drivers/base/ |
H A D | arch_topology.c | 78 for_each_cpu(cpu, cpus) { in topology_set_scale_freq_source() 102 for_each_cpu(cpu, cpus) { in topology_clear_scale_freq_source() 153 for_each_cpu(i, cpus) in topology_set_freq_scale() 197 for_each_cpu(cpu, cpus) in topology_update_hw_pressure() 399 for_each_cpu(cpu, policy->related_cpus) { in init_cpu_capacity_callback() 814 for_each_cpu(sibling, topology_core_cpumask(cpu)) in remove_cpu_topology() 816 for_each_cpu(sibling, topology_sibling_cpumask(cpu)) in remove_cpu_topology() 818 for_each_cpu(sibling, topology_cluster_cpumask(cpu)) in remove_cpu_topology() 820 for_each_cpu(sibling, topology_llc_cpumask(cpu)) in remove_cpu_topology()
|
/linux/arch/mips/kernel/ |
H A D | smp.c | 111 for_each_cpu(i, &cpu_sibling_setup_map) { in set_cpu_sibling_map() 127 for_each_cpu(i, &cpu_core_setup_map) { in set_cpu_core_map() 148 for_each_cpu(k, &temp_foreign_map) in calculate_cpu_foreign_map() 199 for_each_cpu(cpu, mask) { in mips_smp_send_ipi_mask() 291 for_each_cpu(cpu, mask) { in mips_smp_ipi_allocate() 327 for_each_cpu(cpu, mask) { in mips_smp_ipi_free() 753 for_each_cpu(cpu, mask) { in tick_broadcast()
|
/linux/drivers/virt/nitro_enclaves/ |
H A D | ne_misc_dev.c | 214 for_each_cpu(cpu, cpu_pool) in ne_setup_cpu_pool() 227 for_each_cpu(cpu, cpu_pool) in ne_setup_cpu_pool() 261 for_each_cpu(cpu_sibling, topology_sibling_cpumask(0)) { in ne_setup_cpu_pool() 277 for_each_cpu(cpu, cpu_pool) { in ne_setup_cpu_pool() 278 for_each_cpu(cpu_sibling, topology_sibling_cpumask(cpu)) { in ne_setup_cpu_pool() 292 for_each_cpu(cpu_sibling, topology_sibling_cpumask(cpu)) in ne_setup_cpu_pool() 317 for_each_cpu(cpu, cpu_pool) { in ne_setup_cpu_pool() 341 for_each_cpu(cpu, cpu_pool) { in ne_setup_cpu_pool() 360 for_each_cpu(cpu, cpu_pool) in ne_setup_cpu_pool() 401 for_each_cpu(cpu, ne_cpu_pool.avail_threads_per_core[i]) { in ne_teardown_cpu_pool() [all …]
|
/linux/arch/powerpc/kernel/ |
H A D | smp.c | 378 for_each_cpu(cpu, mask) in arch_send_call_function_ipi_mask() 575 for_each_cpu(cpu, mask) in tick_broadcast() 743 for_each_cpu(k, srcmask(i)) in or_cpumasks_related() 750 for_each_cpu(k, srcmask(j)) in or_cpumasks_related() 1427 for_each_cpu(i, per_cpu(thread_group_l2_cache_map, cpu)) { in update_mask_by_l2() 1445 for_each_cpu(i, cpu_sibling_mask(cpu)) in update_mask_by_l2() 1459 for_each_cpu(i, *mask) { in update_mask_by_l2() 1492 for_each_cpu(i, mask_fn(cpu)) { in remove_cpu_from_masks() 1499 for_each_cpu(i, cpu_core_mask(cpu)) in remove_cpu_from_masks() 1503 for_each_cpu(i, cpu_coregroup_mask(cpu)) in remove_cpu_from_masks() [all …]
|
/linux/drivers/md/ |
H A D | dm-ps-io-affinity.c | 79 for_each_cpu(cpu, pi->cpumask) { in ioa_add_path() 143 for_each_cpu(cpu, s->path_mask) in ioa_destroy() 220 for_each_cpu(i, cpumask) { in ioa_select_path() 226 for_each_cpu(i, s->path_mask) { in ioa_select_path()
|
/linux/tools/testing/selftests/cpufreq/ |
H A D | cpu.sh | 15 for_each_cpu() function 70 for_each_cpu cpu_should_have_cpufreq_directory
|
/linux/arch/loongarch/kernel/ |
H A D | smp.c | 104 for_each_cpu(i, &cpu_core_setup_map) { in set_cpu_core_map() 118 for_each_cpu(i, &cpu_llc_shared_setup_map) { in set_cpu_llc_shared_map() 130 for_each_cpu(i, &cpu_llc_shared_setup_map) { in clear_cpu_llc_shared_map() 146 for_each_cpu(i, &cpu_sibling_setup_map) { in set_cpu_sibling_map() 158 for_each_cpu(i, &cpu_sibling_setup_map) { in clear_cpu_sibling_map() 181 for_each_cpu(k, &temp_foreign_map) in calculate_cpu_foreign_map() 244 for_each_cpu(i, mask) in loongson_send_ipi_mask()
|
/linux/arch/riscv/kernel/ |
H A D | sys_hwprobe.c | 30 for_each_cpu(cpu, cpus) { in hwprobe_arch_id() 83 for_each_cpu(cpu, cpus) { in hwprobe_isa_ext0() 188 for_each_cpu(cpu, cpus) { in hwprobe_misaligned() 225 for_each_cpu(cpu, cpus) { in hwprobe_vec_misaligned() 433 for_each_cpu(cpu, &cpus) { in hwprobe_get_cpus()
|
/linux/arch/x86/kernel/ |
H A D | smpboot.c | 538 for_each_cpu(i, cpu_sibling_setup_mask) { in set_cpu_sibling_map() 561 for_each_cpu(i, topology_sibling_cpumask(cpu)) in set_cpu_sibling_map() 568 for_each_cpu(i, cpu_sibling_setup_mask) { in set_cpu_sibling_map() 1143 for_each_cpu(sibling, topology_core_cpumask(cpu)) { in remove_siblinginfo() 1152 for_each_cpu(sibling, topology_die_cpumask(cpu)) in remove_siblinginfo() 1155 for_each_cpu(sibling, topology_sibling_cpumask(cpu)) { in remove_siblinginfo() 1161 for_each_cpu(sibling, cpu_llc_shared_mask(cpu)) in remove_siblinginfo() 1163 for_each_cpu(sibling, cpu_l2c_shared_mask(cpu)) in remove_siblinginfo()
|
/linux/kernel/irq/ |
H A D | matrix.c | 143 for_each_cpu(cpu, msk) { in matrix_find_best_cpu() 164 for_each_cpu(cpu, msk) { in matrix_find_best_cpu_managed() 220 for_each_cpu(cpu, msk) { in irq_matrix_reserve_managed() 237 for_each_cpu(cpu, msk) { in irq_matrix_reserve_managed() 261 for_each_cpu(cpu, msk) { in irq_matrix_remove_managed()
|
H A D | ipi.c | 292 for_each_cpu(cpu, dest) { in __ipi_send_mask() 299 for_each_cpu(cpu, dest) in __ipi_send_mask()
|
/linux/block/ |
H A D | blk-mq-cpumap.c | 72 for_each_cpu(cpu, &masks[queue % nr_masks]) in blk_mq_map_queues() 123 for_each_cpu(cpu, mask) in blk_mq_map_hw_queues()
|
/linux/drivers/perf/ |
H A D | arm_pmu_platform.c | 54 for_each_cpu(cpu, &pmu->supported_cpus) in pmu_parse_percpu_irq() 161 for_each_cpu(cpu, &armpmu->supported_cpus) { in armpmu_request_irqs() 179 for_each_cpu(cpu, &armpmu->supported_cpus) { in armpmu_free_irqs()
|
/linux/lib/ |
H A D | cpu_rmap.c | 99 for_each_cpu(neigh, mask) { in cpu_rmap_copy_neigh() 191 for_each_cpu(cpu, affinity) { in cpu_rmap_update() 201 for_each_cpu(cpu, update_mask) { in cpu_rmap_update()
|
/linux/drivers/powercap/ |
H A D | dtpm_cpu.c | 154 for_each_cpu(dtpm_cpu->cpu, policy->related_cpus) in pd_release() 224 for_each_cpu(cpu, policy->related_cpus) in __dtpm_cpu_setup() 250 for_each_cpu(cpu, policy->related_cpus) in __dtpm_cpu_setup()
|
/linux/drivers/opp/ |
H A D | cpu.c | 115 for_each_cpu(cpu, cpumask) { in _dev_pm_opp_cpumask_remove_table() 166 for_each_cpu(cpu, cpumask) { in dev_pm_opp_set_sharing_cpus()
|
/linux/kernel/sched/ |
H A D | topology.c | 220 for_each_cpu(i, cpu_mask) { in sched_is_eas_possible() 422 for_each_cpu(i, cpu_map) { in build_perf_domains() 920 for_each_cpu(i, sg_span) { in build_balance_mask() 1294 for_each_cpu(cpu, mask) { in init_sched_groups_capacity() 1305 for_each_cpu(cpu, sched_group_span(sg)) { in init_sched_groups_capacity() 1368 for_each_cpu(group_cpu, sched_group_span(sg)) { in sched_update_asym_prefer_cpu() 2273 for_each_cpu(j, cpu_map) { in __sdt_alloc() 2324 for_each_cpu(j, cpu_map) { in __sdt_free() 2413 for_each_cpu(cpu, cpu_map) { in topology_span_sane() 2460 for_each_cpu(i, cpu_map) { in build_sched_domains() [all …]
|
/linux/drivers/cpuidle/ |
H A D | coupled.c | 306 for_each_cpu(i, &coupled->coupled_cpus) in cpuidle_coupled_get_state() 352 for_each_cpu(cpu, &coupled->coupled_cpus) in cpuidle_coupled_poke_others() 645 for_each_cpu(cpu, &dev->coupled_cpus) { in cpuidle_coupled_register_device()
|
H A D | cpuidle-tegra.c | 55 for_each_cpu(lcpu, cpu_possible_mask) { in tegra_cpuidle_report_cpus_state() 101 for_each_cpu(lcpu, cpu_online_mask) { in tegra_cpuidle_unpark_secondary_cpus()
|
/linux/tools/testing/selftests/net/bench/page_pool/ |
H A D | time_bench.c | 304 for_each_cpu(cpu, mask) { in time_bench_print_stats_cpumask() 346 for_each_cpu(cpu, mask) { in time_bench_run_concurrent() 385 for_each_cpu(cpu, mask) { in time_bench_run_concurrent()
|
/linux/tools/workqueue/ |
H A D | wq_dump.py | 52 from drgn.helpers.linux.cpumask import for_each_cpu,for_each_possible_cpu 69 for cpu in for_each_cpu(cpumask[0]):
|
/linux/arch/riscv/include/asm/vendor_extensions/ |
H A D | vendor_hwprobe.h | 30 for_each_cpu(cpu, (cpus)) { \
|
/linux/drivers/acpi/ |
H A D | processor_thermal.c | 164 for_each_cpu(cpu, policy->related_cpus) { in acpi_thermal_cpufreq_init() 188 for_each_cpu(cpu, policy->related_cpus) { in acpi_thermal_cpufreq_exit()
|
/linux/arch/arm64/kernel/ |
H A D | watchdog_hld.c | 79 for_each_cpu(cpu, policy->cpus) in watchdog_freq_notifier_callback()
|
/linux/arch/x86/kernel/cpu/microcode/ |
H A D | core.c | 423 for_each_cpu(sibling, secondaries) { in __load_primary() 433 for_each_cpu(cpu, &cpu_offline_mask) { in kick_offline_cpus() 453 for_each_cpu(cpu, &cpu_offline_mask) in release_offline_cpus()
|