| /linux/drivers/base/ |
| H A D | arch_topology.c | 39 for_each_cpu(i, cpus) { in supports_scale_freq_counters() 85 for_each_cpu(cpu, cpus) { in topology_set_scale_freq_source() 109 for_each_cpu(cpu, cpus) { in topology_clear_scale_freq_source() 160 for_each_cpu(i, cpus) in topology_set_freq_scale() 204 for_each_cpu(cpu, cpus) in topology_update_hw_pressure() 406 for_each_cpu(cpu, policy->related_cpus) { in init_cpu_capacity_callback() 821 for_each_cpu(sibling, topology_core_cpumask(cpu)) in remove_cpu_topology() 823 for_each_cpu(sibling, topology_sibling_cpumask(cpu)) in remove_cpu_topology() 825 for_each_cpu(sibling, topology_cluster_cpumask(cpu)) in remove_cpu_topology() 827 for_each_cpu(sibling, topology_llc_cpumask(cpu)) in remove_cpu_topology()
|
| /linux/arch/mips/kernel/ |
| H A D | smp.c | 111 for_each_cpu(i, &cpu_sibling_setup_map) { in set_cpu_sibling_map() 127 for_each_cpu(i, &cpu_core_setup_map) { in set_cpu_core_map() 148 for_each_cpu(k, &temp_foreign_map) in calculate_cpu_foreign_map() 199 for_each_cpu(cpu, mask) { in mips_smp_send_ipi_mask() 291 for_each_cpu(cpu, mask) { in mips_smp_ipi_allocate() 327 for_each_cpu(cpu, mask) { in mips_smp_ipi_free() 753 for_each_cpu(cpu, mask) { in tick_broadcast()
|
| /linux/drivers/cpufreq/ |
| H A D | cpufreq_governor.c | 102 for_each_cpu(j, policy_dbs->policy->cpus) { in gov_update_cpu_data() 137 for_each_cpu(j, policy->cpus) { in dbs_update() 332 for_each_cpu(cpu, policy->cpus) { in gov_set_update_util() 344 for_each_cpu(i, policy->cpus) in gov_clear_update_util() 368 for_each_cpu(j, policy->related_cpus) { in alloc_policy_dbs_info() 383 for_each_cpu(j, policy_dbs->policy->related_cpus) { in free_policy_dbs_info() 532 for_each_cpu(j, policy->cpus) { in cpufreq_dbs_governor_start()
|
| /linux/drivers/virt/nitro_enclaves/ |
| H A D | ne_misc_dev.c | 214 for_each_cpu(cpu, cpu_pool) in ne_setup_cpu_pool() 227 for_each_cpu(cpu, cpu_pool) in ne_setup_cpu_pool() 261 for_each_cpu(cpu_sibling, topology_sibling_cpumask(0)) { in ne_setup_cpu_pool() 277 for_each_cpu(cpu, cpu_pool) { in ne_setup_cpu_pool() 278 for_each_cpu(cpu_sibling, topology_sibling_cpumask(cpu)) { in ne_setup_cpu_pool() 292 for_each_cpu(cpu_sibling, topology_sibling_cpumask(cpu)) in ne_setup_cpu_pool() 316 for_each_cpu(cpu, cpu_pool) { in ne_setup_cpu_pool() 340 for_each_cpu(cpu, cpu_pool) { in ne_setup_cpu_pool() 359 for_each_cpu(cpu, cpu_pool) in ne_setup_cpu_pool() 400 for_each_cpu(cpu, ne_cpu_pool.avail_threads_per_core[i]) { in ne_teardown_cpu_pool() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | smp.c | 378 for_each_cpu(cpu, mask) in arch_send_call_function_ipi_mask() 575 for_each_cpu(cpu, mask) in tick_broadcast() 743 for_each_cpu(k, srcmask(i)) in or_cpumasks_related() 750 for_each_cpu(k, srcmask(j)) in or_cpumasks_related() 1447 for_each_cpu(i, per_cpu(thread_group_l2_cache_map, cpu)) { in update_mask_by_l2() 1465 for_each_cpu(i, cpu_sibling_mask(cpu)) in update_mask_by_l2() 1479 for_each_cpu(i, *mask) { in update_mask_by_l2() 1512 for_each_cpu(i, mask_fn(cpu)) { in remove_cpu_from_masks() 1519 for_each_cpu(i, cpu_core_mask(cpu)) in remove_cpu_from_masks() 1523 for_each_cpu(i, cpu_coregroup_mask(cpu)) in remove_cpu_from_masks() [all …]
|
| /linux/drivers/md/ |
| H A D | dm-ps-io-affinity.c | 79 for_each_cpu(cpu, pi->cpumask) { in ioa_add_path() 142 for_each_cpu(cpu, s->path_mask) in ioa_destroy() 219 for_each_cpu(i, cpumask) { in ioa_select_path() 225 for_each_cpu(i, s->path_mask) { in ioa_select_path()
|
| /linux/tools/testing/selftests/cpufreq/ |
| H A D | cpu.sh | 15 for_each_cpu() function 70 for_each_cpu cpu_should_have_cpufreq_directory
|
| /linux/kernel/irq/ |
| H A D | matrix.c | 143 for_each_cpu(cpu, msk) { in matrix_find_best_cpu() 164 for_each_cpu(cpu, msk) { in matrix_find_best_cpu_managed() 220 for_each_cpu(cpu, msk) { in irq_matrix_reserve_managed() 237 for_each_cpu(cpu, msk) { in irq_matrix_reserve_managed() 261 for_each_cpu(cpu, msk) { in irq_matrix_remove_managed()
|
| H A D | ipi.c | 292 for_each_cpu(cpu, dest) { in __ipi_send_mask() 299 for_each_cpu(cpu, dest) in __ipi_send_mask()
|
| /linux/arch/x86/kernel/ |
| H A D | smpboot.c | 541 for_each_cpu(cpu, cpus) { in slit_cluster_package() 694 for_each_cpu(i, cpu_sibling_setup_mask) { in set_cpu_sibling_map() 717 for_each_cpu(i, topology_sibling_cpumask(cpu)) in set_cpu_sibling_map() 724 for_each_cpu(i, cpu_sibling_setup_mask) { in set_cpu_sibling_map() 1299 for_each_cpu(sibling, topology_core_cpumask(cpu)) { in remove_siblinginfo() 1308 for_each_cpu(sibling, topology_die_cpumask(cpu)) in remove_siblinginfo() 1311 for_each_cpu(sibling, topology_sibling_cpumask(cpu)) { in remove_siblinginfo() 1317 for_each_cpu(sibling, cpu_llc_shared_mask(cpu)) in remove_siblinginfo() 1319 for_each_cpu(sibling, cpu_l2c_shared_mask(cpu)) in remove_siblinginfo()
|
| /linux/block/ |
| H A D | blk-mq-cpumap.c | 72 for_each_cpu(cpu, &masks[queue % nr_masks]) in blk_mq_map_queues() 123 for_each_cpu(cpu, mask) in blk_mq_map_hw_queues()
|
| /linux/arch/riscv/kernel/ |
| H A D | sys_hwprobe.c | 47 for_each_cpu(cpu, cpus) { in hwprobe_arch_id() 101 for_each_cpu(cpu, cpus) { in hwprobe_isa_ext0() 198 for_each_cpu(cpu, cpus) { in hwprobe_isa_ext1() 228 for_each_cpu(cpu, cpus) { in hwprobe_misaligned() 265 for_each_cpu(cpu, cpus) { in hwprobe_vec_misaligned() 485 for_each_cpu(cpu, &cpus) { in hwprobe_get_cpus()
|
| /linux/lib/ |
| H A D | cpu_rmap.c | 99 for_each_cpu(neigh, mask) { in cpu_rmap_copy_neigh() 191 for_each_cpu(cpu, affinity) { in cpu_rmap_update() 201 for_each_cpu(cpu, update_mask) { in cpu_rmap_update()
|
| /linux/drivers/powercap/ |
| H A D | dtpm_cpu.c | 154 for_each_cpu(dtpm_cpu->cpu, policy->related_cpus) in pd_release() 224 for_each_cpu(cpu, policy->related_cpus) in __dtpm_cpu_setup() 250 for_each_cpu(cpu, policy->related_cpus) in __dtpm_cpu_setup()
|
| /linux/drivers/opp/ |
| H A D | cpu.c | 115 for_each_cpu(cpu, cpumask) { in _dev_pm_opp_cpumask_remove_table() 167 for_each_cpu(cpu, cpumask) { in dev_pm_opp_set_sharing_cpus()
|
| /linux/include/linux/ |
| H A D | cpumask.h | 379 #define for_each_cpu(cpu, mask) \ macro 475 for_each_cpu(i, mask) in cpumask_any_but() 1139 #define for_each_possible_cpu(cpu) for_each_cpu((cpu), cpu_possible_mask) 1140 #define for_each_online_cpu(cpu) for_each_cpu((cpu), cpu_online_mask) 1141 #define for_each_enabled_cpu(cpu) for_each_cpu((cpu), cpu_enabled_mask) 1142 #define for_each_present_cpu(cpu) for_each_cpu((cpu), cpu_present_mask)
|
| /linux/tools/testing/selftests/net/bench/page_pool/ |
| H A D | time_bench.c | 304 for_each_cpu(cpu, mask) { in time_bench_print_stats_cpumask() 346 for_each_cpu(cpu, mask) { in time_bench_run_concurrent() 385 for_each_cpu(cpu, mask) { in time_bench_run_concurrent()
|
| /linux/drivers/cpuidle/ |
| H A D | coupled.c | 306 for_each_cpu(i, &coupled->coupled_cpus) in cpuidle_coupled_get_state() 352 for_each_cpu(cpu, &coupled->coupled_cpus) in cpuidle_coupled_poke_others() 645 for_each_cpu(cpu, &dev->coupled_cpus) { in cpuidle_coupled_register_device()
|
| /linux/kernel/sched/ |
| H A D | topology.c | 220 for_each_cpu(i, cpu_mask) { in sched_is_eas_possible() 422 for_each_cpu(i, cpu_map) { in build_perf_domains() 925 for_each_cpu(i, sg_span) { in build_balance_mask() 1299 for_each_cpu(cpu, mask) { in init_sched_groups_capacity() 1310 for_each_cpu(cpu, sched_group_span(sg)) { in init_sched_groups_capacity() 1373 for_each_cpu(group_cpu, sched_group_span(sg)) { in sched_update_asym_prefer_cpu() 2387 for_each_cpu(j, cpu_map) { in __sdt_alloc() 2438 for_each_cpu(j, cpu_map) { in __sdt_free() 2527 for_each_cpu(cpu, cpu_map) { in topology_span_sane() 2574 for_each_cpu(i, cpu_map) { in build_sched_domains() [all …]
|
| /linux/drivers/crypto/caam/ |
| H A D | qi.c | 521 for_each_cpu(i, cpus) { in caam_qi_shutdown() 699 for_each_cpu(i, cpus) { in alloc_rsp_fqs() 715 for_each_cpu(i, cpus) in free_rsp_fqs() 724 for_each_cpu(i, cpus) { in free_caam_qi_pcpu_netdev() 762 for_each_cpu(i, cpus) { in caam_qi_init()
|
| /linux/tools/workqueue/ |
| H A D | wq_dump.py | 52 from drgn.helpers.linux.cpumask import for_each_cpu,for_each_possible_cpu 69 for cpu in for_each_cpu(cpumask[0]):
|
| /linux/arch/riscv/include/asm/vendor_extensions/ |
| H A D | vendor_hwprobe.h | 30 for_each_cpu(cpu, (cpus)) { \
|
| /linux/drivers/acpi/ |
| H A D | processor_thermal.c | 166 for_each_cpu(cpu, policy->related_cpus) { in acpi_thermal_cpufreq_init() 190 for_each_cpu(cpu, policy->related_cpus) { in acpi_thermal_cpufreq_exit()
|
| /linux/arch/arm64/kernel/ |
| H A D | watchdog_hld.c | 79 for_each_cpu(cpu, policy->cpus) in watchdog_freq_notifier_callback()
|
| /linux/arch/x86/kernel/apic/ |
| H A D | x2apic_savic.c | 152 for_each_cpu(cpu, cpu_online_mask) { in send_ipi_allbut() 267 for_each_cpu(cpu, mask) { in send_ipi_mask()
|