| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | affinity.c | 74 cpu = cpumask_first(diff); in cpu_mask_set_get_first() 105 cpumask_first(&node_affinity.real_cpu_mask))); in init_real_cpu_mask() 129 cpumask_first(&node_affinity.proc.mask) in node_affinity_init() 252 ret_cpu = cpumask_first(possible_cpumask); in per_cpu_affinity_get() 288 max_cpu = cpumask_first(possible_cpumask); in per_cpu_affinity_put_max() 340 cpu = cpumask_first(non_intr_cpus); in _dev_comp_vect_cpu_get() 344 cpu = cpumask_first(available_cpus); in _dev_comp_vect_cpu_get() 590 if (cpumask_first(local_mask) >= nr_cpu_ids) in hfi1_dev_affinity_init() 620 curr_cpu = cpumask_first(&entry->def_intr.mask); in hfi1_dev_affinity_init() 781 int cpu = cpumask_first(mask); in hfi1_irq_notifier_notify() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | irq_affinity.c | 40 best_cpu = cpumask_first(cpu_online_mask); in cpu_get_least_loaded() 73 cpu_get(pool, cpumask_first(&af_desc->mask)); in irq_pool_request_irq() 192 cpu = cpumask_first(mlx5_irq_get_affinity_mask(irq)); in mlx5_irq_affinity_irq_release()
|
| /linux/drivers/clocksource/ |
| H A D | dw_apb_timer.c | 108 cpumask_first(evt->cpumask)); in apbt_shutdown() 122 cpumask_first(evt->cpumask)); in apbt_set_oneshot() 155 cpumask_first(evt->cpumask)); in apbt_set_periodic() 179 cpumask_first(evt->cpumask)); in apbt_resume()
|
| H A D | timer-econet-en751221.c | 68 int cpu = cpumask_first(dev->cpumask); in cevt_interrupt() 86 cpu = cpumask_first(dev->cpumask); in cevt_set_next_event()
|
| /linux/arch/sparc/kernel/ |
| H A D | time_32.c | 187 int cpu = cpumask_first(evt->cpumask); in percpu_ce_shutdown() 195 int cpu = cpumask_first(evt->cpumask); in percpu_ce_set_periodic() 204 int cpu = cpumask_first(evt->cpumask); in percpu_ce_set_next_event()
|
| H A D | cpumap.c | 204 prev_cpu = cpu = cpumask_first(cpu_online_mask); in build_cpuinfo_tree() 391 return cpumask_first(cpu_online_mask); in simple_map_to_cpu()
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | amd_cache_disable.c | 158 cpu = cpumask_first(&ci->shared_cpu_map); in store_cache_disable() 190 int cpu = cpumask_first(&ci->shared_cpu_map); in subcaches_show() 200 int cpu = cpumask_first(&ci->shared_cpu_map); in subcaches_store()
|
| /linux/drivers/cpuidle/ |
| H A D | dt_idle_states.c | 109 cpu = cpumask_first(cpumask) + 1; in idle_state_valid() 164 cpu_node = of_cpu_device_node_get(cpumask_first(cpumask)); in dt_init_idle_driver()
|
| /linux/kernel/power/ |
| H A D | poweroff.c | 29 schedule_work_on(cpumask_first(cpu_online_mask), &poweroff_work); in handle_poweroff()
|
| /linux/lib/ |
| H A D | group_cpus.c | 21 cpu = cpumask_first(nmsk); in grp_spread_init_one() 306 cpu = cpumask_first(msk); in alloc_cluster_groups() 333 cpu = cpumask_first(msk); in alloc_cluster_groups()
|
| /linux/kernel/irq/ |
| H A D | ipi.c | 64 offset = cpumask_first(dest); in irq_reserve_ipi() 143 irq = irq + cpumask_first(dest) - data->common->ipi_offset; in irq_destroy_ipi()
|
| /linux/drivers/irqchip/ |
| H A D | irq-mips-gic.c | 138 cpu = cpumask_first(irq_data_get_effective_affinity_mask(d)); in gic_irq_lock_cluster() 292 cpu = cpumask_first(irq_data_get_effective_affinity_mask(d)); in gic_unmask_irq() 381 cpu = cpumask_first(cpumask); in gic_set_affinity() 388 old_cpu = cpumask_first(irq_data_get_effective_affinity_mask(d)); in gic_set_affinity()
|
| H A D | irq-ls-scfg-msi.c | 97 msg->data |= cpumask_first(mask); in ls_scfg_msi_compose_msg() 116 cpu = cpumask_first(mask); in ls_scfg_msi_set_affinity()
|
| /linux/arch/x86/platform/uv/ |
| H A D | uv_time.c | 294 int ced_cpu = cpumask_first(ced->cpumask); in uv_rtc_next_event() 304 int ced_cpu = cpumask_first(evt->cpumask); in uv_rtc_shutdown()
|
| /linux/drivers/pci/controller/ |
| H A D | pci-xgene-msi.c | 136 cpu = cpumask_first(irq_data_get_effective_affinity_mask(data)); in xgene_compose_msi_msg() 152 int target_cpu = cpumask_first(mask); in xgene_msi_set_affinity()
|
| /linux/arch/powerpc/kernel/ |
| H A D | rtasd.c | 439 cpu = cpumask_first(cpu_online_mask); in rtas_event_scan() 492 schedule_delayed_work_on(cpumask_first(cpu_online_mask), in start_event_scan()
|
| H A D | cacheinfo.c | 466 return cpumask_first(per_cpu(thread_group_l1_cache_map, in get_group_id() 469 return cpumask_first(per_cpu(thread_group_l2_cache_map, in get_group_id() 472 return cpumask_first(per_cpu(thread_group_l3_cache_map, in get_group_id()
|
| /linux/arch/mips/kernel/ |
| H A D | sync-r4k.c | 183 smp_call_function_single(cpumask_first(cpu_online_mask), in synchronise_count_slave()
|
| /linux/drivers/thermal/ |
| H A D | cpuidle_cooling.c | 202 dev = get_cpu_device(cpumask_first(drv->cpumask)); in __cpuidle_cooling_register()
|
| /linux/arch/parisc/kernel/ |
| H A D | irq.c | 107 cpu_dest = cpumask_first(cpu_online_mask); in cpu_check_affinity() 518 int cpu = cpumask_first(&dest); in do_cpu_irq_mask()
|
| /linux/drivers/devfreq/ |
| H A D | governor_passive.c | 55 if (parent_cpu_data->first_cpu == cpumask_first(policy->related_cpus)) in get_parent_cpu_data() 335 parent_cpu_data->first_cpu = cpumask_first(policy->related_cpus); in cpufreq_passive_register_notifier()
|
| /linux/Documentation/translations/zh_CN/power/ |
| H A D | energy-model.rst | 198 29 cpu_dev = get_cpu_device(cpumask_first(policy->cpus));
|
| /linux/arch/mips/sgi-ip27/ |
| H A D | ip27-nmi.c | 208 cpu = cpumask_first(cpumask_of_node(node)); in nmi_dump()
|
| /linux/include/linux/ |
| H A D | topology.h | 257 return cpu == cpumask_first(topology_sibling_cpumask(cpu)); in topology_is_primary_thread()
|
| H A D | cpumask.h | 155 static __always_inline unsigned int cpumask_first(const struct cpumask *srcp) in cpumask_first() function 275 return cpumask_first(srcp); in cpumask_any_distribute() 943 #define cpumask_any(srcp) cpumask_first(srcp)
|