/linux/drivers/cpuidle/ |
H A D | dt_idle_states.c | 110 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); in idle_state_valid() 111 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid()
|
/linux/lib/ |
H A D | cpumask_kunit.c | 97 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_next(-1, &mask_empty), in test_cpumask_next() 99 KUNIT_EXPECT_EQ_MSG(test, 0, cpumask_next(-1, cpu_possible_mask), in test_cpumask_next()
|
H A D | cpumask.c | 27 next = cpumask_next(n, mask); in cpumask_next_wrap()
|
H A D | group_cpus.c | 34 sibl = cpumask_next(sibl, siblmsk); in grp_spread_init_one()
|
/linux/drivers/net/wireguard/ |
H A D | queueing.h | 114 cpu = cpumask_next(cpu, cpu_online_mask); in wg_cpumask_choose_online() 127 int cpu = cpumask_next(READ_ONCE(*last_cpu), cpu_online_mask); in wg_cpumask_next_online()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | affinity.c | 115 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 122 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 643 curr_cpu = cpumask_next(curr_cpu, in hfi1_dev_affinity_init() 658 curr_cpu = cpumask_next(curr_cpu, in hfi1_dev_affinity_init() 979 curr_cpu = cpumask_next(curr_cpu, hw_thread_mask); in find_hw_thread_mask() 983 curr_cpu = cpumask_next(curr_cpu, hw_thread_mask); in find_hw_thread_mask()
|
/linux/kernel/ |
H A D | watchdog_buddy.c | 15 next_cpu = cpumask_next(cpu, &watchdog_cpus); in watchdog_next_cpu()
|
H A D | torture.c | 539 shuffle_idle_cpu = cpumask_next(shuffle_idle_cpu, shuffle_tmp_mask); in torture_shuffle_tasks()
|
/linux/kernel/rcu/ |
H A D | rcu.h | 415 (cpu) = cpumask_next((rnp)->grplo - 1, cpu_possible_mask); \ 417 (cpu) = cpumask_next((cpu), cpu_possible_mask))
|
/linux/kernel/sched/ |
H A D | stats.c | 188 n = cpumask_next(n - 1, cpu_online_mask); in schedstat_start()
|
/linux/kernel/time/ |
H A D | clocksource.c | 328 cpu = cpumask_next(cpu, cpu_online_mask); in clocksource_verify_choose_cpus() 346 cpu = cpumask_next(cpu - 1, cpu_online_mask); in clocksource_verify_choose_cpus() 592 next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in clocksource_watchdog()
|
H A D | timer_list.c | 308 iter->cpu = cpumask_next(iter->cpu, cpu_online_mask); in move_iter()
|
/linux/kernel/irq/ |
H A D | ipi.c | 71 next = cpumask_next(next, dest); in irq_reserve_ipi()
|
/linux/arch/x86/kernel/ |
H A D | tsc_sync.c | 101 next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in tsc_sync_check_timer_fn()
|
/linux/arch/s390/kernel/ |
H A D | processor.c | 353 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_update()
|
/linux/arch/openrisc/kernel/ |
H A D | setup.c | 356 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_start()
|
/linux/arch/riscv/kernel/ |
H A D | cpu.c | 315 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_start()
|
/linux/arch/powerpc/kernel/ |
H A D | rtasd.c | 435 cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in rtas_event_scan()
|
/linux/drivers/infiniband/sw/siw/ |
H A D | siw_main.c | 183 i++, cpu = cpumask_next(cpu, tx_cpumask)) { in siw_get_tx_cpu()
|
/linux/net/openvswitch/ |
H A D | flow.c | 140 cpu = cpumask_next(cpu, flow->cpu_used_mask)) { in ovs_flow_stats_get() 165 cpu = cpumask_next(cpu, flow->cpu_used_mask)) { in ovs_flow_stats_clear()
|
H A D | flow_table.c | 119 cpu = cpumask_next(cpu, flow->cpu_used_mask)) { in flow_free()
|
/linux/kernel/trace/ |
H A D | trace_hwlat.c | 328 next_cpu = cpumask_next(raw_smp_processor_id(), current_mask); in move_to_next_cpu()
|
/linux/kernel/bpf/ |
H A D | bpf_lru_list.c | 24 cpu = cpumask_next(cpu, cpu_possible_mask); in get_next_cpu()
|
/linux/drivers/crypto/caam/ |
H A D | qi.c | 445 *pcpu = cpumask_next(*pcpu, cpus); in caam_drv_ctx_init()
|
/linux/include/linux/ |
H A D | cpumask.h | 217 unsigned int cpumask_next(int n, const struct cpumask *srcp) in cpumask_next() function
|