| /linux/lib/ |
| H A D | group_cpus.c | 105 unsigned ncpus; member 114 return ln->ncpus - rn->ncpus; in ncpus_cmp_func() 141 node_groups[n].ncpus = UINT_MAX; in alloc_nodes_groups() 145 unsigned ncpus; in alloc_nodes_groups() local 148 ncpus = cpumask_weight(nmsk); in alloc_nodes_groups() 150 if (!ncpus) in alloc_nodes_groups() 152 remaining_ncpus += ncpus; in alloc_nodes_groups() 153 node_groups[n].ncpus = ncpus; in alloc_nodes_groups() 230 unsigned ngroups, ncpus; in alloc_nodes_groups() local 232 if (node_groups[n].ncpus == UINT_MAX) in alloc_nodes_groups() [all …]
|
| /linux/tools/testing/selftests/rcutorture/bin/ |
| H A D | cpus2use.sh | 17 ncpus=`grep '^processor' /proc/cpuinfo | wc -l` 21 awk -v ncpus=$ncpus '{ print ncpus * ($7 + $NF) / 100 }'` 24 idlecpus=$ncpus 26 awk -v ncpus=$ncpus -v idlecpus=$idlecpus < /dev/null '
|
| H A D | kvm.sh | 360 awk < $T/cfgcpu.sort > $T/cfgcpu.pack -v ncpus=$cpus ' 468 if (ja[1] == -1 && ncpus == 0) 471 njitter = ncpus; 486 if (cpusr[jn] > ncpus && ncpus != 0) 554 nc = ncpus; 560 if (ncpus == 0) { 569 nc = ncpus; 576 if (ncpus != 0) 584 -v ncpus=$cpus \
|
| H A D | kvm-build.sh | 44 ncpus="`getconf _NPROCESSORS_ONLN`" 45 make -j$((2 * ncpus)) $TORTURE_KMAKE_ARG > $resdir/Make.out 2>&1
|
| /linux/arch/x86/include/asm/trace/ |
| H A D | hyperv.h | 16 __field(unsigned int, ncpus) 21 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus); 27 __entry->ncpus, __entry->mm, 64 __field(unsigned int, ncpus) 67 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus); 71 __entry->ncpus, __entry->vector)
|
| /linux/arch/powerpc/platforms/powermac/ |
| H A D | smp.c | 271 int i, ncpus; in smp_psurge_probe() local 297 ncpus = 4; in smp_psurge_probe() 309 ncpus = 2; in smp_psurge_probe() 323 if (ncpus > NR_CPUS) in smp_psurge_probe() 324 ncpus = NR_CPUS; in smp_psurge_probe() 325 for (i = 1; i < ncpus ; ++i) in smp_psurge_probe() 562 static void __init smp_core99_setup_i2c_hwsync(int ncpus) in smp_core99_setup_i2c_hwsync() argument 696 static void __init smp_core99_setup(int ncpus) in smp_core99_setup() argument 704 smp_core99_setup_i2c_hwsync(ncpus); in smp_core99_setup() 753 for (i = 1; i < ncpus; ++i) in smp_core99_setup() [all …]
|
| /linux/tools/perf/util/ |
| H A D | bpf_ftrace.c | 25 int i, ncpus = 1, ntasks = 1; in perf_ftrace__latency_prepare_bpf() local 62 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf() 63 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in perf_ftrace__latency_prepare_bpf() 89 for (i = 0; i < ncpus; i++) { in perf_ftrace__latency_prepare_bpf() 174 int ncpus = cpu__max_cpu().cpu; in perf_ftrace__latency_read_bpf() local 178 hist = calloc(ncpus, sizeof(*hist)); in perf_ftrace__latency_read_bpf() 189 for (i = 0; i < ncpus; i++) in perf_ftrace__latency_read_bpf()
|
| H A D | counts.c | 10 struct perf_counts *perf_counts__new(int ncpus, int nthreads) in perf_counts__new() argument 17 values = xyarray__new(ncpus, nthreads, sizeof(struct perf_counts_values)); in perf_counts__new() 25 values = xyarray__new(ncpus, nthreads, sizeof(bool)); in perf_counts__new()
|
| H A D | bpf_off_cpu.c | 152 int ncpus = 1, ntasks = 1, ncgrps = 1; in off_cpu_prepare() local 169 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in off_cpu_prepare() 170 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in off_cpu_prepare() 240 for (i = 0; i < ncpus; i++) { in off_cpu_prepare()
|
| /linux/arch/mips/kernel/ |
| H A D | crash.c | 59 unsigned int ncpus; in crash_kexec_prepare_cpus() local 64 ncpus = num_online_cpus() - 1;/* Excluding the panic cpu */ in crash_kexec_prepare_cpus() 75 while ((cpumask_weight(&cpus_in_crash) < ncpus) && (--msecs > 0)) { in crash_kexec_prepare_cpus()
|
| /linux/arch/sparc/kernel/ |
| H A D | setup_32.c | 378 int i, ncpus, err; in topology_init() local 384 ncpus = 0; in topology_init() 385 while (!cpu_find_by_instance(ncpus, NULL, NULL)) in topology_init() 386 ncpus++; in topology_init() 387 ncpus_probed = ncpus; in topology_init()
|
| H A D | sun4m_smp.c | 177 register int ncpus = SUN4M_NCPUS; in sun4m_cross_call() local 196 for (i = 0; i < ncpus; i++) { in sun4m_cross_call() 217 } while (++i < ncpus); in sun4m_cross_call() 225 } while (++i < ncpus); in sun4m_cross_call()
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | topology.h | 21 unsigned int shift, unsigned int ncpus); 50 unsigned int shift, unsigned int ncpus) in topology_update_dom() argument 53 tscan->dom_ncpus[dom] = ncpus; in topology_update_dom()
|
| /linux/drivers/clk/mvebu/ |
| H A D | clk-cpu.c | 173 int ncpus = num_possible_cpus(); in of_cpu_clk_setup() local 186 cpuclk = kcalloc(ncpus, sizeof(*cpuclk), GFP_KERNEL); in of_cpu_clk_setup() 190 clks = kcalloc(ncpus, sizeof(*clks), GFP_KERNEL); in of_cpu_clk_setup() 230 while(ncpus--) in of_cpu_clk_setup() 231 kfree(cpuclk[ncpus].clk_name); in of_cpu_clk_setup()
|
| /linux/drivers/xen/ |
| H A D | mcelog.c | 58 static uint32_t ncpus; variable 239 for (i = 0; i < ncpus; i++) in convert_log() 242 if (unlikely(i == ncpus)) { in convert_log() 377 ncpus = mc_op.u.mc_physcpuinfo.ncpus; in bind_virq_for_mce() 378 g_physinfo = kcalloc(ncpus, sizeof(struct mcinfo_logical_cpu), in bind_virq_for_mce()
|
| /linux/tools/lib/perf/ |
| H A D | evsel.c | 65 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_fd() argument 67 evsel->fd = xyarray__new(ncpus, nthreads, sizeof(int)); in perf_evsel__alloc_fd() 72 for (idx = 0; idx < ncpus; idx++) { in perf_evsel__alloc_fd() 85 static int perf_evsel__alloc_mmap(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_mmap() argument 87 evsel->mmap = xyarray__new(ncpus, nthreads, sizeof(struct perf_mmap)); in perf_evsel__alloc_mmap() 526 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_id() argument 528 if (ncpus == 0 || nthreads == 0) in perf_evsel__alloc_id() 531 evsel->sample_id = xyarray__new(ncpus, nthreads, sizeof(struct perf_sample_id)); in perf_evsel__alloc_id() 535 evsel->id = zalloc(ncpus * nthreads * sizeof(u64)); in perf_evsel__alloc_id()
|
| /linux/arch/xtensa/kernel/ |
| H A D | smp.c | 91 unsigned int ncpus = get_core_count(); in smp_init_cpus() local 94 pr_info("%s: Core Count = %d\n", __func__, ncpus); in smp_init_cpus() 97 if (ncpus > NR_CPUS) { in smp_init_cpus() 98 ncpus = NR_CPUS; in smp_init_cpus() 99 pr_info("%s: limiting core count by %d\n", __func__, ncpus); in smp_init_cpus() 102 for (i = 0; i < ncpus; ++i) in smp_init_cpus()
|
| /linux/tools/testing/selftests/powerpc/ |
| H A D | utils.c | 419 int ncpus, cpu = -1; in pick_online_cpu() local 423 ncpus = get_nprocs_conf(); in pick_online_cpu() 424 size = CPU_ALLOC_SIZE(ncpus); in pick_online_cpu() 425 mask = CPU_ALLOC(ncpus); in pick_online_cpu() 439 for (cpu = 8; cpu < ncpus; cpu += 8) in pick_online_cpu() 444 for (cpu = ncpus - 1; cpu >= 0; cpu--) in pick_online_cpu()
|
| /linux/arch/loongarch/kernel/ |
| H A D | machine_kexec.c | 182 unsigned int ncpus; in crash_smp_send_stop() local 196 ncpus = num_online_cpus() - 1; in crash_smp_send_stop() 207 while ((cpumask_weight(&cpus_in_crash) < ncpus) && timeout--) { in crash_smp_send_stop()
|
| /linux/arch/x86/platform/uv/ |
| H A D | uv_time.c | 52 int ncpus; member 56 } cpu[] __counted_by(ncpus); 158 head->ncpus = uv_blade_nr_possible_cpus(bid); in uv_rtc_allocate_timers() 177 for (c = 0; c < head->ncpus; c++) { in uv_rtc_find_next_timer()
|
| /linux/tools/lib/perf/include/internal/ |
| H A D | evsel.h | 137 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads); 143 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads);
|
| /linux/tools/testing/selftests/powerpc/benchmarks/ |
| H A D | context_switch.c | 108 int pid, ncpus; in start_process_on() local 121 ncpus = get_nprocs(); in start_process_on() 122 size = CPU_ALLOC_SIZE(ncpus); in start_process_on() 123 cpuset = CPU_ALLOC(ncpus); in start_process_on()
|
| /linux/drivers/misc/sgi-gru/ |
| H A D | grukservices.c | 145 int ctxnum, ncpus; in gru_load_kernel_context() local 163 ncpus = uv_blade_nr_possible_cpus(blade_id); in gru_load_kernel_context() 165 GRU_NUM_KERNEL_CBR * ncpus + bs->bs_async_cbrs); in gru_load_kernel_context() 167 GRU_NUM_KERNEL_DSR_BYTES * ncpus + in gru_load_kernel_context() 361 int ncpus; in gru_lock_async_resource() local 364 ncpus = uv_blade_nr_possible_cpus(blade_id); in gru_lock_async_resource() 366 *cb = bs->kernel_cb + ncpus * GRU_HANDLE_STRIDE; in gru_lock_async_resource() 368 *dsr = bs->kernel_dsr + ncpus * GRU_NUM_KERNEL_DSR_BYTES; in gru_lock_async_resource()
|
| /linux/drivers/dma-buf/ |
| H A D | st-dma-fence-chain.c | 445 int ncpus = num_online_cpus(); in find_race() local 455 threads = kmalloc_array(ncpus, sizeof(*threads), GFP_KERNEL); in find_race() 462 for (i = 0; i < ncpus; i++) { in find_race() 465 ncpus = i; in find_race() 476 for (i = 0; i < ncpus; i++) { in find_race()
|
| /linux/drivers/irqchip/ |
| H A D | irq-gic-v5-irs.c | 605 int ret, i, ncpus, niaffids; in gicv5_irs_of_init_affinity() local 607 ncpus = of_count_phandle_with_args(node, "cpus", NULL); in gicv5_irs_of_init_affinity() 608 if (ncpus < 0) in gicv5_irs_of_init_affinity() 613 if (niaffids != ncpus) in gicv5_irs_of_init_affinity() 624 for (i = 0; i < ncpus; i++) { in gicv5_irs_of_init_affinity()
|