| /linux/arch/sparc/kernel/ |
| H A D | cpumap.c | 193 int n, id, cpu, prev_cpu, last_cpu, level; in build_cpuinfo_tree() local 233 for (last_cpu = (num_possible_cpus() - 1); last_cpu >= 0; last_cpu--) { in build_cpuinfo_tree() 234 if (cpu_online(last_cpu)) in build_cpuinfo_tree() 238 while (++cpu <= last_cpu) { in build_cpuinfo_tree() 250 if ((id != prev_id[level]) || (cpu == last_cpu)) { in build_cpuinfo_tree() 256 if (cpu == last_cpu) in build_cpuinfo_tree() 268 (cpu == last_cpu) ? cpu : prev_cpu; in build_cpuinfo_tree()
|
| /linux/arch/alpha/kernel/ |
| H A D | irq.c | 49 static int last_cpu; in irq_select_affinity() local 50 int cpu = last_cpu + 1; in irq_select_affinity() 62 last_cpu = cpu; in irq_select_affinity()
|
| /linux/drivers/opp/ |
| H A D | cpu.c | 108 int last_cpu) in _dev_pm_opp_cpumask_remove_table() argument 116 if (cpu == last_cpu) in _dev_pm_opp_cpumask_remove_table()
|
| H A D | opp.h | 262 void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask, int last_cpu);
|
| /linux/arch/x86/include/asm/fpu/ |
| H A D | sched.h | 49 old_fpu->last_cpu = cpu; in switch_fpu()
|
| H A D | types.h | 484 unsigned int last_cpu; member
|
| /linux/tools/perf/util/ |
| H A D | cpumap.c | 685 struct perf_cpu c, last_cpu = perf_cpu_map__max(map); in cpu_map__snprint_mask() local 690 if (last_cpu.cpu < 0) { in cpu_map__snprint_mask() 695 bitmap = zalloc(last_cpu.cpu / 8 + 1); in cpu_map__snprint_mask() 704 for (int cpu = last_cpu.cpu / 4 * 4; cpu >= 0; cpu -= 4) { in cpu_map__snprint_mask()
|
| /linux/drivers/net/wireguard/ |
| H A D | device.h | 31 int last_cpu; member
|
| H A D | queueing.c | 31 queue->last_cpu = -1; in wg_packet_queue_init()
|
| H A D | receive.c | 568 cpu = wg_cpumask_next_online(&wg->handshake_queue.last_cpu); in wg_packet_receive()
|
| /linux/drivers/xen/ |
| H A D | xen-acpi-processor.c | 293 unsigned int i, last_cpu, max_acpi_id = 0; in get_max_acpi_id() local 303 last_cpu = op.u.pcpu_info.max_present; in get_max_acpi_id() 304 for (i = 0; i <= last_cpu; i++) { in get_max_acpi_id()
|
| /linux/arch/powerpc/kvm/ |
| H A D | mpic.c | 150 int last_cpu; member 448 if (src->destmask == (1 << src->last_cpu)) { in openpic_update_irq() 450 IRQ_local_pipe(opp, src->last_cpu, n_IRQ, active, was_active); in openpic_update_irq() 461 for (i = src->last_cpu + 1; i != src->last_cpu; i++) { in openpic_update_irq() 468 src->last_cpu = i; in openpic_update_irq()
|
| /linux/kernel/ |
| H A D | smp.c | 785 int cpu, last_cpu, this_cpu = smp_processor_id(); in smp_call_function_many_cond() local 847 last_cpu = cpu; in smp_call_function_many_cond() 857 send_call_function_single_ipi(last_cpu); in smp_call_function_many_cond()
|
| /linux/drivers/crypto/caam/ |
| H A D | qi.c | 65 static DEFINE_PER_CPU(int, last_cpu); 443 int *pcpu = &get_cpu_var(last_cpu); in caam_drv_ctx_init() 447 put_cpu_var(last_cpu); in caam_drv_ctx_init()
|
| /linux/tools/virtio/ |
| H A D | vringh_test.c | 144 unsigned int first_cpu, last_cpu; in parallel_test() local 171 find_cpus(&first_cpu, &last_cpu); in parallel_test() 172 printf("Using CPUS %u and %u\n", first_cpu, last_cpu); in parallel_test()
|
| /linux/arch/x86/kernel/fpu/ |
| H A D | core.c | 676 dst_fpu->last_cpu = -1; in fpu_clone() 930 fpu->last_cpu = smp_processor_id(); in fpregs_mark_activate()
|
| /linux/tools/perf/ |
| H A D | builtin-ftrace.c | 393 int last_cpu; in set_tracing_cpumask() local 395 last_cpu = perf_cpu_map__cpu(cpumap, perf_cpu_map__nr(cpumap) - 1).cpu; in set_tracing_cpumask() 396 mask_size = last_cpu / 4 + 2; /* one more byte for EOS */ in set_tracing_cpumask() 397 mask_size += last_cpu / 32; /* ',' is needed for every 32th cpus */ in set_tracing_cpumask()
|
| /linux/arch/x86/events/ |
| H A D | core.c | 694 event->hw.last_cpu = -1; in __x86_pmu_event_init() 1245 hwc->last_cpu = smp_processor_id(); in x86_assign_hw_event() 1302 hwc->last_cpu == smp_processor_id() && in match_prev_assignment()
|
| /linux/drivers/scsi/lpfc/ |
| H A D | lpfc_sli4.h | 164 uint16_t last_cpu; /* most recent cpu */ member
|
| H A D | lpfc_init.c | 1375 ena_delay[eq->last_cpu] = 1; in lpfc_hb_eq_delay_work() 1392 if (unlikely(eq->last_cpu != i)) { in lpfc_hb_eq_delay_work() 1394 eq->last_cpu); in lpfc_hb_eq_delay_work() 10523 qdesc->last_cpu = qdesc->chann; in lpfc_sli4_queue_create() 10528 eqi = per_cpu_ptr(phba->sli4_hba.eq_info, qdesc->last_cpu); in lpfc_sli4_queue_create()
|
| H A D | lpfc_sli.c | 15646 fpeq->last_cpu = raw_smp_processor_id(); in lpfc_sli4_hba_intr_handler() 16291 fpeq->last_cpu = raw_smp_processor_id(); in lpfc_sli4_hba_intr_handler_th()
|
| /linux/include/linux/ |
| H A D | perf_event.h | 159 int last_cpu; member
|