Searched refs:mcpu (Results 1 – 6 of 6) sorted by relevance
529 struct machcpu *mcpu = &cpu->cpu_m; in hilevel_intr_prolog() local563 mcpu->pil_high_start[nestpil - (LOCK_LEVEL + 1)]; in hilevel_intr_prolog()564 mcpu->intrstat[nestpil][0] += intrtime; in hilevel_intr_prolog()582 mcpu->intrstat[t->t_pil][0] += intrtime; in hilevel_intr_prolog()591 mcpu->pil_high_start[pil - (LOCK_LEVEL + 1)] = now; in hilevel_intr_prolog()625 struct machcpu *mcpu = &cpu->cpu_m; in hilevel_intr_epilog() local630 ASSERT(mcpu->mcpu_pri == pil); in hilevel_intr_epilog()653 ASSERT(mcpu->pil_high_start[pil - (LOCK_LEVEL + 1)] != 0); in hilevel_intr_epilog()655 intrtime = now - mcpu->pil_high_start[pil - (LOCK_LEVEL + 1)]; in hilevel_intr_epilog()656 mcpu->intrstat[pil][0] += intrtime; in hilevel_intr_epilog()[all …]
134 xc_decrement(struct machcpu *mcpu) in xc_decrement() argument136 atomic_dec_32(&mcpu->xc_work_cnt); in xc_decrement()143 xc_increment(struct machcpu *mcpu) in xc_increment() argument147 old = mcpu->xc_work_cnt; in xc_increment()148 } while (atomic_cas_32(&mcpu->xc_work_cnt, old, old + 1) != old); in xc_increment()
231 struct machcpu *mcpu = &cpu->cpu_m; in apix_do_softint_prolog() local235 ASSERT(pil > mcpu->mcpu_pri && pil > cpu->cpu_base_spl); in apix_do_softint_prolog()237 atomic_and_32((uint32_t *)&mcpu->mcpu_softinfo.st_pending, ~(1 << pil)); in apix_do_softint_prolog()239 mcpu->mcpu_pri = pil; in apix_do_softint_prolog()256 mcpu->intrstat[pil][0] += intrtime; in apix_do_softint_prolog()302 struct machcpu *mcpu = &cpu->cpu_m; in apix_do_softint_epilog() local317 mcpu->intrstat[pil][0] += intrtime; in apix_do_softint_epilog()350 mcpu->mcpu_pri = pil; in apix_do_softint_epilog()406 struct machcpu *mcpu = &cpu->cpu_m; in apix_hilevel_intr_prolog() local412 ASSERT(pil > mcpu->mcpu_pri && pil > cpu->cpu_base_spl); in apix_hilevel_intr_prolog()[all …]
349 struct machcpu *mcpu; in intr_dequeue_req() local356 mcpu = &CPU->cpu_m; in intr_dequeue_req()361 next = mcpu->intr_head[pil]; in intr_dequeue_req()378 mcpu->intr_head[pil] = next_iv; /* head */ in intr_dequeue_req()381 mcpu->intr_tail[pil] = prev; /* tail */ in intr_dequeue_req()385 if (mcpu->intr_head[pil] == NULL) { in intr_dequeue_req()
177 struct machcpu *mcpu = &(cp->cpu_m); in cstate_wakeup() local178 volatile uint32_t *mcpu_mwait = mcpu->mcpu_mwait; in cstate_wakeup()927 struct machcpu *mcpu; in cpuidle_cstate_instance() local968 mcpu = &(cp->cpu_m); in cpuidle_cstate_instance()969 mcpu->max_cstates = cpu_acpi_get_max_cstates(handle); in cpuidle_cstate_instance()970 if (mcpu->max_cstates > CPU_ACPI_C1) { in cpuidle_cstate_instance()975 } else if (mcpu->max_cstates == CPU_ACPI_C1) { in cpuidle_cstate_instance()
149 struct machcpu *mcpu = &(cp->cpu_m); in cpupm_init() local233 mcpu->max_cstates = CPU_ACPI_C1; in cpupm_init()238 mcpu->max_cstates = cpu_acpi_get_max_cstates( in cpupm_init()240 if (mcpu->max_cstates > CPU_ACPI_C1) { in cpupm_init()244 mcpu->mcpu_idle_type = CPU_ACPI_C1; in cpupm_init()252 mcpu->max_cstates = CPU_ACPI_C1; in cpupm_init()