Lines Matching defs:vecp

654 	apix_vector_t *vecp;
670 vecp = xv_vector(cpuid, vector);
671 if (vecp == NULL) {
677 nipl = vecp->v_pri;
695 ASSERT(vecp->v_state != APIX_STATE_OBSOLETED);
699 (vecp->v_type & APIX_TYPE_FIXED) && apic_level_intr[vecp->v_inum])
700 apix_level_intr_pre_eoi(vecp->v_inum);
716 APIC_DEBUG_BUF_PUT(vecp->v_inum);
785 apix_vector_t *vecp = xv_vector(cpuid, vector);
788 ASSERT(vecp != NULL && LOCK_HELD(&apix_lock));
790 if (vecp->v_type == APIX_TYPE_FIXED)
791 apix_intx_set_shared(vecp->v_inum, 1);
794 if (vecp->v_share > 1)
798 if (vecp->v_type == APIX_TYPE_IPI)
808 (void) apix_setup_io_intr(vecp);
818 apix_vector_t *vecp = xv_vector(cpuid, vector);
821 ASSERT(vecp != NULL && LOCK_HELD(&apix_lock));
823 if (vecp->v_type == APIX_TYPE_FIXED)
824 apix_intx_set_shared(vecp->v_inum, -1);
827 if (vecp->v_share > 1)
831 if (vecp->v_type == APIX_TYPE_IPI)
840 apix_disable_vector(vecp);
855 apix_vector_t *vecp, *newp;
866 vecp = apixp->x_vectbl[i];
867 if (!IS_VECT_ENABLED(vecp))
870 apix_disable_vector(vecp);
877 vecp = apixp->x_vectbl[i];
878 if (!IS_VECT_ENABLED(vecp))
881 if (vecp->v_flags & APIX_VECT_USER_BOUND) {
885 type = vecp->v_type;
896 newp = apix_set_cpu(vecp, bindcpu, &ret);
898 newp = apix_grp_set_cpu(vecp, bindcpu, &ret);
904 vecp->v_cpuid, vecp->v_vector);
926 apix_vector_t *vecp;
937 vecp = xv_vector(cpun, i);
938 if (!IS_VECT_ENABLED(vecp))
941 apix_enable_vector(vecp);
952 vecp = xv_vector(n, i);
953 if (!IS_VECT_ENABLED(vecp) ||
954 vecp->v_bound_cpuid != cpun)
957 if (vecp->v_type != APIX_TYPE_MSI)
958 (void) apix_set_cpu(vecp, cpun, &ret);
960 (void) apix_grp_set_cpu(vecp, cpun, &ret);
1145 apix_vector_t *vecp, *newvecp;
1192 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1193 if (vecp != NULL) {
1194 *result = APIX_VIRTVECTOR(vecp->v_cpuid,
1195 vecp->v_vector);
1206 vecp = apix_intx_xlate_vector(dip, hdlp->ih_inum,
1208 *result = (vecp == NULL) ? APIX_INVALID_VECT :
1209 APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector);
1215 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1216 if (vecp == NULL)
1219 *result = apix_get_pending(vecp);
1225 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1226 if (vecp == NULL)
1229 apix_intx_clear_mask(vecp->v_inum);
1235 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1236 if (vecp == NULL)
1239 apix_intx_set_mask(vecp->v_inum);
1245 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1246 if (vecp == NULL)
1249 *result = apix_intx_get_shared(vecp->v_inum);
1280 vecp = apix_get_req_vector(hdlp, hdlp->ih_flags);
1281 if (!IS_VECT_ENABLED(vecp)) {
1292 newvecp = apix_set_cpu(vecp, target, result);
1294 newvecp = apix_grp_set_cpu(vecp, target, result);
1374 apix_vector_t *vecp;
1381 if ((vecp = xv_vector(i, j)) != NULL)
1382 vecp->v_busy = 0;
1447 apix_get_pending(apix_vector_t *vecp)
1453 affinity_set(vecp->v_cpuid);
1455 index = vecp->v_vector / 32;
1456 bit = vecp->v_vector % 32;
1463 if (!pending && vecp->v_type == APIX_TYPE_FIXED)
1464 pending = apix_intx_get_pending(vecp->v_inum);
1472 apix_vector_t *vecp;
1486 vecp = xv_vector(cpuid, APIX_VIRTVEC_VECTOR(virt_vec));
1489 vecp = apix_get_dev_map(hdlp->ih_dip, hdlp->ih_inum,
1496 return (vecp);
1503 apix_vector_t *vecp;
1507 vecp = apix_get_req_vector(hdlp, intr_params_p->avgi_req_flags);
1508 if (IS_VECT_FREE(vecp)) {
1516 intr_params_p->avgi_cpu_id = vecp->v_cpuid;
1526 intr_params_p->avgi_vector = vecp->v_vector;
1531 intr_params_p->avgi_num_devs = vecp->v_share;
1539 for (i = 0, av_dev = vecp->v_autovect; av_dev;
1573 for (i = 0, av_dev = vecp->v_autovect; av_dev;
1592 apix_set_cpu(apix_vector_t *vecp, int new_cpu, int *result)
1607 if (vecp->v_type == APIX_TYPE_MSI) {
1608 if (i_ddi_intr_get_current_nintrs(APIX_GET_DIP(vecp)) > 1)
1611 return (apix_grp_set_cpu(vecp, new_cpu, result));
1617 if (vecp->v_type == APIX_TYPE_MSIX && IS_VECT_ENABLED(vecp)) {
1618 if ((dip = APIX_GET_DIP(vecp)) == NULL)
1620 inum = vecp->v_devp->dv_inum;
1640 if ((newp = apix_rebind(vecp, new_cpu, 1)) == NULL)
1654 apix_grp_set_cpu(apix_vector_t *vecp, int new_cpu, int *result)
1657 uint32_t orig_cpu = vecp->v_cpuid;
1658 int orig_vect = vecp->v_vector;
1666 " newcpu:%x\n", vecp->v_cpuid, vecp->v_vector, new_cpu));
1672 if (vecp->v_type != APIX_TYPE_MSI) {
1677 if ((dip = APIX_GET_DIP(vecp)) == NULL)
1688 if (vecp->v_inum != apix_get_min_dev_inum(dip, vecp->v_type))
1703 (vp->v_cpuid != vecp->v_cpuid))) {
1733 if ((newp = apix_rebind(vecp, new_cpu, num_vectors)) != NULL)
1790 apix_vector_t *vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector);
1801 apic_vt_ops->apic_intrmap_alloc_entry(&vecp->v_intrmap_private, NULL,
1802 vecp->v_type, 1, ioapicindex);
1803 apic_vt_ops->apic_intrmap_map_entry(vecp->v_intrmap_private,
1804 (void *)&irdt, vecp->v_type, 1);
1805 apic_vt_ops->apic_intrmap_record_rdt(vecp->v_intrmap_private, &irdt);
1814 vecp->v_state = APIX_STATE_ENABLED;
2517 apix_vector_t *vecp;
2522 if ((vecp = apix_alloc_intx(dip, inum, irqno)) == NULL)
2528 vecp->v_cpuid, vecp->v_vector));
2542 apix_vector_t *vecp;
2549 vecp = apix_intx_get_vector(irqno);
2551 return (vecp);