/titanic_44/usr/src/uts/common/os/ |
H A D | putnext.c | 154 boolean_t queued = B_FALSE; in putnext() local 195 queued = qp->q_sqflags & Q_SQQUEUED; in putnext() 245 queued = qp->q_sqflags & Q_SQQUEUED; in putnext() 271 if (!queued) { in putnext() 382 ASSERT((sq->sq_flags & (SQ_EXCL|SQ_CIPUT)) || queued); in putnext() 383 ASSERT((flags & (SQ_EXCL|SQ_CIPUT)) || queued); in putnext() 435 boolean_t queued = B_FALSE; in put() local 463 queued = qp->q_sqflags & Q_SQQUEUED; in put() 512 queued = qp->q_sqflags & Q_SQQUEUED; in put() 538 if (!queued) { in put() [all …]
|
/titanic_44/usr/src/uts/common/io/ib/clients/rdsv3/ |
H A D | send.c | 809 uint16_be_t dport, int *queued) in rdsv3_send_queue_rm() argument 815 if (*queued) in rdsv3_send_queue_rm() 873 *queued = 1; in rdsv3_send_queue_rm() 880 return (*queued); in rdsv3_send_queue_rm() 944 int queued = 0, allocated_mr = 0; in rdsv3_sendmsg() local 1054 &queued); in rdsv3_sendmsg() 1055 if (!queued) { in rdsv3_sendmsg() 1076 dport, &queued))); in rdsv3_sendmsg() 1088 dport, &queued)) { in rdsv3_sendmsg() 1106 queued); in rdsv3_sendmsg() [all …]
|
/titanic_44/usr/src/cmd/mdb/common/modules/genunix/ |
H A D | taskq.c | 147 size_t tact, tcount, queued, maxq; in taskq() local 209 queued = tq.tq_tasks - tq.tq_executed; in taskq() 232 if (active && tact == 0 && queued == 0) in taskq() 243 addr, tq.tq_name, tact, tcount, queued); in taskq()
|
/titanic_44/usr/src/uts/common/avs/ns/rdc/ |
H A D | rdc_clnt.c | 1280 rdc_aio_coalesce(rdc_aio_t *queued, rdc_aio_t *new) in rdc_aio_coalesce() argument 1290 if ((queued == NULL) || in rdc_aio_coalesce() 1291 (queued->handle == NULL) || in rdc_aio_coalesce() 1295 if (queued->index != new->index || queued->len + new->len > in rdc_aio_coalesce() 1299 if ((queued->pos + queued->len == new->pos) || in rdc_aio_coalesce() 1300 (new->pos + new->len == queued->pos)) { in rdc_aio_coalesce() 1301 rc = nsc_alloc_abuf(queued->pos, queued->len + new->len, 0, in rdc_aio_coalesce() 1308 rc = nsc_copy(queued->handle, h, queued->pos, queued->pos, in rdc_aio_coalesce() 1309 queued->len); in rdc_aio_coalesce() 1321 krdc = &rdc_k_info[queued->index]; in rdc_aio_coalesce() [all …]
|
/titanic_44/usr/src/uts/common/io/ral/ |
H A D | rt2560_var.h | 104 int queued; member
|
H A D | rt2560.c | 577 ring->queued = 0; in rt2560_alloc_tx_ring() 644 ring->queued = 0; in rt2560_reset_tx_ring() 1084 sc->txq.queued--; in rt2560_tx_intr() 1088 sc->txq.queued < (RT2560_TX_RING_COUNT - 32)) { in rt2560_tx_intr() 1157 sc->prioq.queued--; in rt2560_prio_intr() 1454 if (sc->prioq.queued >= RT2560_PRIO_RING_COUNT) { in rt2560_mgmt_send() 1537 sc->prioq.queued++; /* IF > RT2560_PRIO_RING_COUNT? FULL */ in rt2560_mgmt_send() 1581 if (sc->txq.queued >= RT2560_TX_RING_COUNT - 1) { in rt2560_send() 1691 sc->txq.queued++; in rt2560_send()
|
/titanic_44/usr/src/uts/common/io/ib/mgt/ibcm/ |
H A D | ibcm_impl.c | 179 uint_t queued; member 1434 ibcm_open.queued--; in ibcm_open_done() 1493 if (ibcm_open.queued == 0 && ibcm_ok_to_start(&ibcm_open)) { in ibcm_open_enqueue() 1498 ibcm_open.queued++; in ibcm_open_enqueue() 1515 ibcm_open.queued--; in ibcm_open_dequeue() 1537 while (ibcm_open.queued > 0) { in ibcm_check_for_opens() 1575 ibcm_open.queued = 0; in ibcm_rc_flow_control_init()
|
/titanic_44/usr/src/uts/common/io/xge/hal/xgehal/ |
H A D | xge-queue.c | 257 int pages_max, xge_queued_f queued, void *queued_data) in xge_queue_create() argument 264 queue->queued_func = queued; in xge_queue_create()
|
/titanic_44/usr/src/uts/common/io/rwd/ |
H A D | rt2661_var.h | 57 int queued; member
|
H A D | rt2661.c | 619 ring->queued = 0; in rt2661_alloc_tx_ring() 705 ring->queued = 0; in rt2661_reset_tx_ring() 947 ring->queued--; in rt2661_tx_intr() 1423 if (ring->queued > RT2661_TX_RING_COUNT - 8) { in rt2661_send() 1511 ring->queued++; in rt2661_send() 1553 if (ring->queued >= RT2661_MGT_RING_COUNT) { in rt2661_mgmt_send() 1634 ring->queued++; in rt2661_mgmt_send()
|
/titanic_44/usr/src/uts/common/io/wpi/ |
H A D | wpivar.h | 100 int queued; member
|
H A D | wpi.c | 1045 ring->queued = 0; in wpi_alloc_tx_ring() 1142 ring->queued = 0; in wpi_reset_tx_ring() 1795 ring->queued--; 1796 if (ring->queued < 0) 1797 ring->queued = 0; 1798 if ((sc->sc_need_reschedule) && (ring->queued <= (ring->count << 3))) { 2104 if (ring->queued > ring->count - 64) { 2277 ring->queued++;
|
/titanic_44/usr/src/uts/common/io/iwk/ |
H A D | iwk2_var.h | 73 int queued; member
|
/titanic_44/usr/src/uts/common/io/rwn/ |
H A D | rt2860_var.h | 88 int queued; member
|
H A D | rt2860.c | 774 ring->queued = 0; in rt2860_alloc_tx_ring() 812 ring->queued = 0; in rt2860_reset_tx_ring() 1110 if (SLIST_EMPTY(&sc->data_pool) || (ring->queued > 15)) { in rt2860_send() 1251 ring->queued++; in rt2860_send() 1670 hw, ring->next, ring->queued); in rt2860_tx_intr() 1692 ring->queued--; in rt2860_tx_intr() 1696 (ring->queued < RT2860_TX_RING_COUNT)) { in rt2860_tx_intr()
|
/titanic_44/usr/src/uts/common/io/iwh/ |
H A D | iwh_var.h | 73 int queued; member
|
H A D | iwh.c | 1569 ring->queued = 0; in iwh_alloc_tx_ring() 1714 ring->queued = 0; in iwh_reset_tx_ring() 2469 ring->queued--; in iwh_tx_intr() 2470 if (ring->queued < 0) { in iwh_tx_intr() 2471 ring->queued = 0; in iwh_tx_intr() 2474 if ((sc->sc_need_reschedule) && (ring->queued <= (ring->count >> 3))) { in iwh_tx_intr() 3170 if (ring->queued > ring->count - IWH_MAX_WIN_SIZE) { in iwh_send() 3186 ring->queued++; in iwh_send() 3199 ring->queued--; in iwh_send() 3200 if ((sc->sc_need_reschedule) && (ring->queued <= 0)) { in iwh_send() [all …]
|
/titanic_44/usr/src/uts/common/io/iwp/ |
H A D | iwp_var.h | 73 int queued; member
|
H A D | iwp.c | 1508 ring->queued = 0; in iwp_alloc_tx_ring() 1651 ring->queued = 0; in iwp_reset_tx_ring() 2373 ring->queued--; in iwp_tx_intr() 2374 if (ring->queued < 0) { in iwp_tx_intr() 2375 ring->queued = 0; in iwp_tx_intr() 2378 if ((sc->sc_need_reschedule) && (ring->queued <= (ring->count >> 3))) { in iwp_tx_intr() 3008 if (ring->queued > ring->count - IWP_MAX_WIN_SIZE) { in iwp_send() 3024 ring->queued++; in iwp_send() 3037 ring->queued--; in iwp_send() 3038 if ((sc->sc_need_reschedule) && (ring->queued <= 0)) { in iwp_send() [all …]
|
/titanic_44/usr/src/cmd/bnu/ |
H A D | Grades | 24 # Grades provides a orderly means of partitioning queued jobs into
|
/titanic_44/usr/src/uts/common/io/mwl/ |
H A D | mwl.c | 551 ring->queued = 0; in mwl_alloc_tx_ring() 2761 if (ring->queued > 15) { in mwl_send() 2763 "no txbuf, %d\n", ring->queued); in mwl_send() 2869 mblen, ring->cur, ring->queued); in mwl_send() 2871 ring->queued++; in mwl_send() 3085 if (!(ring->queued)) { in mwl_tx_intr() 3124 ring->queued--; in mwl_tx_intr() 3128 ring->next, ring->queued); in mwl_tx_intr() 3131 (ring->queued < MWL_TX_RING_COUNT)) { in mwl_tx_intr()
|
/titanic_44/usr/src/cmd/lp/lib/lp/ |
H A D | llib-llp | 49 int Q; /* # requests queued to activate alert */
|
/titanic_44/usr/src/uts/sun4u/ml/ |
H A D | mach_interrupt.s | 256 ! If %g1 is less than zero, no interrupt will be queued.
|
/titanic_44/usr/src/uts/common/io/usb/clients/video/usbvc/ |
H A D | usbvc_v4l2.c | 1135 boolean_t queued = B_FALSE; in usbvc_v4l2_enqueue_buf() local 1155 queued = B_TRUE; in usbvc_v4l2_enqueue_buf() 1163 if (queued) { in usbvc_v4l2_enqueue_buf()
|
/titanic_44/usr/src/uts/common/io/scsi/targets/ |
H A D | st.c | 517 clock_t timeout_interval, int queued); 6448 int queued; in st_start() local 6459 queued = 0; in st_start() 6465 queued = 0; in st_start() 6471 queued = 1; in st_start() 6578 if (queued) { in st_start() 6627 ST_TRAN_BUSY_TIMEOUT, queued) == 0) { in st_start() 6661 clock_t timeout_interval, int queued) in st_handle_start_busy() argument 6679 ((queued) && (un->un_runql != bp))) { in st_handle_start_busy() 6684 if (queued) { in st_handle_start_busy() [all …]
|