/freebsd/sys/dev/mlx5/mlx5_en/ |
H A D | mlx5_en_tx.c | 36 mlx5e_do_send_cqe_inline(struct mlx5e_sq *sq) in mlx5e_do_send_cqe_inline() 38 sq->cev_counter++; in mlx5e_do_send_cqe_inline() 40 if (sq->cev_counter >= sq->cev_factor) { in mlx5e_do_send_cqe_inline() 41 sq->cev_counter = 0; in mlx5e_do_send_cqe_inline() 48 mlx5e_do_send_cqe(struct mlx5e_sq *sq) in mlx5e_do_send_cqe() 51 return (mlx5e_do_send_cqe_inline(sq)); in mlx5e_do_send_cqe() 55 mlx5e_send_nop(struct mlx5e_sq *sq, u32 ds_cnt) in mlx5e_send_nop() 57 u16 pi = sq->pc & sq in mlx5e_send_nop() 35 mlx5e_do_send_cqe_inline(struct mlx5e_sq * sq) mlx5e_do_send_cqe_inline() argument 47 mlx5e_do_send_cqe(struct mlx5e_sq * sq) mlx5e_do_send_cqe() argument 54 mlx5e_send_nop(struct mlx5e_sq * sq,u32 ds_cnt) mlx5e_send_nop() argument 92 struct mlx5e_sq *sq; mlx5e_select_queue_by_send_tag() local 139 struct mlx5e_sq *sq; mlx5e_select_queue() local 178 mlx5e_get_l2_header_size(struct mlx5e_sq * sq,struct mbuf * mb) mlx5e_get_l2_header_size() argument 547 mlx5e_sq_dump_xmit(struct mlx5e_sq * sq,struct mlx5e_xmit_args * parg,struct mbuf ** mbp) mlx5e_sq_dump_xmit() argument 684 mlx5e_sq_xmit(struct mlx5e_sq * sq,struct mbuf ** mbp) mlx5e_sq_xmit() argument 1019 mlx5e_poll_tx_cq(struct mlx5e_sq * sq,int budget) mlx5e_poll_tx_cq() argument 1102 mlx5e_xmit_locked(if_t ifp,struct mlx5e_sq * sq,struct mbuf * mb) mlx5e_xmit_locked() argument 1140 struct mlx5e_sq *sq; mlx5e_xmit() local 1171 struct mlx5e_sq *sq = container_of(mcq, struct mlx5e_sq, cq.mcq); mlx5e_tx_cq_comp() local [all...] |
H A D | mlx5_en_rl.c | 99 mlx5e_rl_build_sq_param(rl, &cparam->sq); in mlx5e_rl_build_channel_param() 104 mlx5e_rl_create_sq(struct mlx5e_priv *priv, struct mlx5e_sq *sq, in mlx5e_rl_create_sq() argument 125 &sq->dma_tag))) in mlx5e_rl_create_sq() 128 sq->mkey_be = cpu_to_be32(priv->mr.key); in mlx5e_rl_create_sq() 129 sq->ifp = priv->ifp; in mlx5e_rl_create_sq() 130 sq->priv = priv; in mlx5e_rl_create_sq() 132 err = mlx5_wq_cyc_create(mdev, ¶m->wq, sqc_wq, &sq->wq, in mlx5e_rl_create_sq() 133 &sq->wq_ctrl); in mlx5e_rl_create_sq() 137 sq->wq.db = &sq->wq.db[MLX5_SND_DBR]; in mlx5e_rl_create_sq() 139 err = mlx5e_alloc_sq_db(sq); in mlx5e_rl_create_sq() [all …]
|
H A D | mlx5_en_hw_tls.c | 570 mlx5e_tls_send_static_parameters(struct mlx5e_sq *sq, struct mlx5e_tls_tag *ptag) in mlx5e_tls_send_static_parameters() argument 577 pi = sq->pc & sq->wq.sz_m1; in mlx5e_tls_send_static_parameters() 578 wqe = mlx5_wq_cyc_get_wqe(&sq->wq, pi); in mlx5e_tls_send_static_parameters() 582 wqe->ctrl.opmod_idx_opcode = cpu_to_be32((sq->pc << 8) | in mlx5e_tls_send_static_parameters() 584 wqe->ctrl.qpn_ds = cpu_to_be32((sq->sqn << 8) | ds_cnt); in mlx5e_tls_send_static_parameters() 587 if (mlx5e_do_send_cqe(sq)) in mlx5e_tls_send_static_parameters() 601 memcpy(sq->doorbell.d32, &wqe->ctrl, sizeof(sq->doorbell.d32)); in mlx5e_tls_send_static_parameters() 603 sq->mbuf[pi].mbuf = NULL; in mlx5e_tls_send_static_parameters() 604 sq->mbuf[pi].num_bytes = 0; in mlx5e_tls_send_static_parameters() 605 sq->mbuf[pi].num_wqebbs = DIV_ROUND_UP(ds_cnt, MLX5_SEND_WQEBB_NUM_DS); in mlx5e_tls_send_static_parameters() [all …]
|
H A D | mlx5_en_main.c | 50 struct mlx5e_sq_param sq; member 866 sq_stats = &pch->sq[j].stats; in mlx5e_update_stats_locked() 884 struct mlx5e_sq *sq = channel->sq; in mlx5e_update_stats_locked() local 886 if (sq == NULL) in mlx5e_update_stats_locked() 889 sq_stats = &sq->stats; in mlx5e_update_stats_locked() 1568 mlx5e_free_sq_db(struct mlx5e_sq *sq) in mlx5e_free_sq_db() argument 1570 int wq_sz = mlx5_wq_cyc_get_size(&sq->wq); in mlx5e_free_sq_db() 1574 if (sq->mbuf[x].mbuf != NULL) { in mlx5e_free_sq_db() 1575 bus_dmamap_unload(sq->dma_tag, sq->mbuf[x].dma_map); in mlx5e_free_sq_db() 1576 m_freem(sq->mbuf[x].mbuf); in mlx5e_free_sq_db() [all …]
|
/freebsd/lib/libthr/thread/ |
H A D | thr_sleepq.c | 63 struct sleepqueue *sq; in _sleepq_alloc() local 65 sq = __thr_calloc(1, sizeof(struct sleepqueue)); in _sleepq_alloc() 66 TAILQ_INIT(&sq->sq_blocked); in _sleepq_alloc() 67 SLIST_INIT(&sq->sq_freeq); in _sleepq_alloc() 68 return (sq); in _sleepq_alloc() 72 _sleepq_free(struct sleepqueue *sq) in _sleepq_free() argument 74 __thr_free(sq); in _sleepq_free() 100 struct sleepqueue *sq; in lookup() local 102 LIST_FOREACH(sq, &sc->sc_queues, sq_hash) in lookup() 103 if (sq->sq_wchan == wchan) in lookup() [all …]
|
/freebsd/contrib/unbound/services/ |
H A D | outside_network.c | 84 static void serviced_tcp_initiate(struct serviced_query* sq, sldns_buffer* buff); 94 static int serviced_udp_send(struct serviced_query* sq, sldns_buffer* buff); 97 static int serviced_tcp_send(struct serviced_query* sq, sldns_buffer* buff); 100 static void serviced_callbacks(struct serviced_query* sq, int error, 1006 if(outnet->dtenv && pend_tcp && w && w->sq && 1011 dt_msg_send_outside_query(outnet->dtenv, &w->sq->addr, 1012 &pend_tcp->pi->addr, comm_tcp, NULL, w->sq->zone, 1013 w->sq->zonelen, &tmp); 1447 log_assert(!pend->sq->busy); 1448 pend->sq->busy = 1; [all …]
|
/freebsd/contrib/unbound/edns-subnet/ |
H A D | subnetmod.c | 93 struct subnet_qstate *sq = (struct subnet_qstate*)regional_alloc( in subnet_new_qstate() local 95 if(!sq) in subnet_new_qstate() 97 qstate->minfo[id] = sq; in subnet_new_qstate() 98 memset(sq, 0, sizeof(*sq)); in subnet_new_qstate() 99 sq->started_no_cache_store = qstate->no_cache_store; in subnet_new_qstate() 100 sq->started_no_cache_lookup = qstate->no_cache_lookup; in subnet_new_qstate() 149 struct subnet_qstate *sq; in ecs_whitelist_check() local 152 if(!(sq=(struct subnet_qstate*)qstate->minfo[id])) in ecs_whitelist_check() 162 sq->subnet_sent_no_subnet = 0; in ecs_whitelist_check() 163 if(sq->ecs_server_out.subnet_validdata && ((sq->subnet_downstream && in ecs_whitelist_check() [all …]
|
/freebsd/sys/kern/ |
H A D | subr_sleepqueue.c | 172 static void sleepq_resume_thread(struct sleepqueue *sq, struct thread *td, 174 static void sleepq_remove_thread(struct sleepqueue *sq, struct thread *td); 250 sleepq_free(struct sleepqueue *sq) in sleepq_free() argument 253 uma_zfree(sleepq_zone, sq); in sleepq_free() 277 struct sleepqueue *sq; in sleepq_lookup() local 282 LIST_FOREACH(sq, &sc->sc_queues, sq_hash) in sleepq_lookup() 283 if (sq->sq_wchan == wchan) in sleepq_lookup() 284 return (sq); in sleepq_lookup() 311 struct sleepqueue *sq; in sleepq_add() local 332 sq = sleepq_lookup(wchan); in sleepq_add() [all …]
|
/freebsd/sys/dev/vnic/ |
H A D | nicvf_queues.c | 699 struct snd_queue *sq; in nicvf_snd_pkt_handler() local 703 sq = &nic->qs->sq[cqe_tx->sq_idx]; in nicvf_snd_pkt_handler() 705 hdr = (struct sq_hdr_subdesc *)GET_SQ_DESC(sq, cqe_tx->sqe_ptr); in nicvf_snd_pkt_handler() 714 dmap = (bus_dmamap_t)sq->snd_buff[cqe_tx->sqe_ptr].dmap; in nicvf_snd_pkt_handler() 715 bus_dmamap_unload(sq->snd_buff_dmat, dmap); in nicvf_snd_pkt_handler() 717 mbuf = (struct mbuf *)sq->snd_buff[cqe_tx->sqe_ptr].mbuf; in nicvf_snd_pkt_handler() 720 sq->snd_buff[cqe_tx->sqe_ptr].mbuf = NULL; in nicvf_snd_pkt_handler() 721 nicvf_put_sq_desc(sq, hdr->subdesc_cnt + 1); in nicvf_snd_pkt_handler() 739 struct snd_queue *sq = &qs->sq[cq_idx]; in nicvf_cq_intr_handler() local 812 taskqueue_enqueue(sq->snd_taskq, &sq->snd_task); in nicvf_cq_intr_handler() [all …]
|
/freebsd/contrib/ofed/libcxgb4/ |
H A D | cq.c | 53 V_CQE_QPID(wq->sq.qid)); in insert_recv_cqe() 86 V_CQE_QPID(wq->sq.qid)); in insert_sq_cqe() 104 if (wq->sq.flush_cidx == -1) in c4iw_flush_sq() 105 wq->sq.flush_cidx = wq->sq.cidx; in c4iw_flush_sq() 106 idx = wq->sq.flush_cidx; in c4iw_flush_sq() 107 BUG_ON(idx >= wq->sq.size); in c4iw_flush_sq() 108 while (idx != wq->sq.pidx) { in c4iw_flush_sq() 109 swsqe = &wq->sq.sw_sq[idx]; in c4iw_flush_sq() 113 if (wq->sq.oldest_read == swsqe) { in c4iw_flush_sq() 118 if (++idx == wq->sq.size) in c4iw_flush_sq() [all …]
|
H A D | verbs.c | 338 qhp->wq.sq.qid = resp.sqid; in create_qp_v0() 339 qhp->wq.sq.size = resp.sq_size; in create_qp_v0() 340 qhp->wq.sq.memsize = resp.sq_memsize; in create_qp_v0() 341 qhp->wq.sq.flags = 0; in create_qp_v0() 355 qhp->wq.sq.udb = dbva; in create_qp_v0() 356 qhp->wq.sq.queue = mmap(NULL, qhp->wq.sq.memsize, in create_qp_v0() 359 if (qhp->wq.sq.queue == MAP_FAILED) in create_qp_v0() 373 qhp->wq.sq.sw_sq = calloc(qhp->wq.sq.size, sizeof (struct t4_swsqe)); in create_qp_v0() 374 if (!qhp->wq.sq.sw_sq) in create_qp_v0() 384 qhp->wq.sq.udb, qhp->wq.sq.queue, in create_qp_v0() [all …]
|
H A D | t4.h | 354 struct t4_sq sq; member 408 return wq->sq.in_use == 0; in t4_sq_empty() 413 return wq->sq.in_use == (wq->sq.size - 1); in t4_sq_full() 418 return wq->sq.size - 1 - wq->sq.in_use; in t4_sq_avail() 423 return wq->sq.flags & T4_SQ_ONCHIP; in t4_sq_onchip() 428 wq->sq.in_use++; in t4_sq_produce() 429 if (++wq->sq.pidx == wq->sq.size) in t4_sq_produce() 430 wq->sq.pidx = 0; in t4_sq_produce() 431 wq->sq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_sq_produce() 432 if (wq->sq.wq_pidx >= wq->sq.size * T4_SQ_NUM_SLOTS) in t4_sq_produce() [all …]
|
H A D | qp.c | 52 dst = &wq->sq.queue->flits[wq->sq.wq_pidx * in copy_wr_to_sq() 67 end = (uintptr_t)&wq->sq.queue[wq->sq.size]; in copy_wr_to_sq() 74 memcpy(wq->sq.queue, src + len, total - len); in copy_wr_to_sq() 103 static int build_immd(struct t4_sq *sq, struct fw_ri_immd *immdp, in build_immd() argument 158 static int build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, in build_rdma_send() argument 180 ret = build_immd(sq, wqe->send.u.immd_src, wr, in build_rdma_send() 207 static int build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, in build_rdma_write() argument 221 ret = build_immd(sq, wqe->write.u.immd_src, wr, in build_rdma_write() 298 if (qid == qhp->wq.sq.qid) { in ring_kernel_db() 348 swsqe = &qhp->wq.sq.sw_sq[qhp->wq.sq.pidx]; in c4iw_post_send() [all …]
|
/freebsd/sys/dev/ice/ |
H A D | ice_controlq.c | 36 (qinfo)->sq.head = prefix##_ATQH; \ 37 (qinfo)->sq.tail = prefix##_ATQT; \ 38 (qinfo)->sq.len = prefix##_ATQLEN; \ 39 (qinfo)->sq.bah = prefix##_ATQBAH; \ 40 (qinfo)->sq.bal = prefix##_ATQBAL; \ 41 (qinfo)->sq.len_mask = prefix##_ATQLEN_ATQLEN_M; \ 42 (qinfo)->sq.len_ena_mask = prefix##_ATQLEN_ATQENABLE_M; \ 43 (qinfo)->sq.len_crit_mask = prefix##_ATQLEN_ATQCRIT_M; \ 44 (qinfo)->sq.head_mask = prefix##_ATQH_ATQH_M; \ 109 if (cq->sq in ice_alloc_ctrlq_sq_ring() 826 struct ice_ctl_q_ring *sq = &cq->sq; ice_clean_sq() local [all...] |
/freebsd/sys/contrib/ck/include/ |
H A D | ck_sequence.h | 42 ck_sequence_init(struct ck_sequence *sq) in ck_sequence_init() argument 45 ck_pr_store_uint(&sq->sequence, 0); in ck_sequence_init() 50 ck_sequence_read_begin(const struct ck_sequence *sq) in ck_sequence_read_begin() argument 55 version = ck_pr_load_uint(&sq->sequence); in ck_sequence_read_begin() 77 ck_sequence_read_retry(const struct ck_sequence *sq, unsigned int version) in ck_sequence_read_retry() argument 85 return ck_pr_load_uint(&sq->sequence) != version; in ck_sequence_read_retry() 97 ck_sequence_write_begin(struct ck_sequence *sq) in ck_sequence_write_begin() argument 104 ck_pr_store_uint(&sq->sequence, sq->sequence + 1); in ck_sequence_write_begin() 113 ck_sequence_write_end(struct ck_sequence *sq) in ck_sequence_write_end() argument 121 ck_pr_store_uint(&sq->sequence, sq->sequence + 1); in ck_sequence_write_end()
|
/freebsd/contrib/bearssl/src/symcipher/ |
H A D | aes_ct_cbcdec.c | 53 uint32_t q[8], sq[8]; in br_aes_ct_cbcdec_run() local 70 memcpy(sq, q, sizeof q); in br_aes_ct_cbcdec_run() 79 iv0 = sq[0]; in br_aes_ct_cbcdec_run() 80 iv1 = sq[2]; in br_aes_ct_cbcdec_run() 81 iv2 = sq[4]; in br_aes_ct_cbcdec_run() 82 iv3 = sq[6]; in br_aes_ct_cbcdec_run() 85 br_enc32le(buf + 16, q[1] ^ sq[0]); in br_aes_ct_cbcdec_run() 86 br_enc32le(buf + 20, q[3] ^ sq[2]); in br_aes_ct_cbcdec_run() 87 br_enc32le(buf + 24, q[5] ^ sq[4]); in br_aes_ct_cbcdec_run() 88 br_enc32le(buf + 28, q[7] ^ sq[6]); in br_aes_ct_cbcdec_run() [all …]
|
/freebsd/usr.bin/fortune/tools/ |
H A D | Troff.mac | 24 \(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq
|
/freebsd/sys/dev/cxgbe/iw_cxgbe/ |
H A D | cq.c | 215 V_CQE_QPID(wq->sq.qid)); in insert_recv_cqe() 248 V_CQE_QPID(wq->sq.qid)); in insert_sq_cqe() 266 if (wq->sq.flush_cidx == -1) in c4iw_flush_sq() 267 wq->sq.flush_cidx = wq->sq.cidx; in c4iw_flush_sq() 268 idx = wq->sq.flush_cidx; in c4iw_flush_sq() 269 BUG_ON(idx >= wq->sq.size); in c4iw_flush_sq() 270 while (idx != wq->sq.pidx) { in c4iw_flush_sq() 271 swsqe = &wq->sq.sw_sq[idx]; in c4iw_flush_sq() 275 if (wq->sq.oldest_read == swsqe) { in c4iw_flush_sq() 280 if (++idx == wq->sq.size) in c4iw_flush_sq() [all …]
|
H A D | t4.h | 361 struct t4_sq sq; member 415 static inline int t4_sq_onchip(struct t4_sq *sq) in t4_sq_onchip() argument 417 return sq->flags & T4_SQ_ONCHIP; in t4_sq_onchip() 422 return wq->sq.in_use == 0; in t4_sq_empty() 427 return wq->sq.in_use == (wq->sq.size - 1); in t4_sq_full() 432 return wq->sq.size - 1 - wq->sq.in_use; in t4_sq_avail() 437 wq->sq.in_use++; in t4_sq_produce() 438 if (++wq->sq.pidx == wq->sq.size) in t4_sq_produce() 439 wq->sq.pidx = 0; in t4_sq_produce() 440 wq->sq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_sq_produce() [all …]
|
H A D | qp.c | 115 wq->sq.memsize, wq->sq.queue, in destroy_qp() 116 dma_unmap_addr(&wq->sq, mapping)); in destroy_qp() 119 kfree(wq->sq.sw_sq); in destroy_qp() 121 c4iw_put_qpid(rdev, wq->sq.qid, uctx); in destroy_qp() 144 wq->sq.qid = c4iw_get_qpid(rdev, uctx); in create_qp() 145 if (!wq->sq.qid) in create_qp() 155 wq->sq.sw_sq = kzalloc(wq->sq.size * sizeof *wq->sq.sw_sq, in create_qp() 157 if (!wq->sq.sw_sq) { in create_qp() 181 wq->sq.queue = dma_alloc_coherent(rhp->ibdev.dma_device, wq->sq.memsize, in create_qp() 182 &(wq->sq.dma_addr), GFP_KERNEL); in create_qp() [all …]
|
/freebsd/contrib/ofed/libmlx4/ |
H A D | qp.c | 66 return qp->buf.buf + qp->sq.offset + (n << qp->sq.wqe_shift); in get_send_wqe() 86 qp->sq.head = 0; in mlx4_init_qp_indices() 87 qp->sq.tail = 0; in mlx4_init_qp_indices() 97 for (i = 0; i < qp->sq.wqe_cnt; ++i) { in mlx4_qp_init_sq_ownership() 100 ctrl->fence_size = 1 << (qp->sq.wqe_shift - 4); in mlx4_qp_init_sq_ownership() 227 pthread_spin_lock(&qp->sq.lock); in mlx4_post_send() 231 ind = qp->sq.head; in mlx4_post_send() 234 if (wq_overflow(&qp->sq, nreq, to_mcq(ibqp->send_cq))) { in mlx4_post_send() 240 if (wr->num_sge > qp->sq.max_gs) { in mlx4_post_send() 252 ctrl = wqe = get_send_wqe(qp, ind & (qp->sq.wqe_cnt - 1)); in mlx4_post_send() [all …]
|
/freebsd/sys/dev/bnxt/bnxt_re/ |
H A D | qplib_fp.c | 53 qp->sq.condition = false; in bnxt_re_legacy_cancel_phantom_processing() 54 qp->sq.legacy_send_phantom = false; in bnxt_re_legacy_cancel_phantom_processing() 55 qp->sq.single = false; in bnxt_re_legacy_cancel_phantom_processing() 65 if (!qp->sq.flushed) { in __bnxt_qplib_add_flush_qp() 71 qp->sq.flushed = true; in __bnxt_qplib_add_flush_qp() 115 if (qp->sq.flushed) { in __bnxt_qplib_del_flush_qp() 116 qp->sq.flushed = false; in __bnxt_qplib_del_flush_qp() 132 qp->sq.hwq.prod = 0; in bnxt_qplib_clean_qp() 133 qp->sq.hwq.cons = 0; in bnxt_qplib_clean_qp() 134 qp->sq.swq_start = 0; in bnxt_qplib_clean_qp() [all …]
|
/freebsd/bin/sh/tests/parser/ |
H A D | heredoc2.0 | 11 s='ast*que?non' sq=\' dq=\" 19 ${s+'$sq'x'$sq'} 21 )" = ${sq}x${sq}' 34 ${s##'$sq'ast'$sq'}
|
/freebsd/sys/dev/mthca/ |
H A D | mthca_qp.c | 221 (n << qp->sq.wqe_shift); in get_send_wqe() 224 (n << qp->sq.wqe_shift)) >> in get_send_wqe() 226 ((qp->send_wqe_offset + (n << qp->sq.wqe_shift)) & in get_send_wqe() 499 qp_attr->cap.max_send_wr = qp->sq.max; in mthca_query_qp() 501 qp_attr->cap.max_send_sge = qp->sq.max_gs; in mthca_query_qp() 607 if (qp->sq.max) in __mthca_modify_qp() 608 qp_context->sq_size_stride = ilog2(qp->sq.max) << 3; in __mthca_modify_qp() 609 qp_context->sq_size_stride |= qp->sq.wqe_shift - 4; in __mthca_modify_qp() 727 qp_context->snd_db_index = cpu_to_be32(qp->sq.db_index); in __mthca_modify_qp() 828 mthca_wq_reset(&qp->sq); in __mthca_modify_qp() [all …]
|
/freebsd/sys/net80211/ |
H A D | ieee80211_superg.c | 636 ieee80211_ff_age(struct ieee80211com *ic, struct ieee80211_stageq *sq, in ieee80211_ff_age() argument 643 if (sq->depth == 0) { in ieee80211_ff_age() 648 KASSERT(sq->head != NULL, ("stageq empty")); in ieee80211_ff_age() 650 head = sq->head; in ieee80211_ff_age() 651 while ((m = sq->head) != NULL && M_AGE_GET(m) < quanta) { in ieee80211_ff_age() 659 sq->head = m->m_nextpkt; in ieee80211_ff_age() 660 sq->depth--; in ieee80211_ff_age() 663 sq->tail = NULL; in ieee80211_ff_age() 674 stageq_add(struct ieee80211com *ic, struct ieee80211_stageq *sq, struct mbuf *m) in stageq_add() argument 680 if (sq->tail != NULL) { in stageq_add() [all …]
|