Lines Matching defs:cc
152 * @cc: svc_rdma_chunk_ctxt to be initialized
155 struct svc_rdma_chunk_ctxt *cc)
157 struct rpc_rdma_cid *cid = &cc->cc_cid;
162 INIT_LIST_HEAD(&cc->cc_rwctxts);
163 cc->cc_sqecount = 0;
169 * @cc: svc_rdma_chunk_ctxt to be released
173 struct svc_rdma_chunk_ctxt *cc,
180 trace_svcrdma_cc_release(&cc->cc_cid, cc->cc_sqecount);
183 while ((ctxt = svc_rdma_next_ctxt(&cc->cc_rwctxts)) != NULL) {
241 struct svc_rdma_chunk_ctxt *cc = &ctxt->sc_reply_info.wi_cc;
243 if (!cc->cc_sqecount)
245 svc_rdma_cc_release(rdma, cc, DMA_TO_DEVICE);
258 struct svc_rdma_chunk_ctxt *cc =
264 trace_svcrdma_wc_reply(&cc->cc_cid);
267 trace_svcrdma_wc_reply_flush(wc, &cc->cc_cid);
270 trace_svcrdma_wc_reply_err(wc, &cc->cc_cid);
287 struct svc_rdma_chunk_ctxt *cc =
290 container_of(cc, struct svc_rdma_write_info, wi_cc);
294 trace_svcrdma_wc_write(&cc->cc_cid);
297 trace_svcrdma_wc_write_flush(wc, &cc->cc_cid);
300 trace_svcrdma_wc_write_err(wc, &cc->cc_cid);
303 svc_rdma_wake_send_waiters(rdma, cc->cc_sqecount);
321 struct svc_rdma_chunk_ctxt *cc =
325 svc_rdma_wake_send_waiters(rdma, cc->cc_sqecount);
327 ctxt = container_of(cc, struct svc_rdma_recv_ctxt, rc_cc);
330 trace_svcrdma_wc_read(wc, &cc->cc_cid, ctxt->rc_readbytes,
331 cc->cc_posttime);
341 trace_svcrdma_wc_read_flush(wc, &cc->cc_cid);
344 trace_svcrdma_wc_read_err(wc, &cc->cc_cid);
351 svc_rdma_cc_release(rdma, cc, DMA_FROM_DEVICE);
363 struct svc_rdma_chunk_ctxt *cc)
373 if (cc->cc_sqecount > rdma->sc_sq_depth)
377 cqe = &cc->cc_cqe;
378 list_for_each(tmp, &cc->cc_rwctxts) {
388 if (atomic_sub_return(cc->cc_sqecount,
390 cc->cc_posttime = ktime_get();
398 trace_svcrdma_sq_full(rdma, &cc->cc_cid);
399 atomic_add(cc->cc_sqecount, &rdma->sc_sq_avail);
401 atomic_read(&rdma->sc_sq_avail) > cc->cc_sqecount);
402 trace_svcrdma_sq_retry(rdma, &cc->cc_cid);
405 trace_svcrdma_sq_post_err(rdma, &cc->cc_cid, ret);
412 atomic_add(cc->cc_sqecount, &rdma->sc_sq_avail);
474 struct svc_rdma_chunk_ctxt *cc = &info->wi_cc;
504 list_add(&ctxt->rw_list, &cc->cc_rwctxts);
505 cc->cc_sqecount += ret;
518 trace_svcrdma_small_wrch_err(&cc->cc_cid, remaining, info->wi_seg_no,
609 struct svc_rdma_chunk_ctxt *cc;
620 cc = &info->wi_cc;
626 trace_svcrdma_post_write_chunk(&cc->cc_cid, cc->cc_sqecount);
627 ret = svc_rdma_post_chunk_ctxt(rdma, cc);
685 struct svc_rdma_chunk_ctxt *cc = &info->wi_cc;
703 cqe = &cc->cc_cqe;
704 list_for_each(pos, &cc->cc_rwctxts) {
713 sctxt->sc_sqecount += cc->cc_sqecount;
715 trace_svcrdma_post_reply_chunk(&cc->cc_cid, cc->cc_sqecount);
736 struct svc_rdma_chunk_ctxt *cc = &head->rc_cc;
778 list_add(&ctxt->rw_list, &cc->cc_rwctxts);
779 cc->cc_sqecount += ret;
783 trace_svcrdma_page_overrun_err(&cc->cc_cid, head->rc_curpage);
1119 struct svc_rdma_chunk_ctxt *cc = &head->rc_cc;
1122 cc->cc_cqe.done = svc_rdma_wc_read_done;
1123 cc->cc_sqecount = 0;
1139 trace_svcrdma_post_read_chunk(&cc->cc_cid, cc->cc_sqecount);
1140 ret = svc_rdma_post_chunk_ctxt(rdma, cc);