Home
last modified time | relevance | path

Searched refs:lldi (Results 1 – 25 of 26) sorted by relevance

12

/linux/drivers/infiniband/hw/cxgb4/
H A Ddevice.c102 le.poll_sge_ts = cxgb4_read_sge_timestamp(wq->rdev->lldi.ports[0]); in c4iw_log_wr_stats()
130 #define ts2ns(ts) div64_u64((ts) * dev->rdev.lldi.cclk_ps, 1000) in wr_log_show()
383 ret = cxgb4_read_tpte(stagd->devp->rdev.lldi.ports[0], (u32)id<<8, in dump_stag()
386 dev_err(&stagd->devp->rdev.lldi.pdev->dev, in dump_stag()
796 if (rdev->lldi.udb_density != rdev->lldi.ucq_density) { in c4iw_rdev_open()
798 pci_name(rdev->lldi.pdev), rdev->lldi.udb_density, in c4iw_rdev_open()
799 rdev->lldi.ucq_density); in c4iw_rdev_open()
802 if (rdev->lldi.vr->qp.start != rdev->lldi.vr->cq.start || in c4iw_rdev_open()
803 rdev->lldi.vr->qp.size != rdev->lldi.vr->cq.size) { in c4iw_rdev_open()
805 pci_name(rdev->lldi.pdev), rdev->lldi.vr->qp.start, in c4iw_rdev_open()
[all …]
H A Dprovider.c181 ret = dma_mmap_coherent(&rdev->lldi.pdev->dev, vma, in c4iw_mmap()
251 memcpy(&(gid->raw[0]), dev->rdev.lldi.ports[port-1]->dev_addr, 6); in c4iw_query_gid()
268 dev->rdev.lldi.ports[0]->dev_addr); in c4iw_query_device()
269 props->hw_ver = CHELSIO_CHIP_RELEASE(dev->rdev.lldi.adapter_type); in c4iw_query_device()
270 props->fw_ver = dev->rdev.lldi.fw_vers; in c4iw_query_device()
276 props->vendor_id = (u32)dev->rdev.lldi.pdev->vendor; in c4iw_query_device()
277 props->vendor_part_id = (u32)dev->rdev.lldi.pdev->device; in c4iw_query_device()
279 props->max_qp = dev->rdev.lldi.vr->qp.size / 2; in c4iw_query_device()
280 props->max_srq = dev->rdev.lldi.vr->srq.size; in c4iw_query_device()
287 props->max_res_rd_atom = dev->rdev.lldi.max_ird_adapter; in c4iw_query_device()
[all …]
H A Dcm.c212 error = cxgb4_l2t_send(rdev->lldi.ports[0], skb, l2e); in c4iw_l2t_send()
229 error = cxgb4_ofld_send(rdev->lldi.ports[0], skb); in c4iw_ofld_send()
250 ep->emss = ep->com.dev->rdev.lldi.mtus[TCPOPT_MSS_G(opt)] - in set_emss()
401 ep->com.dev->rdev.lldi.ports[0], in _c4iw_free_ep()
405 cxgb4_remove_tid(ep->com.dev->rdev.lldi.tids, 0, ep->hwtid, in _c4iw_free_ep()
558 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in act_open_req_arp_failure()
562 cxgb4_free_atid(ep->com.dev->rdev.lldi.tids, ep->atid); in act_open_req_arp_failure()
618 (ep->com.dev->rdev.lldi.pf)); in send_flowc()
736 enum chip_type adapter_type = ep->com.dev->rdev.lldi.adapter_type; in send_connect()
741 netdev = ep->com.dev->rdev.lldi.ports[0]; in send_connect()
[all …]
H A Dresource.c43 rdev->lldi.vr->qp.start, in c4iw_init_qid_table()
44 rdev->lldi.vr->qp.size, in c4iw_init_qid_table()
45 rdev->lldi.vr->qp.size, 0)) in c4iw_init_qid_table()
48 for (i = rdev->lldi.vr->qp.start; in c4iw_init_qid_table()
49 i < rdev->lldi.vr->qp.start + rdev->lldi.vr->qp.size; i++) in c4iw_init_qid_table()
308 pbl_start = rdev->lldi.vr->pbl.start; in c4iw_pblpool_create()
309 pbl_chunk = rdev->lldi.vr->pbl.size; in c4iw_pblpool_create()
350 pci_name(rdev->lldi.pdev)); in c4iw_rqtpool_alloc()
395 if (rdev->lldi.vr->srq.size) in c4iw_rqtpool_create()
398 rqt_start = rdev->lldi.vr->rq.start + skip; in c4iw_rqtpool_create()
[all …]
H A Dmem.c56 return (is_t4(dev->rdev.lldi.adapter_type) || in mr_exceeds_hw_limits()
57 is_t5(dev->rdev.lldi.adapter_type)) && in mr_exceeds_hw_limits()
92 T5_ULP_MEMIO_FID_V(rdev->lldi.rxq_ids[0])); in _c4iw_write_mem_dma_aligned()
120 if (is_t4(rdev->lldi.adapter_type)) in _c4iw_write_mem_inline()
200 daddr = dma_map_single(&rdev->lldi.pdev->dev, data, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma()
201 if (dma_mapping_error(&rdev->lldi.pdev->dev, daddr)) in _c4iw_write_mem_dma()
226 dma_unmap_single(&rdev->lldi.pdev->dev, save, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma()
240 if (!rdev->lldi.ulptx_memwrite_dsgl || !use_dsgl) { in write_adapter_mem()
255 pci_name(rdev->lldi.pdev)); in write_adapter_mem()
333 (rdev->lldi.vr->stag.start >> 5), in write_tpt_entry()
[all …]
H A Diw_cxgb4.h76 #define PBL_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->pbl.start)
77 #define RQT_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->rq.start)
180 struct cxgb4_lld_info lldi; member
205 return (int)(rdev->lldi.vr->stag.size >> 5); in c4iw_num_stags()
274 func, pci_name(rdev->lldi.pdev), hwtid, qpid); in c4iw_wait_for_reply()
281 pci_name(rdev->lldi.pdev), wr_waitp->ret, hwtid, qpid); in c4iw_wait_for_reply()
333 struct cxgb4_lld_info lldi; member
357 return min(dev->rdev.lldi.max_ordird_qp, c4iw_max_read_depth); in cur_max_read_depth()
577 if (addr >= pci_resource_start(rdev->lldi.pdev, 0) && in insert_flag_to_mmap()
578 (addr < (pci_resource_start(rdev->lldi.pdev, 0) + in insert_flag_to_mmap()
[all …]
H A Dqp.c74 dev_warn(&dev->rdev.lldi.pdev->dev, in alloc_ird()
102 dma_free_coherent(&(rdev->lldi.pdev->dev), sq->memsize, sq->queue, in dealloc_host_sq()
116 if (!ocqp_support || !ocqp_supported(&rdev->lldi)) in alloc_oc_sq()
122 rdev->lldi.vr->ocq.start; in alloc_oc_sq()
124 rdev->lldi.vr->ocq.start); in alloc_oc_sq()
131 sq->queue = dma_alloc_coherent(&(rdev->lldi.pdev->dev), sq->memsize, in alloc_host_sq()
162 dma_free_coherent(&rdev->lldi.pdev->dev, in destroy_qp()
184 ret = cxgb4_bar2_sge_qregs(rdev->lldi.ports[0], qid, qtype, in c4iw_bar2_addrs()
193 if (is_t4(rdev->lldi.adapter_type)) in c4iw_bar2_addrs()
264 wq->rq.queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, in create_qp()
[all …]
H A Dev.c43 ret = cxgb4_read_tpte(dev->rdev.lldi.ports[0], stag, in print_tpte()
46 dev_err(&dev->rdev.lldi.pdev->dev, in print_tpte()
66 dev_err(&dev->rdev.lldi.pdev->dev, in dump_err_cqe()
H A Dcq.c64 dma_free_coherent(&(rdev->lldi.pdev->dev), in destroy_cq()
98 cq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, cq->memsize, in create_cq()
142 rdev->lldi.ciq_ids[cq->vector])); in create_cq()
160 cq->gts = rdev->lldi.gts_reg; in create_cq()
168 pci_name(rdev->lldi.pdev), cq->cqid); in create_cq()
174 dma_free_coherent(&rdev->lldi.pdev->dev, cq->memsize, cq->queue, in create_cq()
1021 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
H A Drestrack.c447 ret = cxgb4_read_tpte(dev->rdev.lldi.ports[0], stag, (__be32 *)&tpte); in c4iw_fill_res_mr_entry()
449 dev_err(&dev->rdev.lldi.pdev->dev, in c4iw_fill_res_mr_entry()
/linux/drivers/scsi/cxgbi/cxgb4i/
H A Dcxgb4i.c211 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req() local
231 if (is_t4(lldi->adapter_type)) { in send_act_open_req()
254 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req()
320 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, in send_act_open_req()
330 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req6() local
350 if (is_t4(lldi->adapter_type)) { in send_act_open_req6()
375 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req6()
431 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, csk->state, in send_act_open_req6()
718 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(cdev); in cxgb4i_make_tx_data_wr() local
726 u32 force = is_t5(lldi->adapter_type) ? TX_FORCE_V(!submode) : in cxgb4i_make_tx_data_wr()
[all …]
/linux/drivers/target/iscsi/cxgbit/
H A Dcxgbit_ddp.c298 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_ddp_init() local
299 struct net_device *ndev = cdev->lldi.ports[0]; in cxgbit_ddp_init()
303 if (!lldi->vr->iscsi.size) { in cxgbit_ddp_init()
310 tformat.pgsz_order[i] = (lldi->iscsi_pgsz_order >> (i << 3)) in cxgbit_ddp_init()
312 cxgbi_tagmask_check(lldi->iscsi_tagmask, &tformat); in cxgbit_ddp_init()
314 ret = cxgbi_ppm_init(lldi->iscsi_ppm, cdev->lldi.ports[0], in cxgbit_ddp_init()
315 cdev->lldi.pdev, &cdev->lldi, &tformat, in cxgbit_ddp_init()
316 lldi->vr->iscsi.size, lldi->iscsi_llimit, in cxgbit_ddp_init()
317 lldi->vr->iscsi.start, 2, in cxgbit_ddp_init()
318 lldi->vr->ppod_edram.start, in cxgbit_ddp_init()
[all …]
H A Dcxgbit_cm.c63 func, pci_name(cdev->lldi.pdev), tid); in cxgbit_wait_for_reply()
69 pci_name(cdev->lldi.pdev), wr_waitp->ret, tid); in cxgbit_wait_for_reply()
153 __func__, cdev->lldi.ports[0]->name, stid, sin6->sin6_port); in cxgbit_create_server6()
158 ret = cxgb4_clip_get(cdev->lldi.ports[0], in cxgbit_create_server6()
170 ret = cxgb4_create_server6(cdev->lldi.ports[0], in cxgbit_create_server6()
173 cdev->lldi.rxq_ids[0]); in cxgbit_create_server6()
184 cxgb4_clip_release(cdev->lldi.ports[0], in cxgbit_create_server6()
204 __func__, cdev->lldi.ports[0]->name, stid, sin->sin_port); in cxgbit_create_server4()
209 ret = cxgb4_create_server(cdev->lldi.ports[0], in cxgbit_create_server4()
212 cdev->lldi.rxq_ids[0]); in cxgbit_create_server4()
[all …]
H A Dcxgbit_main.c33 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_set_mdsl() local
38 if (is_t5(lldi->adapter_type)) { in cxgbit_set_mdsl()
39 mdsl = min_t(u32, lldi->iscsi_iolen - CXGBIT_PDU_NONPAYLOAD_LEN, in cxgbit_set_mdsl()
42 mdsl = lldi->iscsi_iolen - CXGBIT_PDU_NONPAYLOAD_LEN; in cxgbit_set_mdsl()
53 static void *cxgbit_uld_add(const struct cxgb4_lld_info *lldi) in cxgbit_uld_add() argument
57 if (is_t4(lldi->adapter_type)) in cxgbit_uld_add()
67 cdev->lldi = *lldi; in cxgbit_uld_add()
78 pci_name(lldi->pdev)); in cxgbit_uld_add()
80 if (lldi->fw_vers >= 0x10d2b00) in cxgbit_uld_add()
91 pci_name(lldi->pdev)); in cxgbit_uld_add()
[all …]
H A Dcxgbit.h133 struct cxgb4_lld_info lldi; member
349 return (struct cxgbi_ppm *)(*cdev->lldi.iscsi_ppm); in cdev2ppm()
H A Dcxgbit_target.c163 const struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_tx_data_wr() local
169 u32 force = is_t5(lldi->adapter_type) ? TX_FORCE_V(!submode) : in cxgbit_tx_data_wr()
719 struct cxgbi_ppm *ppm = *csk->com.cdev->lldi.iscsi_ppm; in cxgbit_set_params()
750 if (is_t5(cdev->lldi.adapter_type)) in cxgbit_set_params()
1580 struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_rx_skb() local
1584 if (is_t5(lldi->adapter_type)) in cxgbit_rx_skb()
/linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/
H A Dchtls_main.c136 for (i = 0; i < cdev->lldi->nports; i++) { in chtls_inline_feature()
170 kfree(cdev->lldi); in chtls_free_uld()
197 strlcat(tlsdev->name, cdev->lldi->ports[0]->name, in chtls_register_dev()
237 struct cxgb4_lld_info *lldi; in chtls_uld_add() local
245 lldi = kzalloc(sizeof(*lldi), GFP_KERNEL); in chtls_uld_add()
246 if (!lldi) in chtls_uld_add()
252 *lldi = *info; in chtls_uld_add()
253 cdev->lldi = lldi; in chtls_uld_add()
254 cdev->pdev = lldi->pdev; in chtls_uld_add()
255 cdev->tids = lldi->tids; in chtls_uld_add()
[all …]
H A Dchtls_cm.c209 cxgb4_ofld_send(cdev->lldi->ports[0], skb); in abort_arp_failure()
691 cdev->lldi->rxq_ids[0]); in chtls_listen_start()
707 cdev->lldi->rxq_ids[0]); in chtls_listen_start()
742 cxgb4_remove_server(cdev->lldi->ports[0], stid, in chtls_listen_stop()
743 cdev->lldi->rxq_ids[0], sk->sk_family == PF_INET6); in chtls_listen_stop()
979 tp->advmss = cxgb4_best_aligned_mtu(cdev->lldi->mtus, in chtls_select_mss()
1011 struct cxgb4_lld_info *lldi; in chtls_pass_accept_rpl() local
1024 lldi = csk->cdev->lldi; in chtls_pass_accept_rpl()
1048 if (!is_t5(lldi->adapter_type)) in chtls_pass_accept_rpl()
1179 csk->l2t_entry = cxgb4_l2t_get(cdev->lldi->l2t, n, ndev, 0); in chtls_recv_sock()
[all …]
/linux/drivers/net/ethernet/chelsio/libcxgb/
H A Dlibcxgb_cm.c78 cxgb_our_interface(struct cxgb4_lld_info *lldi, in cxgb_our_interface() argument
85 for (i = 0; i < lldi->nports; i++) in cxgb_our_interface()
86 if (lldi->ports[i] == egress_dev) in cxgb_our_interface()
92 cxgb_find_route(struct cxgb4_lld_info *lldi, in cxgb_find_route() argument
109 if (!cxgb_our_interface(lldi, get_real_dev, n->dev) && in cxgb_find_route()
121 cxgb_find_route6(struct cxgb4_lld_info *lldi, in cxgb_find_route6() argument
138 (!cxgb_our_interface(lldi, get_real_dev, in cxgb_find_route6()
/linux/drivers/crypto/chelsio/
H A Dchcr_core.h107 struct cxgb4_lld_info lldi; member
127 return pci_get_drvdata(u_ctx->lldi.pdev); in padap()
H A Dchcr_core.c197 u_ctx->lldi = *lld; in chcr_uld_add()
/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/
H A Dchcr_ktls.h76 struct cxgb4_lld_info lldi; member
117 return u_ctx->lldi.rxq_ids[0]; in chcr_get_first_rx_qid()
H A Dchcr_ktls.c2114 static void *chcr_ktls_uld_add(const struct cxgb4_lld_info *lldi) in chcr_ktls_uld_add() argument
2125 u_ctx->lldi = *lldi; in chcr_ktls_uld_add()
2150 adap = pci_get_drvdata(u_ctx->lldi.pdev); in chcr_ktls_uld_rx_handler()
2206 pr_info("%s: Up\n", pci_name(u_ctx->lldi.pdev)); in chcr_ktls_uld_state_change()
2214 pr_info("%s: Down\n", pci_name(u_ctx->lldi.pdev)); in chcr_ktls_uld_state_change()
2255 adap = pci_get_drvdata(u_ctx->lldi.pdev); in chcr_ktls_exit()
/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/
H A Dchcr_ipsec.h22 struct cxgb4_lld_info lldi; member
H A Dchcr_ipsec.c111 u_ctx->lldi = *infop; in ch_ipsec_uld_add()
123 pr_info("%s: Up\n", pci_name(u_ctx->lldi.pdev)); in ch_ipsec_uld_state_change()
131 pr_info("%s: Down\n", pci_name(u_ctx->lldi.pdev)); in ch_ipsec_uld_state_change()
811 adap = pci_get_drvdata(u_ctx->lldi.pdev); in ch_ipsec_exit()

12