/* SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause */ /* Authors: Cheng Xu */ /* Kai Shen */ /* Copyright (c) 2020-2022, Alibaba Group. */ #ifndef __ERDMA_H__ #define __ERDMA_H__ #include #include #include #include #include #include "erdma_hw.h" #define DRV_MODULE_NAME "erdma" #define ERDMA_NODE_DESC "Elastic RDMA(iWARP) stack" struct erdma_eq { void *qbuf; dma_addr_t qbuf_dma_addr; spinlock_t lock; u32 depth; u16 ci; u16 rsvd; atomic64_t event_num; atomic64_t notify_num; void __iomem *db; u64 *dbrec; dma_addr_t dbrec_dma; }; struct erdma_cmdq_sq { void *qbuf; dma_addr_t qbuf_dma_addr; spinlock_t lock; u32 depth; u16 ci; u16 pi; u16 wqebb_cnt; u64 *dbrec; dma_addr_t dbrec_dma; }; struct erdma_cmdq_cq { void *qbuf; dma_addr_t qbuf_dma_addr; spinlock_t lock; u32 depth; u32 ci; u32 cmdsn; u64 *dbrec; dma_addr_t dbrec_dma; atomic64_t armed_num; }; enum { ERDMA_CMD_STATUS_INIT, ERDMA_CMD_STATUS_ISSUED, ERDMA_CMD_STATUS_FINISHED, ERDMA_CMD_STATUS_TIMEOUT }; struct erdma_comp_wait { struct completion wait_event; u32 cmd_status; u32 ctx_id; u16 sq_pi; u8 comp_status; u8 rsvd; u32 comp_data[4]; }; enum { ERDMA_CMDQ_STATE_OK_BIT = 0, ERDMA_CMDQ_STATE_TIMEOUT_BIT = 1, ERDMA_CMDQ_STATE_CTX_ERR_BIT = 2, }; #define ERDMA_CMDQ_TIMEOUT_MS 15000 #define ERDMA_REG_ACCESS_WAIT_MS 20 #define ERDMA_WAIT_DEV_DONE_CNT 500 struct erdma_cmdq { unsigned long *comp_wait_bitmap; struct erdma_comp_wait *wait_pool; spinlock_t lock; bool use_event; struct erdma_cmdq_sq sq; struct erdma_cmdq_cq cq; struct erdma_eq eq; unsigned long state; struct semaphore credits; u16 max_outstandings; }; #define COMPROMISE_CC ERDMA_CC_CUBIC enum erdma_cc_alg { ERDMA_CC_NEWRENO = 0, ERDMA_CC_CUBIC, ERDMA_CC_HPCC_RTT, ERDMA_CC_HPCC_ECN, ERDMA_CC_HPCC_INT, ERDMA_CC_METHODS_NUM }; struct erdma_devattr { u32 fw_version; unsigned char peer_addr[ETH_ALEN]; unsigned long cap_flags; int numa_node; enum erdma_cc_alg cc; u32 irq_num; u32 max_qp; u32 max_send_wr; u32 max_recv_wr; u32 max_ord; u32 max_ird; u32 max_send_sge; u32 max_recv_sge; u32 max_sge_rd; u32 max_cq; u32 max_cqe; u64 max_mr_size; u32 max_mr; u32 max_pd; u32 max_mw; u32 local_dma_key; }; #define ERDMA_IRQNAME_SIZE 50 struct erdma_irq { char name[ERDMA_IRQNAME_SIZE]; u32 msix_vector; cpumask_t affinity_hint_mask; }; struct erdma_eq_cb { bool ready; void *dev; /* All EQs use this fields to get erdma_dev struct */ struct erdma_irq irq; struct erdma_eq eq; struct tasklet_struct tasklet; }; struct erdma_resource_cb { unsigned long *bitmap; spinlock_t lock; u32 next_alloc_idx; u32 max_cap; }; enum { ERDMA_RES_TYPE_PD = 0, ERDMA_RES_TYPE_STAG_IDX = 1, ERDMA_RES_CNT = 2, }; struct erdma_dev { struct ib_device ibdev; struct net_device *netdev; struct pci_dev *pdev; struct notifier_block netdev_nb; struct workqueue_struct *reflush_wq; resource_size_t func_bar_addr; resource_size_t func_bar_len; u8 __iomem *func_bar; struct erdma_devattr attrs; /* physical port state (only one port per device) */ enum ib_port_state state; u32 mtu; /* cmdq and aeq use the same msix vector */ struct erdma_irq comm_irq; struct erdma_cmdq cmdq; struct erdma_eq aeq; struct erdma_eq_cb ceqs[ERDMA_NUM_MSIX_VEC - 1]; spinlock_t lock; struct erdma_resource_cb res_cb[ERDMA_RES_CNT]; struct xarray qp_xa; struct xarray cq_xa; u32 next_alloc_qpn; u32 next_alloc_cqn; atomic_t num_ctx; struct list_head cep_list; struct dma_pool *db_pool; struct dma_pool *resp_pool; }; static inline void *get_queue_entry(void *qbuf, u32 idx, u32 depth, u32 shift) { idx &= (depth - 1); return qbuf + (idx << shift); } static inline struct erdma_dev *to_edev(struct ib_device *ibdev) { return container_of(ibdev, struct erdma_dev, ibdev); } static inline u32 erdma_reg_read32(struct erdma_dev *dev, u32 reg) { return readl(dev->func_bar + reg); } static inline u64 erdma_reg_read64(struct erdma_dev *dev, u32 reg) { return readq(dev->func_bar + reg); } static inline void erdma_reg_write32(struct erdma_dev *dev, u32 reg, u32 value) { writel(value, dev->func_bar + reg); } static inline void erdma_reg_write64(struct erdma_dev *dev, u32 reg, u64 value) { writeq(value, dev->func_bar + reg); } static inline u32 erdma_reg_read32_filed(struct erdma_dev *dev, u32 reg, u32 filed_mask) { u32 val = erdma_reg_read32(dev, reg); return FIELD_GET(filed_mask, val); } #define ERDMA_GET(val, name) FIELD_GET(ERDMA_CMD_##name##_MASK, val) int erdma_cmdq_init(struct erdma_dev *dev); void erdma_finish_cmdq_init(struct erdma_dev *dev); void erdma_cmdq_destroy(struct erdma_dev *dev); void erdma_cmdq_build_reqhdr(u64 *hdr, u32 mod, u32 op); int erdma_post_cmd_wait(struct erdma_cmdq *cmdq, void *req, u32 req_size, u64 *resp0, u64 *resp1); void erdma_cmdq_completion_handler(struct erdma_cmdq *cmdq); int erdma_ceqs_init(struct erdma_dev *dev); void erdma_ceqs_uninit(struct erdma_dev *dev); void notify_eq(struct erdma_eq *eq); void *get_next_valid_eqe(struct erdma_eq *eq); int erdma_aeq_init(struct erdma_dev *dev); int erdma_eq_common_init(struct erdma_dev *dev, struct erdma_eq *eq, u32 depth); void erdma_eq_destroy(struct erdma_dev *dev, struct erdma_eq *eq); void erdma_aeq_event_handler(struct erdma_dev *dev); void erdma_ceq_completion_handler(struct erdma_eq_cb *ceq_cb); #endif