1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * AMD Passthru DMA device driver 4 * -- Based on the CCP driver 5 * 6 * Copyright (C) 2016,2021 Advanced Micro Devices, Inc. 7 * 8 * Author: Sanjay R Mehta <sanju.mehta@amd.com> 9 * Author: Tom Lendacky <thomas.lendacky@amd.com> 10 * Author: Gary R Hook <gary.hook@amd.com> 11 */ 12 13 #ifndef __PT_DEV_H__ 14 #define __PT_DEV_H__ 15 16 #include <linux/device.h> 17 #include <linux/dmaengine.h> 18 #include <linux/pci.h> 19 #include <linux/spinlock.h> 20 #include <linux/mutex.h> 21 #include <linux/list.h> 22 #include <linux/wait.h> 23 #include <linux/dmapool.h> 24 25 #include "../../virt-dma.h" 26 27 #define MAX_PT_NAME_LEN 16 28 #define MAX_DMAPOOL_NAME_LEN 32 29 30 #define MAX_HW_QUEUES 1 31 #define MAX_CMD_QLEN 100 32 33 #define PT_ENGINE_PASSTHRU 5 34 35 /* Register Mappings */ 36 #define IRQ_MASK_REG 0x040 37 #define IRQ_STATUS_REG 0x200 38 39 #define CMD_Q_ERROR(__qs) ((__qs) & 0x0000003f) 40 41 #define CMD_QUEUE_PRIO_OFFSET 0x00 42 #define CMD_REQID_CONFIG_OFFSET 0x04 43 #define CMD_TIMEOUT_OFFSET 0x08 44 #define CMD_PT_VERSION 0x10 45 46 #define CMD_Q_CONTROL_BASE 0x0000 47 #define CMD_Q_TAIL_LO_BASE 0x0004 48 #define CMD_Q_HEAD_LO_BASE 0x0008 49 #define CMD_Q_INT_ENABLE_BASE 0x000C 50 #define CMD_Q_INTERRUPT_STATUS_BASE 0x0010 51 52 #define CMD_Q_STATUS_BASE 0x0100 53 #define CMD_Q_INT_STATUS_BASE 0x0104 54 #define CMD_Q_DMA_STATUS_BASE 0x0108 55 #define CMD_Q_DMA_READ_STATUS_BASE 0x010C 56 #define CMD_Q_DMA_WRITE_STATUS_BASE 0x0110 57 #define CMD_Q_ABORT_BASE 0x0114 58 #define CMD_Q_AX_CACHE_BASE 0x0118 59 60 #define CMD_CONFIG_OFFSET 0x1120 61 #define CMD_CLK_GATE_CTL_OFFSET 0x6004 62 63 #define CMD_DESC_DW0_VAL 0x500012 64 65 /* Address offset for virtual queue registers */ 66 #define CMD_Q_STATUS_INCR 0x1000 67 68 /* Bit masks */ 69 #define CMD_CONFIG_REQID 0 70 #define CMD_TIMEOUT_DISABLE 0 71 #define CMD_CLK_DYN_GATING_DIS 0 72 #define CMD_CLK_SW_GATE_MODE 0 73 #define CMD_CLK_GATE_CTL 0 74 #define CMD_QUEUE_PRIO GENMASK(2, 1) 75 #define CMD_CONFIG_VHB_EN BIT(0) 76 #define CMD_CLK_DYN_GATING_EN BIT(0) 77 #define CMD_CLK_HW_GATE_MODE BIT(0) 78 #define CMD_CLK_GATE_ON_DELAY BIT(12) 79 #define CMD_CLK_GATE_OFF_DELAY BIT(12) 80 81 #define CMD_CLK_GATE_CONFIG (CMD_CLK_GATE_CTL | \ 82 CMD_CLK_HW_GATE_MODE | \ 83 CMD_CLK_GATE_ON_DELAY | \ 84 CMD_CLK_DYN_GATING_EN | \ 85 CMD_CLK_GATE_OFF_DELAY) 86 87 #define CMD_Q_LEN 32 88 #define CMD_Q_RUN BIT(0) 89 #define CMD_Q_HALT BIT(1) 90 #define CMD_Q_MEM_LOCATION BIT(2) 91 #define CMD_Q_SIZE_MASK GENMASK(4, 0) 92 #define CMD_Q_SIZE GENMASK(7, 3) 93 #define CMD_Q_SHIFT GENMASK(1, 0) 94 #define QUEUE_SIZE_VAL ((ffs(CMD_Q_LEN) - 2) & \ 95 CMD_Q_SIZE_MASK) 96 #define Q_PTR_MASK (2 << (QUEUE_SIZE_VAL + 5) - 1) 97 #define Q_DESC_SIZE sizeof(struct ptdma_desc) 98 #define Q_SIZE(n) (CMD_Q_LEN * (n)) 99 100 #define INT_COMPLETION BIT(0) 101 #define INT_ERROR BIT(1) 102 #define INT_QUEUE_STOPPED BIT(2) 103 #define INT_EMPTY_QUEUE BIT(3) 104 #define SUPPORTED_INTERRUPTS (INT_COMPLETION | INT_ERROR) 105 106 /****** Local Storage Block ******/ 107 #define LSB_START 0 108 #define LSB_END 127 109 #define LSB_COUNT (LSB_END - LSB_START + 1) 110 111 #define PT_DMAPOOL_MAX_SIZE 64 112 #define PT_DMAPOOL_ALIGN BIT(5) 113 114 #define PT_PASSTHRU_BLOCKSIZE 512 115 116 struct pt_device; 117 118 struct pt_tasklet_data { 119 struct completion completion; 120 struct pt_cmd *cmd; 121 }; 122 123 /* 124 * struct pt_passthru_engine - pass-through operation 125 * without performing DMA mapping 126 * @mask: mask to be applied to data 127 * @mask_len: length in bytes of mask 128 * @src_dma: data to be used for this operation 129 * @dst_dma: data produced by this operation 130 * @src_len: length in bytes of data used for this operation 131 * 132 * Variables required to be set when calling pt_enqueue_cmd(): 133 * - bit_mod, byte_swap, src, dst, src_len 134 * - mask, mask_len if bit_mod is not PT_PASSTHRU_BITWISE_NOOP 135 */ 136 struct pt_passthru_engine { 137 dma_addr_t mask; 138 u32 mask_len; /* In bytes */ 139 140 dma_addr_t src_dma, dst_dma; 141 u64 src_len; /* In bytes */ 142 }; 143 144 /* 145 * struct pt_cmd - PTDMA operation request 146 * @entry: list element 147 * @work: work element used for callbacks 148 * @pt: PT device to be run on 149 * @ret: operation return code 150 * @flags: cmd processing flags 151 * @engine: PTDMA operation to perform (passthru) 152 * @engine_error: PT engine return code 153 * @passthru: engine specific structures, refer to specific engine struct below 154 * @callback: operation completion callback function 155 * @data: parameter value to be supplied to the callback function 156 * 157 * Variables required to be set when calling pt_enqueue_cmd(): 158 * - engine, callback 159 * - See the operation structures below for what is required for each 160 * operation. 161 */ 162 struct pt_cmd { 163 struct list_head entry; 164 struct work_struct work; 165 struct pt_device *pt; 166 int ret; 167 u32 engine; 168 u32 engine_error; 169 struct pt_passthru_engine passthru; 170 /* Completion callback support */ 171 void (*pt_cmd_callback)(void *data, int err); 172 void *data; 173 }; 174 175 struct pt_dma_desc { 176 struct virt_dma_desc vd; 177 struct pt_device *pt; 178 enum dma_status status; 179 size_t len; 180 bool issued_to_hw; 181 struct pt_cmd pt_cmd; 182 }; 183 184 struct pt_dma_chan { 185 struct virt_dma_chan vc; 186 struct pt_device *pt; 187 u32 id; 188 }; 189 190 struct pt_cmd_queue { 191 struct pt_device *pt; 192 193 /* Queue dma pool */ 194 struct dma_pool *dma_pool; 195 196 /* Queue base address (not necessarily aligned)*/ 197 struct ptdma_desc *qbase; 198 199 /* Aligned queue start address (per requirement) */ 200 spinlock_t q_lock ____cacheline_aligned; 201 unsigned int qidx; 202 203 unsigned int qsize; 204 dma_addr_t qbase_dma; 205 dma_addr_t qdma_tail; 206 207 unsigned int active; 208 unsigned int suspended; 209 210 /* Interrupt flag */ 211 bool int_en; 212 213 /* Register addresses for queue */ 214 void __iomem *reg_control; 215 u32 qcontrol; /* Cached control register */ 216 217 /* Status values from job */ 218 u32 int_status; 219 u32 q_status; 220 u32 q_int_status; 221 u32 cmd_error; 222 /* Queue Statistics */ 223 unsigned long total_pt_ops; 224 } ____cacheline_aligned; 225 226 struct pt_device { 227 struct list_head entry; 228 229 unsigned int ord; 230 char name[MAX_PT_NAME_LEN]; 231 232 struct device *dev; 233 234 /* Bus specific device information */ 235 struct pt_msix *pt_msix; 236 237 struct pt_dev_vdata *dev_vdata; 238 239 unsigned int pt_irq; 240 241 /* I/O area used for device communication */ 242 void __iomem *io_regs; 243 244 spinlock_t cmd_lock ____cacheline_aligned; 245 unsigned int cmd_count; 246 struct list_head cmd; 247 248 /* 249 * The command queue. This represent the queue available on the 250 * PTDMA that are available for processing cmds 251 */ 252 struct pt_cmd_queue cmd_q; 253 254 /* Support for the DMA Engine capabilities */ 255 struct dma_device dma_dev; 256 struct pt_dma_chan *pt_dma_chan; 257 struct kmem_cache *dma_cmd_cache; 258 struct kmem_cache *dma_desc_cache; 259 260 wait_queue_head_t lsb_queue; 261 262 /* Device Statistics */ 263 unsigned long total_interrupts; 264 265 struct pt_tasklet_data tdata; 266 int ver; 267 }; 268 269 /* 270 * descriptor for PTDMA commands 271 * 8 32-bit words: 272 * word 0: function; engine; control bits 273 * word 1: length of source data 274 * word 2: low 32 bits of source pointer 275 * word 3: upper 16 bits of source pointer; source memory type 276 * word 4: low 32 bits of destination pointer 277 * word 5: upper 16 bits of destination pointer; destination memory type 278 * word 6: reserved 32 bits 279 * word 7: reserved 32 bits 280 */ 281 282 #define DWORD0_SOC BIT(0) 283 #define DWORD0_IOC BIT(1) 284 285 struct dword3 { 286 unsigned int src_hi:16; 287 unsigned int src_mem:2; 288 unsigned int lsb_cxt_id:8; 289 unsigned int rsvd1:5; 290 unsigned int fixed:1; 291 }; 292 293 struct dword5 { 294 unsigned int dst_hi:16; 295 unsigned int dst_mem:2; 296 unsigned int rsvd1:13; 297 unsigned int fixed:1; 298 }; 299 300 struct ptdma_desc { 301 u32 dw0; 302 u32 length; 303 u32 src_lo; 304 struct dword3 dw3; 305 u32 dst_lo; 306 struct dword5 dw5; 307 __le32 rsvd1; 308 __le32 rsvd2; 309 }; 310 311 /* Structure to hold PT device data */ 312 struct pt_dev_vdata { 313 const unsigned int bar; 314 }; 315 316 int pt_dmaengine_register(struct pt_device *pt); 317 void pt_dmaengine_unregister(struct pt_device *pt); 318 319 void ptdma_debugfs_setup(struct pt_device *pt); 320 int pt_core_init(struct pt_device *pt); 321 void pt_core_destroy(struct pt_device *pt); 322 323 int pt_core_perform_passthru(struct pt_cmd_queue *cmd_q, 324 struct pt_passthru_engine *pt_engine); 325 326 void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q); 327 void pt_start_queue(struct pt_cmd_queue *cmd_q); 328 void pt_stop_queue(struct pt_cmd_queue *cmd_q); 329 330 static inline void pt_core_disable_queue_interrupts(struct pt_device *pt) 331 { 332 iowrite32(0, pt->cmd_q.reg_control + 0x000C); 333 } 334 335 static inline void pt_core_enable_queue_interrupts(struct pt_device *pt) 336 { 337 iowrite32(SUPPORTED_INTERRUPTS, pt->cmd_q.reg_control + 0x000C); 338 } 339 #endif 340