1 /* 2 * caam descriptor construction helper functions 3 * 4 * Copyright 2008-2012 Freescale Semiconductor, Inc. 5 */ 6 7 #include "desc.h" 8 #include "regs.h" 9 10 #define IMMEDIATE (1 << 23) 11 #define CAAM_CMD_SZ sizeof(u32) 12 #define CAAM_PTR_SZ sizeof(dma_addr_t) 13 #define CAAM_DESC_BYTES_MAX (CAAM_CMD_SZ * MAX_CAAM_DESCSIZE) 14 #define DESC_JOB_IO_LEN (CAAM_CMD_SZ * 5 + CAAM_PTR_SZ * 3) 15 16 #ifdef DEBUG 17 #define PRINT_POS do { printk(KERN_DEBUG "%02d: %s\n", desc_len(desc),\ 18 &__func__[sizeof("append")]); } while (0) 19 #else 20 #define PRINT_POS 21 #endif 22 23 #define SET_OK_NO_PROP_ERRORS (IMMEDIATE | LDST_CLASS_DECO | \ 24 LDST_SRCDST_WORD_DECOCTRL | \ 25 (LDOFF_CHG_SHARE_OK_NO_PROP << \ 26 LDST_OFFSET_SHIFT)) 27 #define DISABLE_AUTO_INFO_FIFO (IMMEDIATE | LDST_CLASS_DECO | \ 28 LDST_SRCDST_WORD_DECOCTRL | \ 29 (LDOFF_DISABLE_AUTO_NFIFO << LDST_OFFSET_SHIFT)) 30 #define ENABLE_AUTO_INFO_FIFO (IMMEDIATE | LDST_CLASS_DECO | \ 31 LDST_SRCDST_WORD_DECOCTRL | \ 32 (LDOFF_ENABLE_AUTO_NFIFO << LDST_OFFSET_SHIFT)) 33 34 extern bool caam_little_end; 35 36 static inline int desc_len(u32 *desc) 37 { 38 return caam32_to_cpu(*desc) & HDR_DESCLEN_MASK; 39 } 40 41 static inline int desc_bytes(void *desc) 42 { 43 return desc_len(desc) * CAAM_CMD_SZ; 44 } 45 46 static inline u32 *desc_end(u32 *desc) 47 { 48 return desc + desc_len(desc); 49 } 50 51 static inline void *sh_desc_pdb(u32 *desc) 52 { 53 return desc + 1; 54 } 55 56 static inline void init_desc(u32 *desc, u32 options) 57 { 58 *desc = cpu_to_caam32((options | HDR_ONE) + 1); 59 } 60 61 static inline void init_sh_desc(u32 *desc, u32 options) 62 { 63 PRINT_POS; 64 init_desc(desc, CMD_SHARED_DESC_HDR | options); 65 } 66 67 static inline void init_sh_desc_pdb(u32 *desc, u32 options, size_t pdb_bytes) 68 { 69 u32 pdb_len = (pdb_bytes + CAAM_CMD_SZ - 1) / CAAM_CMD_SZ; 70 71 init_sh_desc(desc, (((pdb_len + 1) << HDR_START_IDX_SHIFT) + pdb_len) | 72 options); 73 } 74 75 static inline void init_job_desc(u32 *desc, u32 options) 76 { 77 init_desc(desc, CMD_DESC_HDR | options); 78 } 79 80 static inline void append_ptr(u32 *desc, dma_addr_t ptr) 81 { 82 dma_addr_t *offset = (dma_addr_t *)desc_end(desc); 83 84 *offset = cpu_to_caam_dma(ptr); 85 86 (*desc) = cpu_to_caam32(caam32_to_cpu(*desc) + 87 CAAM_PTR_SZ / CAAM_CMD_SZ); 88 } 89 90 static inline void init_job_desc_shared(u32 *desc, dma_addr_t ptr, int len, 91 u32 options) 92 { 93 PRINT_POS; 94 init_job_desc(desc, HDR_SHARED | options | 95 (len << HDR_START_IDX_SHIFT)); 96 append_ptr(desc, ptr); 97 } 98 99 static inline void append_data(u32 *desc, void *data, int len) 100 { 101 u32 *offset = desc_end(desc); 102 103 if (len) /* avoid sparse warning: memcpy with byte count of 0 */ 104 memcpy(offset, data, len); 105 106 (*desc) = cpu_to_caam32(caam32_to_cpu(*desc) + 107 (len + CAAM_CMD_SZ - 1) / CAAM_CMD_SZ); 108 } 109 110 static inline void append_cmd(u32 *desc, u32 command) 111 { 112 u32 *cmd = desc_end(desc); 113 114 *cmd = cpu_to_caam32(command); 115 116 (*desc) = cpu_to_caam32(caam32_to_cpu(*desc) + 1); 117 } 118 119 #define append_u32 append_cmd 120 121 static inline void append_u64(u32 *desc, u64 data) 122 { 123 u32 *offset = desc_end(desc); 124 125 /* Only 32-bit alignment is guaranteed in descriptor buffer */ 126 if (caam_little_end) { 127 *offset = cpu_to_caam32(lower_32_bits(data)); 128 *(++offset) = cpu_to_caam32(upper_32_bits(data)); 129 } else { 130 *offset = cpu_to_caam32(upper_32_bits(data)); 131 *(++offset) = cpu_to_caam32(lower_32_bits(data)); 132 } 133 134 (*desc) = cpu_to_caam32(caam32_to_cpu(*desc) + 2); 135 } 136 137 /* Write command without affecting header, and return pointer to next word */ 138 static inline u32 *write_cmd(u32 *desc, u32 command) 139 { 140 *desc = cpu_to_caam32(command); 141 142 return desc + 1; 143 } 144 145 static inline void append_cmd_ptr(u32 *desc, dma_addr_t ptr, int len, 146 u32 command) 147 { 148 append_cmd(desc, command | len); 149 append_ptr(desc, ptr); 150 } 151 152 /* Write length after pointer, rather than inside command */ 153 static inline void append_cmd_ptr_extlen(u32 *desc, dma_addr_t ptr, 154 unsigned int len, u32 command) 155 { 156 append_cmd(desc, command); 157 if (!(command & (SQIN_RTO | SQIN_PRE))) 158 append_ptr(desc, ptr); 159 append_cmd(desc, len); 160 } 161 162 static inline void append_cmd_data(u32 *desc, void *data, int len, 163 u32 command) 164 { 165 append_cmd(desc, command | IMMEDIATE | len); 166 append_data(desc, data, len); 167 } 168 169 #define APPEND_CMD_RET(cmd, op) \ 170 static inline u32 *append_##cmd(u32 *desc, u32 options) \ 171 { \ 172 u32 *cmd = desc_end(desc); \ 173 PRINT_POS; \ 174 append_cmd(desc, CMD_##op | options); \ 175 return cmd; \ 176 } 177 APPEND_CMD_RET(jump, JUMP) 178 APPEND_CMD_RET(move, MOVE) 179 180 static inline void set_jump_tgt_here(u32 *desc, u32 *jump_cmd) 181 { 182 *jump_cmd = cpu_to_caam32(caam32_to_cpu(*jump_cmd) | 183 (desc_len(desc) - (jump_cmd - desc))); 184 } 185 186 static inline void set_move_tgt_here(u32 *desc, u32 *move_cmd) 187 { 188 u32 val = caam32_to_cpu(*move_cmd); 189 190 val &= ~MOVE_OFFSET_MASK; 191 val |= (desc_len(desc) << (MOVE_OFFSET_SHIFT + 2)) & MOVE_OFFSET_MASK; 192 *move_cmd = cpu_to_caam32(val); 193 } 194 195 #define APPEND_CMD(cmd, op) \ 196 static inline void append_##cmd(u32 *desc, u32 options) \ 197 { \ 198 PRINT_POS; \ 199 append_cmd(desc, CMD_##op | options); \ 200 } 201 APPEND_CMD(operation, OPERATION) 202 203 #define APPEND_CMD_LEN(cmd, op) \ 204 static inline void append_##cmd(u32 *desc, unsigned int len, u32 options) \ 205 { \ 206 PRINT_POS; \ 207 append_cmd(desc, CMD_##op | len | options); \ 208 } 209 210 APPEND_CMD_LEN(seq_load, SEQ_LOAD) 211 APPEND_CMD_LEN(seq_store, SEQ_STORE) 212 APPEND_CMD_LEN(seq_fifo_load, SEQ_FIFO_LOAD) 213 APPEND_CMD_LEN(seq_fifo_store, SEQ_FIFO_STORE) 214 215 #define APPEND_CMD_PTR(cmd, op) \ 216 static inline void append_##cmd(u32 *desc, dma_addr_t ptr, unsigned int len, \ 217 u32 options) \ 218 { \ 219 PRINT_POS; \ 220 append_cmd_ptr(desc, ptr, len, CMD_##op | options); \ 221 } 222 APPEND_CMD_PTR(key, KEY) 223 APPEND_CMD_PTR(load, LOAD) 224 APPEND_CMD_PTR(fifo_load, FIFO_LOAD) 225 APPEND_CMD_PTR(fifo_store, FIFO_STORE) 226 227 static inline void append_store(u32 *desc, dma_addr_t ptr, unsigned int len, 228 u32 options) 229 { 230 u32 cmd_src; 231 232 cmd_src = options & LDST_SRCDST_MASK; 233 234 append_cmd(desc, CMD_STORE | options | len); 235 236 /* The following options do not require pointer */ 237 if (!(cmd_src == LDST_SRCDST_WORD_DESCBUF_SHARED || 238 cmd_src == LDST_SRCDST_WORD_DESCBUF_JOB || 239 cmd_src == LDST_SRCDST_WORD_DESCBUF_JOB_WE || 240 cmd_src == LDST_SRCDST_WORD_DESCBUF_SHARED_WE)) 241 append_ptr(desc, ptr); 242 } 243 244 #define APPEND_SEQ_PTR_INTLEN(cmd, op) \ 245 static inline void append_seq_##cmd##_ptr_intlen(u32 *desc, dma_addr_t ptr, \ 246 unsigned int len, \ 247 u32 options) \ 248 { \ 249 PRINT_POS; \ 250 if (options & (SQIN_RTO | SQIN_PRE)) \ 251 append_cmd(desc, CMD_SEQ_##op##_PTR | len | options); \ 252 else \ 253 append_cmd_ptr(desc, ptr, len, CMD_SEQ_##op##_PTR | options); \ 254 } 255 APPEND_SEQ_PTR_INTLEN(in, IN) 256 APPEND_SEQ_PTR_INTLEN(out, OUT) 257 258 #define APPEND_CMD_PTR_TO_IMM(cmd, op) \ 259 static inline void append_##cmd##_as_imm(u32 *desc, void *data, \ 260 unsigned int len, u32 options) \ 261 { \ 262 PRINT_POS; \ 263 append_cmd_data(desc, data, len, CMD_##op | options); \ 264 } 265 APPEND_CMD_PTR_TO_IMM(load, LOAD); 266 APPEND_CMD_PTR_TO_IMM(fifo_load, FIFO_LOAD); 267 268 #define APPEND_CMD_PTR_EXTLEN(cmd, op) \ 269 static inline void append_##cmd##_extlen(u32 *desc, dma_addr_t ptr, \ 270 unsigned int len, u32 options) \ 271 { \ 272 PRINT_POS; \ 273 append_cmd_ptr_extlen(desc, ptr, len, CMD_##op | SQIN_EXT | options); \ 274 } 275 APPEND_CMD_PTR_EXTLEN(seq_in_ptr, SEQ_IN_PTR) 276 APPEND_CMD_PTR_EXTLEN(seq_out_ptr, SEQ_OUT_PTR) 277 278 /* 279 * Determine whether to store length internally or externally depending on 280 * the size of its type 281 */ 282 #define APPEND_CMD_PTR_LEN(cmd, op, type) \ 283 static inline void append_##cmd(u32 *desc, dma_addr_t ptr, \ 284 type len, u32 options) \ 285 { \ 286 PRINT_POS; \ 287 if (sizeof(type) > sizeof(u16)) \ 288 append_##cmd##_extlen(desc, ptr, len, options); \ 289 else \ 290 append_##cmd##_intlen(desc, ptr, len, options); \ 291 } 292 APPEND_CMD_PTR_LEN(seq_in_ptr, SEQ_IN_PTR, u32) 293 APPEND_CMD_PTR_LEN(seq_out_ptr, SEQ_OUT_PTR, u32) 294 295 /* 296 * 2nd variant for commands whose specified immediate length differs 297 * from length of immediate data provided, e.g., split keys 298 */ 299 #define APPEND_CMD_PTR_TO_IMM2(cmd, op) \ 300 static inline void append_##cmd##_as_imm(u32 *desc, void *data, \ 301 unsigned int data_len, \ 302 unsigned int len, u32 options) \ 303 { \ 304 PRINT_POS; \ 305 append_cmd(desc, CMD_##op | IMMEDIATE | len | options); \ 306 append_data(desc, data, data_len); \ 307 } 308 APPEND_CMD_PTR_TO_IMM2(key, KEY); 309 310 #define APPEND_CMD_RAW_IMM(cmd, op, type) \ 311 static inline void append_##cmd##_imm_##type(u32 *desc, type immediate, \ 312 u32 options) \ 313 { \ 314 PRINT_POS; \ 315 append_cmd(desc, CMD_##op | IMMEDIATE | options | sizeof(type)); \ 316 append_cmd(desc, immediate); \ 317 } 318 APPEND_CMD_RAW_IMM(load, LOAD, u32); 319 320 /* 321 * Append math command. Only the last part of destination and source need to 322 * be specified 323 */ 324 #define APPEND_MATH(op, desc, dest, src_0, src_1, len) \ 325 append_cmd(desc, CMD_MATH | MATH_FUN_##op | MATH_DEST_##dest | \ 326 MATH_SRC0_##src_0 | MATH_SRC1_##src_1 | (u32)len); 327 328 #define append_math_add(desc, dest, src0, src1, len) \ 329 APPEND_MATH(ADD, desc, dest, src0, src1, len) 330 #define append_math_sub(desc, dest, src0, src1, len) \ 331 APPEND_MATH(SUB, desc, dest, src0, src1, len) 332 #define append_math_add_c(desc, dest, src0, src1, len) \ 333 APPEND_MATH(ADDC, desc, dest, src0, src1, len) 334 #define append_math_sub_b(desc, dest, src0, src1, len) \ 335 APPEND_MATH(SUBB, desc, dest, src0, src1, len) 336 #define append_math_and(desc, dest, src0, src1, len) \ 337 APPEND_MATH(AND, desc, dest, src0, src1, len) 338 #define append_math_or(desc, dest, src0, src1, len) \ 339 APPEND_MATH(OR, desc, dest, src0, src1, len) 340 #define append_math_xor(desc, dest, src0, src1, len) \ 341 APPEND_MATH(XOR, desc, dest, src0, src1, len) 342 #define append_math_lshift(desc, dest, src0, src1, len) \ 343 APPEND_MATH(LSHIFT, desc, dest, src0, src1, len) 344 #define append_math_rshift(desc, dest, src0, src1, len) \ 345 APPEND_MATH(RSHIFT, desc, dest, src0, src1, len) 346 #define append_math_ldshift(desc, dest, src0, src1, len) \ 347 APPEND_MATH(SHLD, desc, dest, src0, src1, len) 348 349 /* Exactly one source is IMM. Data is passed in as u32 value */ 350 #define APPEND_MATH_IMM_u32(op, desc, dest, src_0, src_1, data) \ 351 do { \ 352 APPEND_MATH(op, desc, dest, src_0, src_1, CAAM_CMD_SZ); \ 353 append_cmd(desc, data); \ 354 } while (0) 355 356 #define append_math_add_imm_u32(desc, dest, src0, src1, data) \ 357 APPEND_MATH_IMM_u32(ADD, desc, dest, src0, src1, data) 358 #define append_math_sub_imm_u32(desc, dest, src0, src1, data) \ 359 APPEND_MATH_IMM_u32(SUB, desc, dest, src0, src1, data) 360 #define append_math_add_c_imm_u32(desc, dest, src0, src1, data) \ 361 APPEND_MATH_IMM_u32(ADDC, desc, dest, src0, src1, data) 362 #define append_math_sub_b_imm_u32(desc, dest, src0, src1, data) \ 363 APPEND_MATH_IMM_u32(SUBB, desc, dest, src0, src1, data) 364 #define append_math_and_imm_u32(desc, dest, src0, src1, data) \ 365 APPEND_MATH_IMM_u32(AND, desc, dest, src0, src1, data) 366 #define append_math_or_imm_u32(desc, dest, src0, src1, data) \ 367 APPEND_MATH_IMM_u32(OR, desc, dest, src0, src1, data) 368 #define append_math_xor_imm_u32(desc, dest, src0, src1, data) \ 369 APPEND_MATH_IMM_u32(XOR, desc, dest, src0, src1, data) 370 #define append_math_lshift_imm_u32(desc, dest, src0, src1, data) \ 371 APPEND_MATH_IMM_u32(LSHIFT, desc, dest, src0, src1, data) 372 #define append_math_rshift_imm_u32(desc, dest, src0, src1, data) \ 373 APPEND_MATH_IMM_u32(RSHIFT, desc, dest, src0, src1, data) 374 375 /* Exactly one source is IMM. Data is passed in as u64 value */ 376 #define APPEND_MATH_IMM_u64(op, desc, dest, src_0, src_1, data) \ 377 do { \ 378 u32 upper = (data >> 16) >> 16; \ 379 APPEND_MATH(op, desc, dest, src_0, src_1, CAAM_CMD_SZ * 2 | \ 380 (upper ? 0 : MATH_IFB)); \ 381 if (upper) \ 382 append_u64(desc, data); \ 383 else \ 384 append_u32(desc, lower_32_bits(data)); \ 385 } while (0) 386 387 #define append_math_add_imm_u64(desc, dest, src0, src1, data) \ 388 APPEND_MATH_IMM_u64(ADD, desc, dest, src0, src1, data) 389 #define append_math_sub_imm_u64(desc, dest, src0, src1, data) \ 390 APPEND_MATH_IMM_u64(SUB, desc, dest, src0, src1, data) 391 #define append_math_add_c_imm_u64(desc, dest, src0, src1, data) \ 392 APPEND_MATH_IMM_u64(ADDC, desc, dest, src0, src1, data) 393 #define append_math_sub_b_imm_u64(desc, dest, src0, src1, data) \ 394 APPEND_MATH_IMM_u64(SUBB, desc, dest, src0, src1, data) 395 #define append_math_and_imm_u64(desc, dest, src0, src1, data) \ 396 APPEND_MATH_IMM_u64(AND, desc, dest, src0, src1, data) 397 #define append_math_or_imm_u64(desc, dest, src0, src1, data) \ 398 APPEND_MATH_IMM_u64(OR, desc, dest, src0, src1, data) 399 #define append_math_xor_imm_u64(desc, dest, src0, src1, data) \ 400 APPEND_MATH_IMM_u64(XOR, desc, dest, src0, src1, data) 401 #define append_math_lshift_imm_u64(desc, dest, src0, src1, data) \ 402 APPEND_MATH_IMM_u64(LSHIFT, desc, dest, src0, src1, data) 403 #define append_math_rshift_imm_u64(desc, dest, src0, src1, data) \ 404 APPEND_MATH_IMM_u64(RSHIFT, desc, dest, src0, src1, data) 405