1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) 2 /* 3 * Wave5 series multi-standard codec IP - wave5 backend logic 4 * 5 * Copyright (C) 2021-2023 CHIPS&MEDIA INC 6 */ 7 8 #include <linux/iopoll.h> 9 #include <linux/bitfield.h> 10 #include "wave5-vpu.h" 11 #include "wave5.h" 12 #include "wave5-regdefine.h" 13 14 #define FIO_TIMEOUT 10000000 15 #define FIO_CTRL_READY BIT(31) 16 #define FIO_CTRL_WRITE BIT(16) 17 #define VPU_BUSY_CHECK_TIMEOUT 10000000 18 #define QUEUE_REPORT_MASK 0xffff 19 20 /* Encoder support fields */ 21 #define FEATURE_HEVC10BIT_ENC BIT(3) 22 #define FEATURE_AVC10BIT_ENC BIT(11) 23 #define FEATURE_AVC_ENCODER BIT(1) 24 #define FEATURE_HEVC_ENCODER BIT(0) 25 26 /* Decoder support fields */ 27 #define FEATURE_AVC_DECODER BIT(3) 28 #define FEATURE_HEVC_DECODER BIT(2) 29 30 #define FEATURE_BACKBONE BIT(16) 31 #define FEATURE_VCORE_BACKBONE BIT(22) 32 #define FEATURE_VCPU_BACKBONE BIT(28) 33 34 #define REMAP_CTRL_MAX_SIZE_BITS ((W5_REMAP_MAX_SIZE >> 12) & 0x1ff) 35 #define REMAP_CTRL_REGISTER_VALUE(index) ( \ 36 (BIT(31) | (index << 12) | BIT(11) | REMAP_CTRL_MAX_SIZE_BITS) \ 37 ) 38 39 #define FASTIO_ADDRESS_MASK GENMASK(15, 0) 40 #define SEQ_PARAM_PROFILE_MASK GENMASK(30, 24) 41 42 static void _wave5_print_reg_err(struct vpu_device *vpu_dev, u32 reg_fail_reason, 43 const char *func); 44 #define PRINT_REG_ERR(dev, reason) _wave5_print_reg_err((dev), (reason), __func__) 45 46 static inline const char *cmd_to_str(int cmd, bool is_dec) 47 { 48 switch (cmd) { 49 case W5_INIT_VPU: 50 return "W5_INIT_VPU"; 51 case W5_WAKEUP_VPU: 52 return "W5_WAKEUP_VPU"; 53 case W5_SLEEP_VPU: 54 return "W5_SLEEP_VPU"; 55 case W5_CREATE_INSTANCE: 56 return "W5_CREATE_INSTANCE"; 57 case W5_FLUSH_INSTANCE: 58 return "W5_FLUSH_INSTANCE"; 59 case W5_DESTROY_INSTANCE: 60 return "W5_DESTROY_INSTANCE"; 61 case W5_INIT_SEQ: 62 return "W5_INIT_SEQ"; 63 case W5_SET_FB: 64 return "W5_SET_FB"; 65 case W5_DEC_ENC_PIC: 66 if (is_dec) 67 return "W5_DEC_PIC"; 68 return "W5_ENC_PIC"; 69 case W5_ENC_SET_PARAM: 70 return "W5_ENC_SET_PARAM"; 71 case W5_QUERY: 72 return "W5_QUERY"; 73 case W5_UPDATE_BS: 74 return "W5_UPDATE_BS"; 75 case W5_MAX_VPU_COMD: 76 return "W5_MAX_VPU_COMD"; 77 default: 78 return "UNKNOWN"; 79 } 80 } 81 82 static void _wave5_print_reg_err(struct vpu_device *vpu_dev, u32 reg_fail_reason, 83 const char *func) 84 { 85 struct device *dev = vpu_dev->dev; 86 u32 reg_val; 87 88 switch (reg_fail_reason) { 89 case WAVE5_SYSERR_QUEUEING_FAIL: 90 reg_val = vpu_read_reg(vpu_dev, W5_RET_QUEUE_FAIL_REASON); 91 dev_dbg(dev, "%s: queueing failure: 0x%x\n", func, reg_val); 92 break; 93 case WAVE5_SYSERR_RESULT_NOT_READY: 94 dev_err(dev, "%s: result not ready: 0x%x\n", func, reg_fail_reason); 95 break; 96 case WAVE5_SYSERR_ACCESS_VIOLATION_HW: 97 dev_err(dev, "%s: access violation: 0x%x\n", func, reg_fail_reason); 98 break; 99 case WAVE5_SYSERR_WATCHDOG_TIMEOUT: 100 dev_err(dev, "%s: watchdog timeout: 0x%x\n", func, reg_fail_reason); 101 break; 102 case WAVE5_SYSERR_BUS_ERROR: 103 dev_err(dev, "%s: bus error: 0x%x\n", func, reg_fail_reason); 104 break; 105 case WAVE5_SYSERR_DOUBLE_FAULT: 106 dev_err(dev, "%s: double fault: 0x%x\n", func, reg_fail_reason); 107 break; 108 case WAVE5_SYSERR_VPU_STILL_RUNNING: 109 dev_err(dev, "%s: still running: 0x%x\n", func, reg_fail_reason); 110 break; 111 case WAVE5_SYSERR_VLC_BUF_FULL: 112 dev_err(dev, "%s: vlc buf full: 0x%x\n", func, reg_fail_reason); 113 break; 114 default: 115 dev_err(dev, "%s: failure:: 0x%x\n", func, reg_fail_reason); 116 break; 117 } 118 } 119 120 static int wave5_wait_fio_readl(struct vpu_device *vpu_dev, u32 addr, u32 val) 121 { 122 u32 ctrl; 123 int ret; 124 125 ctrl = addr & 0xffff; 126 wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_CTRL_ADDR, ctrl); 127 ret = read_poll_timeout(wave5_vdi_read_register, ctrl, ctrl & FIO_CTRL_READY, 128 0, FIO_TIMEOUT, false, vpu_dev, W5_VPU_FIO_CTRL_ADDR); 129 if (ret) 130 return ret; 131 132 if (wave5_vdi_read_register(vpu_dev, W5_VPU_FIO_DATA) != val) 133 return -ETIMEDOUT; 134 135 return 0; 136 } 137 138 static void wave5_fio_writel(struct vpu_device *vpu_dev, unsigned int addr, unsigned int data) 139 { 140 int ret; 141 unsigned int ctrl; 142 143 wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_DATA, data); 144 ctrl = FIELD_GET(FASTIO_ADDRESS_MASK, addr); 145 ctrl |= FIO_CTRL_WRITE; 146 wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_CTRL_ADDR, ctrl); 147 ret = read_poll_timeout(wave5_vdi_read_register, ctrl, ctrl & FIO_CTRL_READY, 0, 148 FIO_TIMEOUT, false, vpu_dev, W5_VPU_FIO_CTRL_ADDR); 149 if (ret) 150 dev_dbg_ratelimited(vpu_dev->dev, "FIO write timeout: addr=0x%x data=%x\n", 151 ctrl, data); 152 } 153 154 static int wave5_wait_bus_busy(struct vpu_device *vpu_dev, unsigned int addr) 155 { 156 u32 gdi_status_check_value = 0x3f; 157 158 if (vpu_dev->product_code == WAVE521C_CODE || 159 vpu_dev->product_code == WAVE521_CODE || 160 vpu_dev->product_code == WAVE521E1_CODE) 161 gdi_status_check_value = 0x00ff1f3f; 162 163 return wave5_wait_fio_readl(vpu_dev, addr, gdi_status_check_value); 164 } 165 166 static int wave5_wait_vpu_busy(struct vpu_device *vpu_dev, unsigned int addr) 167 { 168 u32 data; 169 170 return read_poll_timeout(wave5_vdi_read_register, data, data == 0, 171 0, VPU_BUSY_CHECK_TIMEOUT, false, vpu_dev, addr); 172 } 173 174 static int wave5_wait_vcpu_bus_busy(struct vpu_device *vpu_dev, unsigned int addr) 175 { 176 return wave5_wait_fio_readl(vpu_dev, addr, 0); 177 } 178 179 bool wave5_vpu_is_init(struct vpu_device *vpu_dev) 180 { 181 return vpu_read_reg(vpu_dev, W5_VCPU_CUR_PC) != 0; 182 } 183 184 unsigned int wave5_vpu_get_product_id(struct vpu_device *vpu_dev) 185 { 186 u32 val = vpu_read_reg(vpu_dev, W5_PRODUCT_NUMBER); 187 188 switch (val) { 189 case WAVE521C_CODE: 190 return PRODUCT_ID_521; 191 case WAVE521_CODE: 192 case WAVE521C_DUAL_CODE: 193 case WAVE521E1_CODE: 194 case WAVE511_CODE: 195 case WAVE517_CODE: 196 case WAVE537_CODE: 197 dev_err(vpu_dev->dev, "Unsupported product id (%x)\n", val); 198 break; 199 default: 200 dev_err(vpu_dev->dev, "Invalid product id (%x)\n", val); 201 break; 202 } 203 204 return PRODUCT_ID_NONE; 205 } 206 207 static void wave5_bit_issue_command(struct vpu_device *vpu_dev, struct vpu_instance *inst, u32 cmd) 208 { 209 u32 instance_index; 210 u32 codec_mode; 211 212 if (inst) { 213 instance_index = inst->id; 214 codec_mode = inst->std; 215 216 vpu_write_reg(vpu_dev, W5_CMD_INSTANCE_INFO, (codec_mode << 16) | 217 (instance_index & 0xffff)); 218 vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1); 219 } 220 221 vpu_write_reg(vpu_dev, W5_COMMAND, cmd); 222 223 if (inst) { 224 dev_dbg(vpu_dev->dev, "%s: cmd=0x%x (%s)\n", __func__, cmd, 225 cmd_to_str(cmd, inst->type == VPU_INST_TYPE_DEC)); 226 } else { 227 dev_dbg(vpu_dev->dev, "%s: cmd=0x%x\n", __func__, cmd); 228 } 229 230 vpu_write_reg(vpu_dev, W5_VPU_HOST_INT_REQ, 1); 231 } 232 233 static int wave5_vpu_firmware_command_queue_error_check(struct vpu_device *dev, u32 *fail_res) 234 { 235 u32 reason = 0; 236 237 /* Check if we were able to add a command into the VCPU QUEUE */ 238 if (!vpu_read_reg(dev, W5_RET_SUCCESS)) { 239 reason = vpu_read_reg(dev, W5_RET_FAIL_REASON); 240 PRINT_REG_ERR(dev, reason); 241 242 /* 243 * The fail_res argument will be either NULL or 0. 244 * If the fail_res argument is NULL, then just return -EIO. 245 * Otherwise, assign the reason to fail_res, so that the 246 * calling function can use it. 247 */ 248 if (fail_res) 249 *fail_res = reason; 250 else 251 return -EIO; 252 253 if (reason == WAVE5_SYSERR_VPU_STILL_RUNNING) 254 return -EBUSY; 255 } 256 return 0; 257 } 258 259 static int send_firmware_command(struct vpu_instance *inst, u32 cmd, bool check_success, 260 u32 *queue_status, u32 *fail_result) 261 { 262 int ret; 263 264 wave5_bit_issue_command(inst->dev, inst, cmd); 265 ret = wave5_wait_vpu_busy(inst->dev, W5_VPU_BUSY_STATUS); 266 if (ret) { 267 dev_warn(inst->dev->dev, "%s: command: '%s', timed out\n", __func__, 268 cmd_to_str(cmd, inst->type == VPU_INST_TYPE_DEC)); 269 return -ETIMEDOUT; 270 } 271 272 if (queue_status) 273 *queue_status = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS); 274 275 /* In some cases we want to send multiple commands before checking 276 * whether they are queued properly 277 */ 278 if (!check_success) 279 return 0; 280 281 return wave5_vpu_firmware_command_queue_error_check(inst->dev, fail_result); 282 } 283 284 static int wave5_send_query(struct vpu_device *vpu_dev, struct vpu_instance *inst, 285 enum query_opt query_opt) 286 { 287 int ret; 288 289 vpu_write_reg(vpu_dev, W5_QUERY_OPTION, query_opt); 290 vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1); 291 wave5_bit_issue_command(vpu_dev, inst, W5_QUERY); 292 293 ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS); 294 if (ret) { 295 dev_warn(vpu_dev->dev, "command: 'W5_QUERY', timed out opt=0x%x\n", query_opt); 296 return ret; 297 } 298 299 return wave5_vpu_firmware_command_queue_error_check(vpu_dev, NULL); 300 } 301 302 static int setup_wave5_properties(struct device *dev) 303 { 304 struct vpu_device *vpu_dev = dev_get_drvdata(dev); 305 struct vpu_attr *p_attr = &vpu_dev->attr; 306 u32 reg_val; 307 u8 *str; 308 int ret; 309 u32 hw_config_def0, hw_config_def1, hw_config_feature; 310 311 ret = wave5_send_query(vpu_dev, NULL, GET_VPU_INFO); 312 if (ret) 313 return ret; 314 315 reg_val = vpu_read_reg(vpu_dev, W5_RET_PRODUCT_NAME); 316 str = (u8 *)®_val; 317 p_attr->product_name[0] = str[3]; 318 p_attr->product_name[1] = str[2]; 319 p_attr->product_name[2] = str[1]; 320 p_attr->product_name[3] = str[0]; 321 p_attr->product_name[4] = 0; 322 323 p_attr->product_id = wave5_vpu_get_product_id(vpu_dev); 324 p_attr->product_version = vpu_read_reg(vpu_dev, W5_RET_PRODUCT_VERSION); 325 p_attr->fw_version = vpu_read_reg(vpu_dev, W5_RET_FW_VERSION); 326 p_attr->customer_id = vpu_read_reg(vpu_dev, W5_RET_CUSTOMER_ID); 327 hw_config_def0 = vpu_read_reg(vpu_dev, W5_RET_STD_DEF0); 328 hw_config_def1 = vpu_read_reg(vpu_dev, W5_RET_STD_DEF1); 329 hw_config_feature = vpu_read_reg(vpu_dev, W5_RET_CONF_FEATURE); 330 331 p_attr->support_hevc10bit_enc = FIELD_GET(FEATURE_HEVC10BIT_ENC, hw_config_feature); 332 p_attr->support_avc10bit_enc = FIELD_GET(FEATURE_AVC10BIT_ENC, hw_config_feature); 333 334 p_attr->support_decoders = FIELD_GET(FEATURE_AVC_DECODER, hw_config_def1) << STD_AVC; 335 p_attr->support_decoders |= FIELD_GET(FEATURE_HEVC_DECODER, hw_config_def1) << STD_HEVC; 336 p_attr->support_encoders = FIELD_GET(FEATURE_AVC_ENCODER, hw_config_def1) << STD_AVC; 337 p_attr->support_encoders |= FIELD_GET(FEATURE_HEVC_ENCODER, hw_config_def1) << STD_HEVC; 338 339 p_attr->support_backbone = FIELD_GET(FEATURE_BACKBONE, hw_config_def0); 340 p_attr->support_vcpu_backbone = FIELD_GET(FEATURE_VCPU_BACKBONE, hw_config_def0); 341 p_attr->support_vcore_backbone = FIELD_GET(FEATURE_VCORE_BACKBONE, hw_config_def0); 342 343 return 0; 344 } 345 346 int wave5_vpu_get_version(struct vpu_device *vpu_dev, u32 *revision) 347 { 348 u32 reg_val; 349 int ret; 350 351 ret = wave5_send_query(vpu_dev, NULL, GET_VPU_INFO); 352 if (ret) 353 return ret; 354 355 reg_val = vpu_read_reg(vpu_dev, W5_RET_FW_VERSION); 356 if (revision) { 357 *revision = reg_val; 358 return 0; 359 } 360 361 return -EINVAL; 362 } 363 364 static void remap_page(struct vpu_device *vpu_dev, dma_addr_t code_base, u32 index) 365 { 366 vpu_write_reg(vpu_dev, W5_VPU_REMAP_CTRL, REMAP_CTRL_REGISTER_VALUE(index)); 367 vpu_write_reg(vpu_dev, W5_VPU_REMAP_VADDR, index * W5_REMAP_MAX_SIZE); 368 vpu_write_reg(vpu_dev, W5_VPU_REMAP_PADDR, code_base + index * W5_REMAP_MAX_SIZE); 369 } 370 371 int wave5_vpu_init(struct device *dev, u8 *fw, size_t size) 372 { 373 struct vpu_buf *common_vb; 374 dma_addr_t code_base, temp_base; 375 u32 code_size, temp_size; 376 u32 i, reg_val, reason_code; 377 int ret; 378 struct vpu_device *vpu_dev = dev_get_drvdata(dev); 379 380 common_vb = &vpu_dev->common_mem; 381 382 code_base = common_vb->daddr; 383 /* ALIGN TO 4KB */ 384 code_size = (WAVE5_MAX_CODE_BUF_SIZE & ~0xfff); 385 if (code_size < size * 2) 386 return -EINVAL; 387 388 temp_base = common_vb->daddr + WAVE5_TEMPBUF_OFFSET; 389 temp_size = WAVE5_TEMPBUF_SIZE; 390 391 ret = wave5_vdi_write_memory(vpu_dev, common_vb, 0, fw, size); 392 if (ret < 0) { 393 dev_err(vpu_dev->dev, "VPU init, Writing firmware to common buffer, fail: %d\n", 394 ret); 395 return ret; 396 } 397 398 vpu_write_reg(vpu_dev, W5_PO_CONF, 0); 399 400 /* clear registers */ 401 402 for (i = W5_CMD_REG_BASE; i < W5_CMD_REG_END; i += 4) 403 vpu_write_reg(vpu_dev, i, 0x00); 404 405 remap_page(vpu_dev, code_base, W5_REMAP_INDEX0); 406 remap_page(vpu_dev, code_base, W5_REMAP_INDEX1); 407 408 vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base); 409 vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size); 410 vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0); 411 vpu_write_reg(vpu_dev, W5_ADDR_TEMP_BASE, temp_base); 412 vpu_write_reg(vpu_dev, W5_TEMP_SIZE, temp_size); 413 414 /* These register must be reset explicitly */ 415 vpu_write_reg(vpu_dev, W5_HW_OPTION, 0); 416 wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0); 417 wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0); 418 vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0); 419 420 /* Encoder interrupt */ 421 reg_val = BIT(INT_WAVE5_ENC_SET_PARAM); 422 reg_val |= BIT(INT_WAVE5_ENC_PIC); 423 reg_val |= BIT(INT_WAVE5_BSBUF_FULL); 424 /* Decoder interrupt */ 425 reg_val |= BIT(INT_WAVE5_INIT_SEQ); 426 reg_val |= BIT(INT_WAVE5_DEC_PIC); 427 reg_val |= BIT(INT_WAVE5_BSBUF_EMPTY); 428 vpu_write_reg(vpu_dev, W5_VPU_VINT_ENABLE, reg_val); 429 430 reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0); 431 if (FIELD_GET(FEATURE_BACKBONE, reg_val)) { 432 reg_val = ((WAVE5_PROC_AXI_ID << 28) | 433 (WAVE5_PRP_AXI_ID << 24) | 434 (WAVE5_FBD_Y_AXI_ID << 20) | 435 (WAVE5_FBC_Y_AXI_ID << 16) | 436 (WAVE5_FBD_C_AXI_ID << 12) | 437 (WAVE5_FBC_C_AXI_ID << 8) | 438 (WAVE5_PRI_AXI_ID << 4) | 439 WAVE5_SEC_AXI_ID); 440 wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val); 441 } 442 443 vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1); 444 vpu_write_reg(vpu_dev, W5_COMMAND, W5_INIT_VPU); 445 vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1); 446 ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS); 447 if (ret) { 448 dev_err(vpu_dev->dev, "VPU init(W5_VPU_REMAP_CORE_START) timeout\n"); 449 return ret; 450 } 451 452 ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code); 453 if (ret) 454 return ret; 455 456 return setup_wave5_properties(dev); 457 } 458 459 int wave5_vpu_build_up_dec_param(struct vpu_instance *inst, 460 struct dec_open_param *param) 461 { 462 int ret; 463 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 464 struct vpu_device *vpu_dev = inst->dev; 465 466 p_dec_info->cycle_per_tick = 256; 467 if (vpu_dev->sram_buf.size) { 468 p_dec_info->sec_axi_info.use_bit_enable = 1; 469 p_dec_info->sec_axi_info.use_ip_enable = 1; 470 p_dec_info->sec_axi_info.use_lf_row_enable = 1; 471 } 472 switch (inst->std) { 473 case W_HEVC_DEC: 474 p_dec_info->seq_change_mask = SEQ_CHANGE_ENABLE_ALL_HEVC; 475 break; 476 case W_AVC_DEC: 477 p_dec_info->seq_change_mask = SEQ_CHANGE_ENABLE_ALL_AVC; 478 break; 479 default: 480 return -EINVAL; 481 } 482 483 p_dec_info->vb_work.size = WAVE521DEC_WORKBUF_SIZE; 484 ret = wave5_vdi_allocate_dma_memory(inst->dev, &p_dec_info->vb_work); 485 if (ret) 486 return ret; 487 488 vpu_write_reg(inst->dev, W5_CMD_DEC_VCORE_INFO, 1); 489 490 wave5_vdi_clear_memory(inst->dev, &p_dec_info->vb_work); 491 492 vpu_write_reg(inst->dev, W5_ADDR_WORK_BASE, p_dec_info->vb_work.daddr); 493 vpu_write_reg(inst->dev, W5_WORK_SIZE, p_dec_info->vb_work.size); 494 495 vpu_write_reg(inst->dev, W5_CMD_ADDR_SEC_AXI, vpu_dev->sram_buf.daddr); 496 vpu_write_reg(inst->dev, W5_CMD_SEC_AXI_SIZE, vpu_dev->sram_buf.size); 497 498 vpu_write_reg(inst->dev, W5_CMD_DEC_BS_START_ADDR, p_dec_info->stream_buf_start_addr); 499 vpu_write_reg(inst->dev, W5_CMD_DEC_BS_SIZE, p_dec_info->stream_buf_size); 500 501 /* NOTE: SDMA reads MSB first */ 502 vpu_write_reg(inst->dev, W5_CMD_BS_PARAM, BITSTREAM_ENDIANNESS_BIG_ENDIAN); 503 /* This register must be reset explicitly */ 504 vpu_write_reg(inst->dev, W5_CMD_EXT_ADDR, 0); 505 vpu_write_reg(inst->dev, W5_CMD_NUM_CQ_DEPTH_M1, (COMMAND_QUEUE_DEPTH - 1)); 506 507 ret = send_firmware_command(inst, W5_CREATE_INSTANCE, true, NULL, NULL); 508 if (ret) { 509 wave5_vdi_free_dma_memory(vpu_dev, &p_dec_info->vb_work); 510 return ret; 511 } 512 513 p_dec_info->product_code = vpu_read_reg(inst->dev, W5_PRODUCT_NUMBER); 514 515 return 0; 516 } 517 518 int wave5_vpu_hw_flush_instance(struct vpu_instance *inst) 519 { 520 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 521 u32 instance_queue_count, report_queue_count; 522 u32 reg_val = 0; 523 u32 fail_res = 0; 524 int ret; 525 526 ret = send_firmware_command(inst, W5_FLUSH_INSTANCE, true, ®_val, &fail_res); 527 if (ret) 528 return ret; 529 530 instance_queue_count = (reg_val >> 16) & 0xff; 531 report_queue_count = (reg_val & QUEUE_REPORT_MASK); 532 if (instance_queue_count != 0 || report_queue_count != 0) { 533 dev_warn(inst->dev->dev, 534 "FLUSH_INSTANCE cmd didn't reset the amount of queued commands & reports"); 535 } 536 537 /* reset our local copy of the counts */ 538 p_dec_info->instance_queue_count = 0; 539 p_dec_info->report_queue_count = 0; 540 541 return 0; 542 } 543 544 static u32 get_bitstream_options(struct dec_info *info) 545 { 546 u32 bs_option = BSOPTION_ENABLE_EXPLICIT_END; 547 548 if (info->stream_endflag) 549 bs_option |= BSOPTION_HIGHLIGHT_STREAM_END; 550 return bs_option; 551 } 552 553 int wave5_vpu_dec_init_seq(struct vpu_instance *inst) 554 { 555 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 556 u32 cmd_option = INIT_SEQ_NORMAL; 557 u32 reg_val, fail_res; 558 int ret; 559 560 if (!inst->codec_info) 561 return -EINVAL; 562 563 vpu_write_reg(inst->dev, W5_BS_RD_PTR, p_dec_info->stream_rd_ptr); 564 vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr); 565 566 vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info)); 567 568 vpu_write_reg(inst->dev, W5_COMMAND_OPTION, cmd_option); 569 vpu_write_reg(inst->dev, W5_CMD_DEC_USER_MASK, p_dec_info->user_data_enable); 570 571 ret = send_firmware_command(inst, W5_INIT_SEQ, true, ®_val, &fail_res); 572 if (ret) 573 return ret; 574 575 p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff; 576 p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK); 577 578 dev_dbg(inst->dev->dev, "%s: init seq sent (queue %u : %u)\n", __func__, 579 p_dec_info->instance_queue_count, p_dec_info->report_queue_count); 580 581 return 0; 582 } 583 584 static void wave5_get_dec_seq_result(struct vpu_instance *inst, struct dec_initial_info *info) 585 { 586 u32 reg_val; 587 u32 profile_compatibility_flag; 588 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 589 590 p_dec_info->stream_rd_ptr = wave5_dec_get_rd_ptr(inst); 591 info->rd_ptr = p_dec_info->stream_rd_ptr; 592 593 p_dec_info->frame_display_flag = vpu_read_reg(inst->dev, W5_RET_DEC_DISP_IDC); 594 595 reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_SIZE); 596 info->pic_width = ((reg_val >> 16) & 0xffff); 597 info->pic_height = (reg_val & 0xffff); 598 info->min_frame_buffer_count = vpu_read_reg(inst->dev, W5_RET_DEC_NUM_REQUIRED_FB); 599 600 reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_CROP_LEFT_RIGHT); 601 info->pic_crop_rect.left = (reg_val >> 16) & 0xffff; 602 info->pic_crop_rect.right = reg_val & 0xffff; 603 reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_CROP_TOP_BOTTOM); 604 info->pic_crop_rect.top = (reg_val >> 16) & 0xffff; 605 info->pic_crop_rect.bottom = reg_val & 0xffff; 606 607 reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_COLOR_SAMPLE_INFO); 608 info->luma_bitdepth = reg_val & 0xf; 609 info->chroma_bitdepth = (reg_val >> 4) & 0xf; 610 611 reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_SEQ_PARAM); 612 profile_compatibility_flag = (reg_val >> 12) & 0xff; 613 info->profile = (reg_val >> 24) & 0x1f; 614 615 if (inst->std == W_HEVC_DEC) { 616 /* guessing profile */ 617 if (!info->profile) { 618 if ((profile_compatibility_flag & 0x06) == 0x06) 619 info->profile = HEVC_PROFILE_MAIN; /* main profile */ 620 else if (profile_compatibility_flag & 0x04) 621 info->profile = HEVC_PROFILE_MAIN10; /* main10 profile */ 622 else if (profile_compatibility_flag & 0x08) 623 /* main still picture profile */ 624 info->profile = HEVC_PROFILE_STILLPICTURE; 625 else 626 info->profile = HEVC_PROFILE_MAIN; /* for old version HM */ 627 } 628 } else if (inst->std == W_AVC_DEC) { 629 info->profile = FIELD_GET(SEQ_PARAM_PROFILE_MASK, reg_val); 630 } 631 632 info->vlc_buf_size = vpu_read_reg(inst->dev, W5_RET_VLC_BUF_SIZE); 633 info->param_buf_size = vpu_read_reg(inst->dev, W5_RET_PARAM_BUF_SIZE); 634 p_dec_info->vlc_buf_size = info->vlc_buf_size; 635 p_dec_info->param_buf_size = info->param_buf_size; 636 } 637 638 int wave5_vpu_dec_get_seq_info(struct vpu_instance *inst, struct dec_initial_info *info) 639 { 640 int ret; 641 u32 reg_val; 642 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 643 644 vpu_write_reg(inst->dev, W5_CMD_DEC_ADDR_REPORT_BASE, p_dec_info->user_data_buf_addr); 645 vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_SIZE, p_dec_info->user_data_buf_size); 646 vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_PARAM, REPORT_PARAM_ENDIANNESS_BIG_ENDIAN); 647 648 /* send QUERY cmd */ 649 ret = wave5_send_query(inst->dev, inst, GET_RESULT); 650 if (ret) 651 return ret; 652 653 reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS); 654 655 p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff; 656 p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK); 657 658 dev_dbg(inst->dev->dev, "%s: init seq complete (queue %u : %u)\n", __func__, 659 p_dec_info->instance_queue_count, p_dec_info->report_queue_count); 660 661 /* this is not a fatal error, set ret to -EIO but don't return immediately */ 662 if (vpu_read_reg(inst->dev, W5_RET_DEC_DECODING_SUCCESS) != 1) { 663 info->seq_init_err_reason = vpu_read_reg(inst->dev, W5_RET_DEC_ERR_INFO); 664 ret = -EIO; 665 } 666 667 wave5_get_dec_seq_result(inst, info); 668 669 return ret; 670 } 671 672 int wave5_vpu_dec_register_framebuffer(struct vpu_instance *inst, struct frame_buffer *fb_arr, 673 enum tiled_map_type map_type, unsigned int count) 674 { 675 int ret; 676 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 677 struct dec_initial_info *init_info = &p_dec_info->initial_info; 678 size_t remain, idx, j, i, cnt_8_chunk, size; 679 u32 start_no, end_no; 680 u32 reg_val, cbcr_interleave, nv21, pic_size; 681 u32 addr_y, addr_cb, addr_cr; 682 u32 mv_col_size, frame_width, frame_height, fbc_y_tbl_size, fbc_c_tbl_size; 683 struct vpu_buf vb_buf; 684 bool justified = WTL_RIGHT_JUSTIFIED; 685 u32 format_no = WTL_PIXEL_8BIT; 686 u32 color_format = 0; 687 u32 pixel_order = 1; 688 u32 bwb_flag = (map_type == LINEAR_FRAME_MAP) ? 1 : 0; 689 690 cbcr_interleave = inst->cbcr_interleave; 691 nv21 = inst->nv21; 692 mv_col_size = 0; 693 fbc_y_tbl_size = 0; 694 fbc_c_tbl_size = 0; 695 696 if (map_type >= COMPRESSED_FRAME_MAP) { 697 cbcr_interleave = 0; 698 nv21 = 0; 699 700 switch (inst->std) { 701 case W_HEVC_DEC: 702 mv_col_size = WAVE5_DEC_HEVC_BUF_SIZE(init_info->pic_width, 703 init_info->pic_height); 704 break; 705 case W_AVC_DEC: 706 mv_col_size = WAVE5_DEC_AVC_BUF_SIZE(init_info->pic_width, 707 init_info->pic_height); 708 break; 709 default: 710 return -EINVAL; 711 } 712 713 if (inst->std == W_HEVC_DEC || inst->std == W_AVC_DEC) { 714 size = ALIGN(ALIGN(mv_col_size, 16), BUFFER_MARGIN) + BUFFER_MARGIN; 715 ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_mv, count, size); 716 if (ret) 717 goto free_mv_buffers; 718 } 719 720 frame_width = init_info->pic_width; 721 frame_height = init_info->pic_height; 722 fbc_y_tbl_size = ALIGN(WAVE5_FBC_LUMA_TABLE_SIZE(frame_width, frame_height), 16); 723 fbc_c_tbl_size = ALIGN(WAVE5_FBC_CHROMA_TABLE_SIZE(frame_width, frame_height), 16); 724 725 size = ALIGN(fbc_y_tbl_size, BUFFER_MARGIN) + BUFFER_MARGIN; 726 ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_fbc_y_tbl, count, size); 727 if (ret) 728 goto free_fbc_y_tbl_buffers; 729 730 size = ALIGN(fbc_c_tbl_size, BUFFER_MARGIN) + BUFFER_MARGIN; 731 ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_fbc_c_tbl, count, size); 732 if (ret) 733 goto free_fbc_c_tbl_buffers; 734 735 pic_size = (init_info->pic_width << 16) | (init_info->pic_height); 736 737 vb_buf.size = (p_dec_info->vlc_buf_size * VLC_BUF_NUM) + 738 (p_dec_info->param_buf_size * COMMAND_QUEUE_DEPTH); 739 vb_buf.daddr = 0; 740 741 if (vb_buf.size != p_dec_info->vb_task.size) { 742 wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_task); 743 ret = wave5_vdi_allocate_dma_memory(inst->dev, &vb_buf); 744 if (ret) 745 goto free_fbc_c_tbl_buffers; 746 747 p_dec_info->vb_task = vb_buf; 748 } 749 750 vpu_write_reg(inst->dev, W5_CMD_SET_FB_ADDR_TASK_BUF, 751 p_dec_info->vb_task.daddr); 752 vpu_write_reg(inst->dev, W5_CMD_SET_FB_TASK_BUF_SIZE, vb_buf.size); 753 } else { 754 pic_size = (init_info->pic_width << 16) | (init_info->pic_height); 755 756 if (inst->output_format == FORMAT_422) 757 color_format = 1; 758 } 759 vpu_write_reg(inst->dev, W5_PIC_SIZE, pic_size); 760 761 reg_val = (bwb_flag << 28) | 762 (pixel_order << 23) | 763 (justified << 22) | 764 (format_no << 20) | 765 (color_format << 19) | 766 (nv21 << 17) | 767 (cbcr_interleave << 16) | 768 (fb_arr[0].stride); 769 vpu_write_reg(inst->dev, W5_COMMON_PIC_INFO, reg_val); 770 771 remain = count; 772 cnt_8_chunk = DIV_ROUND_UP(count, 8); 773 idx = 0; 774 for (j = 0; j < cnt_8_chunk; j++) { 775 reg_val = (j == cnt_8_chunk - 1) << 4 | ((j == 0) << 3); 776 vpu_write_reg(inst->dev, W5_SFB_OPTION, reg_val); 777 start_no = j * 8; 778 end_no = start_no + ((remain >= 8) ? 8 : remain) - 1; 779 780 vpu_write_reg(inst->dev, W5_SET_FB_NUM, (start_no << 8) | end_no); 781 782 for (i = 0; i < 8 && i < remain; i++) { 783 addr_y = fb_arr[i + start_no].buf_y; 784 addr_cb = fb_arr[i + start_no].buf_cb; 785 addr_cr = fb_arr[i + start_no].buf_cr; 786 vpu_write_reg(inst->dev, W5_ADDR_LUMA_BASE0 + (i << 4), addr_y); 787 vpu_write_reg(inst->dev, W5_ADDR_CB_BASE0 + (i << 4), addr_cb); 788 if (map_type >= COMPRESSED_FRAME_MAP) { 789 /* luma FBC offset table */ 790 vpu_write_reg(inst->dev, W5_ADDR_FBC_Y_OFFSET0 + (i << 4), 791 p_dec_info->vb_fbc_y_tbl[idx].daddr); 792 /* chroma FBC offset table */ 793 vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4), 794 p_dec_info->vb_fbc_c_tbl[idx].daddr); 795 vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2), 796 p_dec_info->vb_mv[idx].daddr); 797 } else { 798 vpu_write_reg(inst->dev, W5_ADDR_CR_BASE0 + (i << 4), addr_cr); 799 vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4), 0); 800 vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2), 0); 801 } 802 idx++; 803 } 804 remain -= i; 805 806 ret = send_firmware_command(inst, W5_SET_FB, false, NULL, NULL); 807 if (ret) 808 goto free_buffers; 809 } 810 811 reg_val = vpu_read_reg(inst->dev, W5_RET_SUCCESS); 812 if (!reg_val) { 813 ret = -EIO; 814 goto free_buffers; 815 } 816 817 return 0; 818 819 free_buffers: 820 wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_task); 821 free_fbc_c_tbl_buffers: 822 for (i = 0; i < count; i++) 823 wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_fbc_c_tbl[i]); 824 free_fbc_y_tbl_buffers: 825 for (i = 0; i < count; i++) 826 wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_fbc_y_tbl[i]); 827 free_mv_buffers: 828 for (i = 0; i < count; i++) 829 wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_mv[i]); 830 return ret; 831 } 832 833 int wave5_vpu_decode(struct vpu_instance *inst, u32 *fail_res) 834 { 835 u32 reg_val; 836 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 837 int ret; 838 839 vpu_write_reg(inst->dev, W5_BS_RD_PTR, p_dec_info->stream_rd_ptr); 840 vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr); 841 842 vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info)); 843 844 /* secondary AXI */ 845 reg_val = p_dec_info->sec_axi_info.use_bit_enable | 846 (p_dec_info->sec_axi_info.use_ip_enable << 9) | 847 (p_dec_info->sec_axi_info.use_lf_row_enable << 15); 848 vpu_write_reg(inst->dev, W5_USE_SEC_AXI, reg_val); 849 850 /* set attributes of user buffer */ 851 vpu_write_reg(inst->dev, W5_CMD_DEC_USER_MASK, p_dec_info->user_data_enable); 852 853 vpu_write_reg(inst->dev, W5_COMMAND_OPTION, DEC_PIC_NORMAL); 854 vpu_write_reg(inst->dev, W5_CMD_DEC_TEMPORAL_ID_PLUS1, 855 (p_dec_info->target_spatial_id << 9) | 856 (p_dec_info->temp_id_select_mode << 8) | p_dec_info->target_temp_id); 857 vpu_write_reg(inst->dev, W5_CMD_SEQ_CHANGE_ENABLE_FLAG, p_dec_info->seq_change_mask); 858 /* When reordering is disabled we force the latency of the framebuffers */ 859 vpu_write_reg(inst->dev, W5_CMD_DEC_FORCE_FB_LATENCY_PLUS1, !p_dec_info->reorder_enable); 860 861 ret = send_firmware_command(inst, W5_DEC_ENC_PIC, true, ®_val, fail_res); 862 if (ret == -ETIMEDOUT) 863 return ret; 864 865 p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff; 866 p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK); 867 868 dev_dbg(inst->dev->dev, "%s: dec pic sent (queue %u : %u)\n", __func__, 869 p_dec_info->instance_queue_count, p_dec_info->report_queue_count); 870 871 if (ret) 872 return ret; 873 874 return 0; 875 } 876 877 int wave5_vpu_dec_get_result(struct vpu_instance *inst, struct dec_output_info *result) 878 { 879 int ret; 880 u32 index, nal_unit_type, reg_val, sub_layer_info; 881 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 882 struct vpu_device *vpu_dev = inst->dev; 883 884 vpu_write_reg(inst->dev, W5_CMD_DEC_ADDR_REPORT_BASE, p_dec_info->user_data_buf_addr); 885 vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_SIZE, p_dec_info->user_data_buf_size); 886 vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_PARAM, REPORT_PARAM_ENDIANNESS_BIG_ENDIAN); 887 888 /* send QUERY cmd */ 889 ret = wave5_send_query(vpu_dev, inst, GET_RESULT); 890 if (ret) 891 return ret; 892 893 reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS); 894 895 p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff; 896 p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK); 897 898 dev_dbg(inst->dev->dev, "%s: dec pic complete (queue %u : %u)\n", __func__, 899 p_dec_info->instance_queue_count, p_dec_info->report_queue_count); 900 901 reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_TYPE); 902 903 nal_unit_type = (reg_val >> 4) & 0x3f; 904 905 if (inst->std == W_HEVC_DEC) { 906 if (reg_val & 0x04) 907 result->pic_type = PIC_TYPE_B; 908 else if (reg_val & 0x02) 909 result->pic_type = PIC_TYPE_P; 910 else if (reg_val & 0x01) 911 result->pic_type = PIC_TYPE_I; 912 else 913 result->pic_type = PIC_TYPE_MAX; 914 if ((nal_unit_type == 19 || nal_unit_type == 20) && result->pic_type == PIC_TYPE_I) 915 /* IDR_W_RADL, IDR_N_LP */ 916 result->pic_type = PIC_TYPE_IDR; 917 } else if (inst->std == W_AVC_DEC) { 918 if (reg_val & 0x04) 919 result->pic_type = PIC_TYPE_B; 920 else if (reg_val & 0x02) 921 result->pic_type = PIC_TYPE_P; 922 else if (reg_val & 0x01) 923 result->pic_type = PIC_TYPE_I; 924 else 925 result->pic_type = PIC_TYPE_MAX; 926 if (nal_unit_type == 5 && result->pic_type == PIC_TYPE_I) 927 result->pic_type = PIC_TYPE_IDR; 928 } 929 index = vpu_read_reg(inst->dev, W5_RET_DEC_DISPLAY_INDEX); 930 result->index_frame_display = index; 931 index = vpu_read_reg(inst->dev, W5_RET_DEC_DECODED_INDEX); 932 result->index_frame_decoded = index; 933 result->index_frame_decoded_for_tiled = index; 934 935 sub_layer_info = vpu_read_reg(inst->dev, W5_RET_DEC_SUB_LAYER_INFO); 936 result->temporal_id = sub_layer_info & 0x7; 937 938 if (inst->std == W_HEVC_DEC || inst->std == W_AVC_DEC) { 939 result->decoded_poc = -1; 940 if (result->index_frame_decoded >= 0 || 941 result->index_frame_decoded == DECODED_IDX_FLAG_SKIP) 942 result->decoded_poc = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_POC); 943 } 944 945 result->sequence_changed = vpu_read_reg(inst->dev, W5_RET_DEC_NOTIFICATION); 946 reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_SIZE); 947 result->dec_pic_width = reg_val >> 16; 948 result->dec_pic_height = reg_val & 0xffff; 949 950 if (result->sequence_changed) { 951 memcpy((void *)&p_dec_info->new_seq_info, (void *)&p_dec_info->initial_info, 952 sizeof(struct dec_initial_info)); 953 wave5_get_dec_seq_result(inst, &p_dec_info->new_seq_info); 954 } 955 956 result->dec_host_cmd_tick = vpu_read_reg(inst->dev, W5_RET_DEC_HOST_CMD_TICK); 957 result->dec_decode_end_tick = vpu_read_reg(inst->dev, W5_RET_DEC_DECODING_ENC_TICK); 958 959 if (!p_dec_info->first_cycle_check) { 960 result->frame_cycle = 961 (result->dec_decode_end_tick - result->dec_host_cmd_tick) * 962 p_dec_info->cycle_per_tick; 963 vpu_dev->last_performance_cycles = result->dec_decode_end_tick; 964 p_dec_info->first_cycle_check = true; 965 } else if (result->index_frame_decoded_for_tiled != -1) { 966 result->frame_cycle = 967 (result->dec_decode_end_tick - vpu_dev->last_performance_cycles) * 968 p_dec_info->cycle_per_tick; 969 vpu_dev->last_performance_cycles = result->dec_decode_end_tick; 970 if (vpu_dev->last_performance_cycles < result->dec_host_cmd_tick) 971 result->frame_cycle = 972 (result->dec_decode_end_tick - result->dec_host_cmd_tick) * 973 p_dec_info->cycle_per_tick; 974 } 975 976 /* no remaining command. reset frame cycle. */ 977 if (p_dec_info->instance_queue_count == 0 && p_dec_info->report_queue_count == 0) 978 p_dec_info->first_cycle_check = false; 979 980 return 0; 981 } 982 983 int wave5_vpu_re_init(struct device *dev, u8 *fw, size_t size) 984 { 985 struct vpu_buf *common_vb; 986 dma_addr_t code_base, temp_base; 987 dma_addr_t old_code_base, temp_size; 988 u32 code_size, reason_code; 989 u32 reg_val; 990 struct vpu_device *vpu_dev = dev_get_drvdata(dev); 991 992 common_vb = &vpu_dev->common_mem; 993 994 code_base = common_vb->daddr; 995 /* ALIGN TO 4KB */ 996 code_size = (WAVE5_MAX_CODE_BUF_SIZE & ~0xfff); 997 if (code_size < size * 2) 998 return -EINVAL; 999 temp_base = common_vb->daddr + WAVE5_TEMPBUF_OFFSET; 1000 temp_size = WAVE5_TEMPBUF_SIZE; 1001 1002 old_code_base = vpu_read_reg(vpu_dev, W5_VPU_REMAP_PADDR); 1003 1004 if (old_code_base != code_base + W5_REMAP_INDEX1 * W5_REMAP_MAX_SIZE) { 1005 int ret; 1006 1007 ret = wave5_vdi_write_memory(vpu_dev, common_vb, 0, fw, size); 1008 if (ret < 0) { 1009 dev_err(vpu_dev->dev, 1010 "VPU init, Writing firmware to common buffer, fail: %d\n", ret); 1011 return ret; 1012 } 1013 1014 vpu_write_reg(vpu_dev, W5_PO_CONF, 0); 1015 1016 ret = wave5_vpu_reset(dev, SW_RESET_ON_BOOT); 1017 if (ret < 0) { 1018 dev_err(vpu_dev->dev, "VPU init, Resetting the VPU, fail: %d\n", ret); 1019 return ret; 1020 } 1021 1022 remap_page(vpu_dev, code_base, W5_REMAP_INDEX0); 1023 remap_page(vpu_dev, code_base, W5_REMAP_INDEX1); 1024 1025 vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base); 1026 vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size); 1027 vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0); 1028 vpu_write_reg(vpu_dev, W5_ADDR_TEMP_BASE, temp_base); 1029 vpu_write_reg(vpu_dev, W5_TEMP_SIZE, temp_size); 1030 1031 /* These register must be reset explicitly */ 1032 vpu_write_reg(vpu_dev, W5_HW_OPTION, 0); 1033 wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0); 1034 wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0); 1035 vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0); 1036 1037 /* Encoder interrupt */ 1038 reg_val = BIT(INT_WAVE5_ENC_SET_PARAM); 1039 reg_val |= BIT(INT_WAVE5_ENC_PIC); 1040 reg_val |= BIT(INT_WAVE5_BSBUF_FULL); 1041 /* Decoder interrupt */ 1042 reg_val |= BIT(INT_WAVE5_INIT_SEQ); 1043 reg_val |= BIT(INT_WAVE5_DEC_PIC); 1044 reg_val |= BIT(INT_WAVE5_BSBUF_EMPTY); 1045 vpu_write_reg(vpu_dev, W5_VPU_VINT_ENABLE, reg_val); 1046 1047 reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0); 1048 if (FIELD_GET(FEATURE_BACKBONE, reg_val)) { 1049 reg_val = ((WAVE5_PROC_AXI_ID << 28) | 1050 (WAVE5_PRP_AXI_ID << 24) | 1051 (WAVE5_FBD_Y_AXI_ID << 20) | 1052 (WAVE5_FBC_Y_AXI_ID << 16) | 1053 (WAVE5_FBD_C_AXI_ID << 12) | 1054 (WAVE5_FBC_C_AXI_ID << 8) | 1055 (WAVE5_PRI_AXI_ID << 4) | 1056 WAVE5_SEC_AXI_ID); 1057 wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val); 1058 } 1059 1060 vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1); 1061 vpu_write_reg(vpu_dev, W5_COMMAND, W5_INIT_VPU); 1062 vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1); 1063 1064 ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS); 1065 if (ret) { 1066 dev_err(vpu_dev->dev, "VPU reinit(W5_VPU_REMAP_CORE_START) timeout\n"); 1067 return ret; 1068 } 1069 1070 ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code); 1071 if (ret) 1072 return ret; 1073 } 1074 1075 return setup_wave5_properties(dev); 1076 } 1077 1078 static int wave5_vpu_sleep_wake(struct device *dev, bool i_sleep_wake, const uint16_t *code, 1079 size_t size) 1080 { 1081 u32 reg_val; 1082 struct vpu_buf *common_vb; 1083 dma_addr_t code_base; 1084 u32 code_size, reason_code; 1085 struct vpu_device *vpu_dev = dev_get_drvdata(dev); 1086 int ret; 1087 1088 if (i_sleep_wake) { 1089 ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS); 1090 if (ret) 1091 return ret; 1092 1093 /* 1094 * Declare who has ownership for the host interface access 1095 * 1 = VPU 1096 * 0 = Host processor 1097 */ 1098 vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1); 1099 vpu_write_reg(vpu_dev, W5_COMMAND, W5_SLEEP_VPU); 1100 /* Send an interrupt named HOST to the VPU */ 1101 vpu_write_reg(vpu_dev, W5_VPU_HOST_INT_REQ, 1); 1102 1103 ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS); 1104 if (ret) 1105 return ret; 1106 1107 ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code); 1108 if (ret) 1109 return ret; 1110 } else { /* restore */ 1111 common_vb = &vpu_dev->common_mem; 1112 1113 code_base = common_vb->daddr; 1114 /* ALIGN TO 4KB */ 1115 code_size = (WAVE5_MAX_CODE_BUF_SIZE & ~0xfff); 1116 if (code_size < size * 2) { 1117 dev_err(dev, "size too small\n"); 1118 return -EINVAL; 1119 } 1120 1121 /* Power on without DEBUG mode */ 1122 vpu_write_reg(vpu_dev, W5_PO_CONF, 0); 1123 1124 remap_page(vpu_dev, code_base, W5_REMAP_INDEX0); 1125 remap_page(vpu_dev, code_base, W5_REMAP_INDEX1); 1126 1127 vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base); 1128 vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size); 1129 vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0); 1130 1131 /* These register must be reset explicitly */ 1132 vpu_write_reg(vpu_dev, W5_HW_OPTION, 0); 1133 wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0); 1134 wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0); 1135 vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0); 1136 1137 /* Encoder interrupt */ 1138 reg_val = BIT(INT_WAVE5_ENC_SET_PARAM); 1139 reg_val |= BIT(INT_WAVE5_ENC_PIC); 1140 reg_val |= BIT(INT_WAVE5_BSBUF_FULL); 1141 /* Decoder interrupt */ 1142 reg_val |= BIT(INT_WAVE5_INIT_SEQ); 1143 reg_val |= BIT(INT_WAVE5_DEC_PIC); 1144 reg_val |= BIT(INT_WAVE5_BSBUF_EMPTY); 1145 vpu_write_reg(vpu_dev, W5_VPU_VINT_ENABLE, reg_val); 1146 1147 reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0); 1148 if (FIELD_GET(FEATURE_BACKBONE, reg_val)) { 1149 reg_val = ((WAVE5_PROC_AXI_ID << 28) | 1150 (WAVE5_PRP_AXI_ID << 24) | 1151 (WAVE5_FBD_Y_AXI_ID << 20) | 1152 (WAVE5_FBC_Y_AXI_ID << 16) | 1153 (WAVE5_FBD_C_AXI_ID << 12) | 1154 (WAVE5_FBC_C_AXI_ID << 8) | 1155 (WAVE5_PRI_AXI_ID << 4) | 1156 WAVE5_SEC_AXI_ID); 1157 wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val); 1158 } 1159 1160 vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1); 1161 vpu_write_reg(vpu_dev, W5_COMMAND, W5_WAKEUP_VPU); 1162 /* Start VPU after settings */ 1163 vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1); 1164 1165 ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS); 1166 if (ret) { 1167 dev_err(vpu_dev->dev, "VPU wakeup(W5_VPU_REMAP_CORE_START) timeout\n"); 1168 return ret; 1169 } 1170 1171 return wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code); 1172 } 1173 1174 return 0; 1175 } 1176 1177 int wave5_vpu_reset(struct device *dev, enum sw_reset_mode reset_mode) 1178 { 1179 u32 val = 0; 1180 int ret = 0; 1181 struct vpu_device *vpu_dev = dev_get_drvdata(dev); 1182 struct vpu_attr *p_attr = &vpu_dev->attr; 1183 /* VPU doesn't send response. force to set BUSY flag to 0. */ 1184 vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 0); 1185 1186 if (reset_mode == SW_RESET_SAFETY) { 1187 ret = wave5_vpu_sleep_wake(dev, true, NULL, 0); 1188 if (ret) 1189 return ret; 1190 } 1191 1192 val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0); 1193 if ((val >> 16) & 0x1) 1194 p_attr->support_backbone = true; 1195 if ((val >> 22) & 0x1) 1196 p_attr->support_vcore_backbone = true; 1197 if ((val >> 28) & 0x1) 1198 p_attr->support_vcpu_backbone = true; 1199 1200 /* waiting for completion of bus transaction */ 1201 if (p_attr->support_backbone) { 1202 dev_dbg(dev, "%s: backbone supported\n", __func__); 1203 1204 if (p_attr->support_vcore_backbone) { 1205 if (p_attr->support_vcpu_backbone) { 1206 /* step1 : disable request */ 1207 wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0xFF); 1208 1209 /* step2 : waiting for completion of bus transaction */ 1210 ret = wave5_wait_vcpu_bus_busy(vpu_dev, 1211 W5_BACKBONE_BUS_STATUS_VCPU); 1212 if (ret) { 1213 wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0x00); 1214 return ret; 1215 } 1216 } 1217 /* step1 : disable request */ 1218 wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x7); 1219 1220 /* step2 : waiting for completion of bus transaction */ 1221 if (wave5_wait_bus_busy(vpu_dev, W5_BACKBONE_BUS_STATUS_VCORE0)) { 1222 wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x00); 1223 return -EBUSY; 1224 } 1225 } else { 1226 /* step1 : disable request */ 1227 wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x7); 1228 1229 /* step2 : waiting for completion of bus transaction */ 1230 if (wave5_wait_bus_busy(vpu_dev, W5_COMBINED_BACKBONE_BUS_STATUS)) { 1231 wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x00); 1232 return -EBUSY; 1233 } 1234 } 1235 } else { 1236 dev_dbg(dev, "%s: backbone NOT supported\n", __func__); 1237 /* step1 : disable request */ 1238 wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x100); 1239 1240 /* step2 : waiting for completion of bus transaction */ 1241 ret = wave5_wait_bus_busy(vpu_dev, W5_GDI_BUS_STATUS); 1242 if (ret) { 1243 wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x00); 1244 return ret; 1245 } 1246 } 1247 1248 switch (reset_mode) { 1249 case SW_RESET_ON_BOOT: 1250 case SW_RESET_FORCE: 1251 case SW_RESET_SAFETY: 1252 val = W5_RST_BLOCK_ALL; 1253 break; 1254 default: 1255 return -EINVAL; 1256 } 1257 1258 if (val) { 1259 vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, val); 1260 1261 ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_RESET_STATUS); 1262 if (ret) { 1263 vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, 0); 1264 return ret; 1265 } 1266 vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, 0); 1267 } 1268 /* step3 : must clear GDI_BUS_CTRL after done SW_RESET */ 1269 if (p_attr->support_backbone) { 1270 if (p_attr->support_vcore_backbone) { 1271 if (p_attr->support_vcpu_backbone) 1272 wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0x00); 1273 wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x00); 1274 } else { 1275 wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x00); 1276 } 1277 } else { 1278 wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x00); 1279 } 1280 if (reset_mode == SW_RESET_SAFETY || reset_mode == SW_RESET_FORCE) 1281 ret = wave5_vpu_sleep_wake(dev, false, NULL, 0); 1282 1283 return ret; 1284 } 1285 1286 int wave5_vpu_dec_finish_seq(struct vpu_instance *inst, u32 *fail_res) 1287 { 1288 return send_firmware_command(inst, W5_DESTROY_INSTANCE, true, NULL, fail_res); 1289 } 1290 1291 int wave5_vpu_dec_set_bitstream_flag(struct vpu_instance *inst, bool eos) 1292 { 1293 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 1294 1295 p_dec_info->stream_endflag = eos ? 1 : 0; 1296 vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info)); 1297 vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr); 1298 1299 return send_firmware_command(inst, W5_UPDATE_BS, true, NULL, NULL); 1300 } 1301 1302 int wave5_dec_clr_disp_flag(struct vpu_instance *inst, unsigned int index) 1303 { 1304 struct dec_info *p_dec_info = &inst->codec_info->dec_info; 1305 int ret; 1306 1307 vpu_write_reg(inst->dev, W5_CMD_DEC_CLR_DISP_IDC, BIT(index)); 1308 vpu_write_reg(inst->dev, W5_CMD_DEC_SET_DISP_IDC, 0); 1309 1310 ret = wave5_send_query(inst->dev, inst, UPDATE_DISP_FLAG); 1311 if (ret) 1312 return ret; 1313 1314 p_dec_info->frame_display_flag = vpu_read_reg(inst->dev, W5_RET_DEC_DISP_IDC); 1315 1316 return 0; 1317 } 1318 1319 int wave5_dec_set_disp_flag(struct vpu_instance *inst, unsigned int index) 1320 { 1321 int ret; 1322 1323 vpu_write_reg(inst->dev, W5_CMD_DEC_CLR_DISP_IDC, 0); 1324 vpu_write_reg(inst->dev, W5_CMD_DEC_SET_DISP_IDC, BIT(index)); 1325 1326 ret = wave5_send_query(inst->dev, inst, UPDATE_DISP_FLAG); 1327 if (ret) 1328 return ret; 1329 1330 return 0; 1331 } 1332 1333 int wave5_vpu_clear_interrupt(struct vpu_instance *inst, u32 flags) 1334 { 1335 u32 interrupt_reason; 1336 1337 interrupt_reason = vpu_read_reg(inst->dev, W5_VPU_VINT_REASON_USR); 1338 interrupt_reason &= ~flags; 1339 vpu_write_reg(inst->dev, W5_VPU_VINT_REASON_USR, interrupt_reason); 1340 1341 return 0; 1342 } 1343 1344 dma_addr_t wave5_dec_get_rd_ptr(struct vpu_instance *inst) 1345 { 1346 int ret; 1347 1348 ret = wave5_send_query(inst->dev, inst, GET_BS_RD_PTR); 1349 if (ret) 1350 return inst->codec_info->dec_info.stream_rd_ptr; 1351 1352 return vpu_read_reg(inst->dev, W5_RET_QUERY_DEC_BS_RD_PTR); 1353 } 1354 1355 int wave5_dec_set_rd_ptr(struct vpu_instance *inst, dma_addr_t addr) 1356 { 1357 int ret; 1358 1359 vpu_write_reg(inst->dev, W5_RET_QUERY_DEC_SET_BS_RD_PTR, addr); 1360 1361 ret = wave5_send_query(inst->dev, inst, SET_BS_RD_PTR); 1362 1363 return ret; 1364 } 1365 1366 /************************************************************************/ 1367 /* ENCODER functions */ 1368 /************************************************************************/ 1369 1370 int wave5_vpu_build_up_enc_param(struct device *dev, struct vpu_instance *inst, 1371 struct enc_open_param *open_param) 1372 { 1373 int ret; 1374 struct enc_info *p_enc_info = &inst->codec_info->enc_info; 1375 u32 reg_val; 1376 struct vpu_device *vpu_dev = dev_get_drvdata(dev); 1377 dma_addr_t buffer_addr; 1378 size_t buffer_size; 1379 1380 p_enc_info->cycle_per_tick = 256; 1381 if (vpu_dev->sram_buf.size) { 1382 p_enc_info->sec_axi_info.use_enc_rdo_enable = 1; 1383 p_enc_info->sec_axi_info.use_enc_lf_enable = 1; 1384 } 1385 1386 p_enc_info->vb_work.size = WAVE521ENC_WORKBUF_SIZE; 1387 ret = wave5_vdi_allocate_dma_memory(vpu_dev, &p_enc_info->vb_work); 1388 if (ret) { 1389 memset(&p_enc_info->vb_work, 0, sizeof(p_enc_info->vb_work)); 1390 return ret; 1391 } 1392 1393 wave5_vdi_clear_memory(vpu_dev, &p_enc_info->vb_work); 1394 1395 vpu_write_reg(inst->dev, W5_ADDR_WORK_BASE, p_enc_info->vb_work.daddr); 1396 vpu_write_reg(inst->dev, W5_WORK_SIZE, p_enc_info->vb_work.size); 1397 1398 vpu_write_reg(inst->dev, W5_CMD_ADDR_SEC_AXI, vpu_dev->sram_buf.daddr); 1399 vpu_write_reg(inst->dev, W5_CMD_SEC_AXI_SIZE, vpu_dev->sram_buf.size); 1400 1401 reg_val = (open_param->line_buf_int_en << 6) | BITSTREAM_ENDIANNESS_BIG_ENDIAN; 1402 vpu_write_reg(inst->dev, W5_CMD_BS_PARAM, reg_val); 1403 vpu_write_reg(inst->dev, W5_CMD_EXT_ADDR, 0); 1404 vpu_write_reg(inst->dev, W5_CMD_NUM_CQ_DEPTH_M1, (COMMAND_QUEUE_DEPTH - 1)); 1405 1406 /* This register must be reset explicitly */ 1407 vpu_write_reg(inst->dev, W5_CMD_ENC_SRC_OPTIONS, 0); 1408 vpu_write_reg(inst->dev, W5_CMD_ENC_VCORE_INFO, 1); 1409 1410 ret = send_firmware_command(inst, W5_CREATE_INSTANCE, true, NULL, NULL); 1411 if (ret) 1412 goto free_vb_work; 1413 1414 buffer_addr = open_param->bitstream_buffer; 1415 buffer_size = open_param->bitstream_buffer_size; 1416 p_enc_info->stream_rd_ptr = buffer_addr; 1417 p_enc_info->stream_wr_ptr = buffer_addr; 1418 p_enc_info->line_buf_int_en = open_param->line_buf_int_en; 1419 p_enc_info->stream_buf_start_addr = buffer_addr; 1420 p_enc_info->stream_buf_size = buffer_size; 1421 p_enc_info->stream_buf_end_addr = buffer_addr + buffer_size; 1422 p_enc_info->stride = 0; 1423 p_enc_info->initial_info_obtained = false; 1424 p_enc_info->product_code = vpu_read_reg(inst->dev, W5_PRODUCT_NUMBER); 1425 1426 return 0; 1427 free_vb_work: 1428 if (wave5_vdi_free_dma_memory(vpu_dev, &p_enc_info->vb_work)) 1429 memset(&p_enc_info->vb_work, 0, sizeof(p_enc_info->vb_work)); 1430 return ret; 1431 } 1432 1433 static void wave5_set_enc_crop_info(u32 codec, struct enc_wave_param *param, int rot_mode, 1434 int src_width, int src_height) 1435 { 1436 int aligned_width = (codec == W_HEVC_ENC) ? ALIGN(src_width, 32) : ALIGN(src_width, 16); 1437 int aligned_height = (codec == W_HEVC_ENC) ? ALIGN(src_height, 32) : ALIGN(src_height, 16); 1438 int pad_right, pad_bot; 1439 int crop_right, crop_left, crop_top, crop_bot; 1440 int prp_mode = rot_mode >> 1; /* remove prp_enable bit */ 1441 1442 if (codec == W_HEVC_ENC && 1443 (!rot_mode || prp_mode == 14)) /* prp_mode 14 : hor_mir && ver_mir && rot_180 */ 1444 return; 1445 1446 pad_right = aligned_width - src_width; 1447 pad_bot = aligned_height - src_height; 1448 1449 if (param->conf_win_right > 0) 1450 crop_right = param->conf_win_right + pad_right; 1451 else 1452 crop_right = pad_right; 1453 1454 if (param->conf_win_bot > 0) 1455 crop_bot = param->conf_win_bot + pad_bot; 1456 else 1457 crop_bot = pad_bot; 1458 1459 crop_top = param->conf_win_top; 1460 crop_left = param->conf_win_left; 1461 1462 param->conf_win_top = crop_top; 1463 param->conf_win_left = crop_left; 1464 param->conf_win_bot = crop_bot; 1465 param->conf_win_right = crop_right; 1466 1467 switch (prp_mode) { 1468 case 0: 1469 return; 1470 case 1: 1471 case 15: 1472 param->conf_win_top = crop_right; 1473 param->conf_win_left = crop_top; 1474 param->conf_win_bot = crop_left; 1475 param->conf_win_right = crop_bot; 1476 break; 1477 case 2: 1478 case 12: 1479 param->conf_win_top = crop_bot; 1480 param->conf_win_left = crop_right; 1481 param->conf_win_bot = crop_top; 1482 param->conf_win_right = crop_left; 1483 break; 1484 case 3: 1485 case 13: 1486 param->conf_win_top = crop_left; 1487 param->conf_win_left = crop_bot; 1488 param->conf_win_bot = crop_right; 1489 param->conf_win_right = crop_top; 1490 break; 1491 case 4: 1492 case 10: 1493 param->conf_win_top = crop_bot; 1494 param->conf_win_bot = crop_top; 1495 break; 1496 case 8: 1497 case 6: 1498 param->conf_win_left = crop_right; 1499 param->conf_win_right = crop_left; 1500 break; 1501 case 5: 1502 case 11: 1503 param->conf_win_top = crop_left; 1504 param->conf_win_left = crop_top; 1505 param->conf_win_bot = crop_right; 1506 param->conf_win_right = crop_bot; 1507 break; 1508 case 7: 1509 case 9: 1510 param->conf_win_top = crop_right; 1511 param->conf_win_left = crop_bot; 1512 param->conf_win_bot = crop_left; 1513 param->conf_win_right = crop_top; 1514 break; 1515 default: 1516 WARN(1, "Invalid prp_mode: %d, must be in range of 1 - 15\n", prp_mode); 1517 } 1518 } 1519 1520 int wave5_vpu_enc_init_seq(struct vpu_instance *inst) 1521 { 1522 u32 reg_val = 0, rot_mir_mode, fixed_cu_size_mode = 0x7; 1523 struct enc_info *p_enc_info = &inst->codec_info->enc_info; 1524 struct enc_open_param *p_open_param = &p_enc_info->open_param; 1525 struct enc_wave_param *p_param = &p_open_param->wave_param; 1526 1527 /* 1528 * OPT_COMMON: 1529 * the last SET_PARAM command should be called with OPT_COMMON 1530 */ 1531 rot_mir_mode = 0; 1532 if (p_enc_info->rotation_enable) { 1533 switch (p_enc_info->rotation_angle) { 1534 case 0: 1535 rot_mir_mode |= NONE_ROTATE; 1536 break; 1537 case 90: 1538 rot_mir_mode |= ROT_CLOCKWISE_90; 1539 break; 1540 case 180: 1541 rot_mir_mode |= ROT_CLOCKWISE_180; 1542 break; 1543 case 270: 1544 rot_mir_mode |= ROT_CLOCKWISE_270; 1545 break; 1546 } 1547 } 1548 1549 if (p_enc_info->mirror_enable) { 1550 switch (p_enc_info->mirror_direction) { 1551 case MIRDIR_NONE: 1552 rot_mir_mode |= NONE_ROTATE; 1553 break; 1554 case MIRDIR_VER: 1555 rot_mir_mode |= MIR_VER_FLIP; 1556 break; 1557 case MIRDIR_HOR: 1558 rot_mir_mode |= MIR_HOR_FLIP; 1559 break; 1560 case MIRDIR_HOR_VER: 1561 rot_mir_mode |= MIR_HOR_VER_FLIP; 1562 break; 1563 } 1564 } 1565 1566 wave5_set_enc_crop_info(inst->std, p_param, rot_mir_mode, p_open_param->pic_width, 1567 p_open_param->pic_height); 1568 1569 /* SET_PARAM + COMMON */ 1570 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SET_PARAM_OPTION, OPT_COMMON); 1571 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SRC_SIZE, p_open_param->pic_height << 16 1572 | p_open_param->pic_width); 1573 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MAP_ENDIAN, VDI_LITTLE_ENDIAN); 1574 1575 reg_val = p_param->profile | 1576 (p_param->level << 3) | 1577 (p_param->internal_bit_depth << 14); 1578 if (inst->std == W_HEVC_ENC) 1579 reg_val |= (p_param->tier << 12) | 1580 (p_param->tmvp_enable << 23) | 1581 (p_param->sao_enable << 24) | 1582 (p_param->skip_intra_trans << 25) | 1583 (p_param->strong_intra_smooth_enable << 27) | 1584 (p_param->en_still_picture << 30); 1585 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SPS_PARAM, reg_val); 1586 1587 reg_val = (p_param->lossless_enable) | 1588 (p_param->const_intra_pred_flag << 1) | 1589 (p_param->lf_cross_slice_boundary_enable << 2) | 1590 (p_param->wpp_enable << 4) | 1591 (p_param->disable_deblk << 5) | 1592 ((p_param->beta_offset_div2 & 0xF) << 6) | 1593 ((p_param->tc_offset_div2 & 0xF) << 10) | 1594 ((p_param->chroma_cb_qp_offset & 0x1F) << 14) | 1595 ((p_param->chroma_cr_qp_offset & 0x1F) << 19) | 1596 (p_param->transform8x8_enable << 29) | 1597 (p_param->entropy_coding_mode << 30); 1598 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_PPS_PARAM, reg_val); 1599 1600 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_GOP_PARAM, p_param->gop_preset_idx); 1601 1602 if (inst->std == W_AVC_ENC) 1603 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_PARAM, p_param->intra_qp | 1604 ((p_param->intra_period & 0x7ff) << 6) | 1605 ((p_param->avc_idr_period & 0x7ff) << 17)); 1606 else if (inst->std == W_HEVC_ENC) 1607 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_PARAM, 1608 p_param->decoding_refresh_type | (p_param->intra_qp << 3) | 1609 (p_param->intra_period << 16)); 1610 1611 reg_val = (p_param->rdo_skip << 2) | 1612 (p_param->lambda_scaling_enable << 3) | 1613 (fixed_cu_size_mode << 5) | 1614 (p_param->intra_nx_n_enable << 8) | 1615 (p_param->max_num_merge << 18); 1616 1617 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RDO_PARAM, reg_val); 1618 1619 if (inst->std == W_AVC_ENC) 1620 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_REFRESH, 1621 p_param->intra_mb_refresh_arg << 16 | p_param->intra_mb_refresh_mode); 1622 else if (inst->std == W_HEVC_ENC) 1623 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_REFRESH, 1624 p_param->intra_refresh_arg << 16 | p_param->intra_refresh_mode); 1625 1626 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_FRAME_RATE, p_open_param->frame_rate_info); 1627 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_TARGET_RATE, p_open_param->bit_rate); 1628 1629 reg_val = p_open_param->rc_enable | 1630 (p_param->hvs_qp_enable << 2) | 1631 (p_param->hvs_qp_scale << 4) | 1632 ((p_param->initial_rc_qp & 0x3F) << 14) | 1633 (p_open_param->vbv_buffer_size << 20); 1634 if (inst->std == W_AVC_ENC) 1635 reg_val |= (p_param->mb_level_rc_enable << 1); 1636 else if (inst->std == W_HEVC_ENC) 1637 reg_val |= (p_param->cu_level_rc_enable << 1); 1638 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_PARAM, reg_val); 1639 1640 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_WEIGHT_PARAM, 1641 p_param->rc_weight_buf << 8 | p_param->rc_weight_param); 1642 1643 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_MIN_MAX_QP, p_param->min_qp_i | 1644 (p_param->max_qp_i << 6) | (p_param->hvs_max_delta_qp << 12)); 1645 1646 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_INTER_MIN_MAX_QP, p_param->min_qp_p | 1647 (p_param->max_qp_p << 6) | (p_param->min_qp_b << 12) | 1648 (p_param->max_qp_b << 18)); 1649 1650 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_BIT_RATIO_LAYER_0_3, 0); 1651 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_BIT_RATIO_LAYER_4_7, 0); 1652 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_ROT_PARAM, rot_mir_mode); 1653 1654 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_BG_PARAM, 0); 1655 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_LAMBDA_ADDR, 0); 1656 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CONF_WIN_TOP_BOT, 1657 p_param->conf_win_bot << 16 | p_param->conf_win_top); 1658 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CONF_WIN_LEFT_RIGHT, 1659 p_param->conf_win_right << 16 | p_param->conf_win_left); 1660 1661 if (inst->std == W_AVC_ENC) 1662 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INDEPENDENT_SLICE, 1663 p_param->avc_slice_arg << 16 | p_param->avc_slice_mode); 1664 else if (inst->std == W_HEVC_ENC) 1665 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INDEPENDENT_SLICE, 1666 p_param->independ_slice_mode_arg << 16 | 1667 p_param->independ_slice_mode); 1668 1669 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_USER_SCALING_LIST_ADDR, 0); 1670 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NUM_UNITS_IN_TICK, 0); 1671 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_TIME_SCALE, 0); 1672 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NUM_TICKS_POC_DIFF_ONE, 0); 1673 1674 if (inst->std == W_HEVC_ENC) { 1675 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU04, 0); 1676 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU08, 0); 1677 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU16, 0); 1678 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU32, 0); 1679 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU08, 0); 1680 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU16, 0); 1681 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU32, 0); 1682 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_DEPENDENT_SLICE, 1683 p_param->depend_slice_mode_arg << 16 | p_param->depend_slice_mode); 1684 1685 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NR_PARAM, 0); 1686 1687 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NR_WEIGHT, 1688 p_param->nr_intra_weight_y | 1689 (p_param->nr_intra_weight_cb << 5) | 1690 (p_param->nr_intra_weight_cr << 10) | 1691 (p_param->nr_inter_weight_y << 15) | 1692 (p_param->nr_inter_weight_cb << 20) | 1693 (p_param->nr_inter_weight_cr << 25)); 1694 } 1695 vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_VUI_HRD_PARAM, 0); 1696 1697 return send_firmware_command(inst, W5_ENC_SET_PARAM, true, NULL, NULL); 1698 } 1699 1700 int wave5_vpu_enc_get_seq_info(struct vpu_instance *inst, struct enc_initial_info *info) 1701 { 1702 int ret; 1703 u32 reg_val; 1704 struct enc_info *p_enc_info = &inst->codec_info->enc_info; 1705 1706 /* send QUERY cmd */ 1707 ret = wave5_send_query(inst->dev, inst, GET_RESULT); 1708 if (ret) 1709 return ret; 1710 1711 dev_dbg(inst->dev->dev, "%s: init seq\n", __func__); 1712 1713 reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS); 1714 1715 p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff; 1716 p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK); 1717 1718 if (vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_SUCCESS) != 1) { 1719 info->seq_init_err_reason = vpu_read_reg(inst->dev, W5_RET_ENC_ERR_INFO); 1720 ret = -EIO; 1721 } else { 1722 info->warn_info = vpu_read_reg(inst->dev, W5_RET_ENC_WARN_INFO); 1723 } 1724 1725 info->min_frame_buffer_count = vpu_read_reg(inst->dev, W5_RET_ENC_NUM_REQUIRED_FB); 1726 info->min_src_frame_count = vpu_read_reg(inst->dev, W5_RET_ENC_MIN_SRC_BUF_NUM); 1727 info->vlc_buf_size = vpu_read_reg(inst->dev, W5_RET_VLC_BUF_SIZE); 1728 info->param_buf_size = vpu_read_reg(inst->dev, W5_RET_PARAM_BUF_SIZE); 1729 p_enc_info->vlc_buf_size = info->vlc_buf_size; 1730 p_enc_info->param_buf_size = info->param_buf_size; 1731 1732 return ret; 1733 } 1734 1735 static u32 calculate_luma_stride(u32 width, u32 bit_depth) 1736 { 1737 return ALIGN(ALIGN(width, 16) * ((bit_depth > 8) ? 5 : 4), 32); 1738 } 1739 1740 static u32 calculate_chroma_stride(u32 width, u32 bit_depth) 1741 { 1742 return ALIGN(ALIGN(width / 2, 16) * ((bit_depth > 8) ? 5 : 4), 32); 1743 } 1744 1745 int wave5_vpu_enc_register_framebuffer(struct device *dev, struct vpu_instance *inst, 1746 struct frame_buffer *fb_arr, enum tiled_map_type map_type, 1747 unsigned int count) 1748 { 1749 struct vpu_device *vpu_dev = dev_get_drvdata(dev); 1750 int ret = 0; 1751 u32 stride; 1752 u32 start_no, end_no; 1753 size_t remain, idx, j, i, cnt_8_chunk; 1754 u32 reg_val = 0, pic_size = 0, mv_col_size, fbc_y_tbl_size, fbc_c_tbl_size; 1755 u32 sub_sampled_size = 0; 1756 u32 luma_stride, chroma_stride; 1757 u32 buf_height = 0, buf_width = 0; 1758 u32 bit_depth; 1759 bool avc_encoding = (inst->std == W_AVC_ENC); 1760 struct vpu_buf vb_mv = {0}; 1761 struct vpu_buf vb_fbc_y_tbl = {0}; 1762 struct vpu_buf vb_fbc_c_tbl = {0}; 1763 struct vpu_buf vb_sub_sam_buf = {0}; 1764 struct vpu_buf vb_task = {0}; 1765 struct enc_open_param *p_open_param; 1766 struct enc_info *p_enc_info = &inst->codec_info->enc_info; 1767 1768 p_open_param = &p_enc_info->open_param; 1769 mv_col_size = 0; 1770 fbc_y_tbl_size = 0; 1771 fbc_c_tbl_size = 0; 1772 stride = p_enc_info->stride; 1773 bit_depth = p_open_param->wave_param.internal_bit_depth; 1774 1775 if (avc_encoding) { 1776 buf_width = ALIGN(p_open_param->pic_width, 16); 1777 buf_height = ALIGN(p_open_param->pic_height, 16); 1778 1779 if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) && 1780 !(p_enc_info->rotation_angle == 180 && 1781 p_enc_info->mirror_direction == MIRDIR_HOR_VER)) { 1782 buf_width = ALIGN(p_open_param->pic_width, 16); 1783 buf_height = ALIGN(p_open_param->pic_height, 16); 1784 } 1785 1786 if (p_enc_info->rotation_angle == 90 || p_enc_info->rotation_angle == 270) { 1787 buf_width = ALIGN(p_open_param->pic_height, 16); 1788 buf_height = ALIGN(p_open_param->pic_width, 16); 1789 } 1790 } else { 1791 buf_width = ALIGN(p_open_param->pic_width, 8); 1792 buf_height = ALIGN(p_open_param->pic_height, 8); 1793 1794 if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) && 1795 !(p_enc_info->rotation_angle == 180 && 1796 p_enc_info->mirror_direction == MIRDIR_HOR_VER)) { 1797 buf_width = ALIGN(p_open_param->pic_width, 32); 1798 buf_height = ALIGN(p_open_param->pic_height, 32); 1799 } 1800 1801 if (p_enc_info->rotation_angle == 90 || p_enc_info->rotation_angle == 270) { 1802 buf_width = ALIGN(p_open_param->pic_height, 32); 1803 buf_height = ALIGN(p_open_param->pic_width, 32); 1804 } 1805 } 1806 1807 pic_size = (buf_width << 16) | buf_height; 1808 1809 if (avc_encoding) { 1810 mv_col_size = WAVE5_ENC_AVC_BUF_SIZE(buf_width, buf_height); 1811 vb_mv.daddr = 0; 1812 vb_mv.size = ALIGN(mv_col_size * count, BUFFER_MARGIN) + BUFFER_MARGIN; 1813 } else { 1814 mv_col_size = WAVE5_ENC_HEVC_BUF_SIZE(buf_width, buf_height); 1815 mv_col_size = ALIGN(mv_col_size, 16); 1816 vb_mv.daddr = 0; 1817 vb_mv.size = ALIGN(mv_col_size * count, BUFFER_MARGIN) + BUFFER_MARGIN; 1818 } 1819 1820 ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_mv); 1821 if (ret) 1822 return ret; 1823 1824 p_enc_info->vb_mv = vb_mv; 1825 1826 fbc_y_tbl_size = ALIGN(WAVE5_FBC_LUMA_TABLE_SIZE(buf_width, buf_height), 16); 1827 fbc_c_tbl_size = ALIGN(WAVE5_FBC_CHROMA_TABLE_SIZE(buf_width, buf_height), 16); 1828 1829 vb_fbc_y_tbl.daddr = 0; 1830 vb_fbc_y_tbl.size = ALIGN(fbc_y_tbl_size * count, BUFFER_MARGIN) + BUFFER_MARGIN; 1831 ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_fbc_y_tbl); 1832 if (ret) 1833 goto free_vb_fbc_y_tbl; 1834 1835 p_enc_info->vb_fbc_y_tbl = vb_fbc_y_tbl; 1836 1837 vb_fbc_c_tbl.daddr = 0; 1838 vb_fbc_c_tbl.size = ALIGN(fbc_c_tbl_size * count, BUFFER_MARGIN) + BUFFER_MARGIN; 1839 ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_fbc_c_tbl); 1840 if (ret) 1841 goto free_vb_fbc_c_tbl; 1842 1843 p_enc_info->vb_fbc_c_tbl = vb_fbc_c_tbl; 1844 1845 if (avc_encoding) 1846 sub_sampled_size = WAVE5_SUBSAMPLED_ONE_SIZE_AVC(buf_width, buf_height); 1847 else 1848 sub_sampled_size = WAVE5_SUBSAMPLED_ONE_SIZE(buf_width, buf_height); 1849 vb_sub_sam_buf.size = ALIGN(sub_sampled_size * count, BUFFER_MARGIN) + BUFFER_MARGIN; 1850 vb_sub_sam_buf.daddr = 0; 1851 ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_sub_sam_buf); 1852 if (ret) 1853 goto free_vb_sam_buf; 1854 1855 p_enc_info->vb_sub_sam_buf = vb_sub_sam_buf; 1856 1857 vb_task.size = (p_enc_info->vlc_buf_size * VLC_BUF_NUM) + 1858 (p_enc_info->param_buf_size * COMMAND_QUEUE_DEPTH); 1859 vb_task.daddr = 0; 1860 if (p_enc_info->vb_task.size == 0) { 1861 ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_task); 1862 if (ret) 1863 goto free_vb_task; 1864 1865 p_enc_info->vb_task = vb_task; 1866 1867 vpu_write_reg(inst->dev, W5_CMD_SET_FB_ADDR_TASK_BUF, 1868 p_enc_info->vb_task.daddr); 1869 vpu_write_reg(inst->dev, W5_CMD_SET_FB_TASK_BUF_SIZE, vb_task.size); 1870 } 1871 1872 /* set sub-sampled buffer base addr */ 1873 vpu_write_reg(inst->dev, W5_ADDR_SUB_SAMPLED_FB_BASE, vb_sub_sam_buf.daddr); 1874 /* set sub-sampled buffer size for one frame */ 1875 vpu_write_reg(inst->dev, W5_SUB_SAMPLED_ONE_FB_SIZE, sub_sampled_size); 1876 1877 vpu_write_reg(inst->dev, W5_PIC_SIZE, pic_size); 1878 1879 /* set stride of luma/chroma for compressed buffer */ 1880 if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) && 1881 !(p_enc_info->rotation_angle == 180 && 1882 p_enc_info->mirror_direction == MIRDIR_HOR_VER)) { 1883 luma_stride = calculate_luma_stride(buf_width, bit_depth); 1884 chroma_stride = calculate_chroma_stride(buf_width / 2, bit_depth); 1885 } else { 1886 luma_stride = calculate_luma_stride(p_open_param->pic_width, bit_depth); 1887 chroma_stride = calculate_chroma_stride(p_open_param->pic_width / 2, bit_depth); 1888 } 1889 1890 vpu_write_reg(inst->dev, W5_FBC_STRIDE, luma_stride << 16 | chroma_stride); 1891 vpu_write_reg(inst->dev, W5_COMMON_PIC_INFO, stride); 1892 1893 remain = count; 1894 cnt_8_chunk = DIV_ROUND_UP(count, 8); 1895 idx = 0; 1896 for (j = 0; j < cnt_8_chunk; j++) { 1897 reg_val = (j == cnt_8_chunk - 1) << 4 | ((j == 0) << 3); 1898 vpu_write_reg(inst->dev, W5_SFB_OPTION, reg_val); 1899 start_no = j * 8; 1900 end_no = start_no + ((remain >= 8) ? 8 : remain) - 1; 1901 1902 vpu_write_reg(inst->dev, W5_SET_FB_NUM, (start_no << 8) | end_no); 1903 1904 for (i = 0; i < 8 && i < remain; i++) { 1905 vpu_write_reg(inst->dev, W5_ADDR_LUMA_BASE0 + (i << 4), fb_arr[i + 1906 start_no].buf_y); 1907 vpu_write_reg(inst->dev, W5_ADDR_CB_BASE0 + (i << 4), 1908 fb_arr[i + start_no].buf_cb); 1909 /* luma FBC offset table */ 1910 vpu_write_reg(inst->dev, W5_ADDR_FBC_Y_OFFSET0 + (i << 4), 1911 vb_fbc_y_tbl.daddr + idx * fbc_y_tbl_size); 1912 /* chroma FBC offset table */ 1913 vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4), 1914 vb_fbc_c_tbl.daddr + idx * fbc_c_tbl_size); 1915 1916 vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2), 1917 vb_mv.daddr + idx * mv_col_size); 1918 idx++; 1919 } 1920 remain -= i; 1921 1922 ret = send_firmware_command(inst, W5_SET_FB, false, NULL, NULL); 1923 if (ret) 1924 goto free_vb_mem; 1925 } 1926 1927 ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, NULL); 1928 if (ret) 1929 goto free_vb_mem; 1930 1931 return ret; 1932 1933 free_vb_mem: 1934 wave5_vdi_free_dma_memory(vpu_dev, &vb_task); 1935 free_vb_task: 1936 wave5_vdi_free_dma_memory(vpu_dev, &vb_sub_sam_buf); 1937 free_vb_sam_buf: 1938 wave5_vdi_free_dma_memory(vpu_dev, &vb_fbc_c_tbl); 1939 free_vb_fbc_c_tbl: 1940 wave5_vdi_free_dma_memory(vpu_dev, &vb_fbc_y_tbl); 1941 free_vb_fbc_y_tbl: 1942 wave5_vdi_free_dma_memory(vpu_dev, &vb_mv); 1943 return ret; 1944 } 1945 1946 int wave5_vpu_encode(struct vpu_instance *inst, struct enc_param *option, u32 *fail_res) 1947 { 1948 u32 src_frame_format; 1949 u32 reg_val = 0; 1950 u32 src_stride_c = 0; 1951 struct enc_info *p_enc_info = &inst->codec_info->enc_info; 1952 struct frame_buffer *p_src_frame = option->source_frame; 1953 struct enc_open_param *p_open_param = &p_enc_info->open_param; 1954 bool justified = WTL_RIGHT_JUSTIFIED; 1955 u32 format_no = WTL_PIXEL_8BIT; 1956 int ret; 1957 1958 vpu_write_reg(inst->dev, W5_CMD_ENC_BS_START_ADDR, option->pic_stream_buffer_addr); 1959 vpu_write_reg(inst->dev, W5_CMD_ENC_BS_SIZE, option->pic_stream_buffer_size); 1960 p_enc_info->stream_buf_start_addr = option->pic_stream_buffer_addr; 1961 p_enc_info->stream_buf_size = option->pic_stream_buffer_size; 1962 p_enc_info->stream_buf_end_addr = 1963 option->pic_stream_buffer_addr + option->pic_stream_buffer_size; 1964 1965 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_AXI_SEL, DEFAULT_SRC_AXI); 1966 /* secondary AXI */ 1967 reg_val = (p_enc_info->sec_axi_info.use_enc_rdo_enable << 11) | 1968 (p_enc_info->sec_axi_info.use_enc_lf_enable << 15); 1969 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_USE_SEC_AXI, reg_val); 1970 1971 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_REPORT_PARAM, 0); 1972 1973 /* 1974 * CODEOPT_ENC_VCL is used to implicitly encode header/headers to generate bitstream. 1975 * (use ENC_PUT_VIDEO_HEADER for give_command to encode only a header) 1976 */ 1977 if (option->code_option.implicit_header_encode) 1978 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CODE_OPTION, 1979 CODEOPT_ENC_HEADER_IMPLICIT | CODEOPT_ENC_VCL | 1980 (option->code_option.encode_aud << 5) | 1981 (option->code_option.encode_eos << 6) | 1982 (option->code_option.encode_eob << 7)); 1983 else 1984 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CODE_OPTION, 1985 option->code_option.implicit_header_encode | 1986 (option->code_option.encode_vcl << 1) | 1987 (option->code_option.encode_vps << 2) | 1988 (option->code_option.encode_sps << 3) | 1989 (option->code_option.encode_pps << 4) | 1990 (option->code_option.encode_aud << 5) | 1991 (option->code_option.encode_eos << 6) | 1992 (option->code_option.encode_eob << 7)); 1993 1994 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PIC_PARAM, 0); 1995 1996 if (option->src_end_flag) 1997 /* no more source images. */ 1998 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_PIC_IDX, 0xFFFFFFFF); 1999 else 2000 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_PIC_IDX, option->src_idx); 2001 2002 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_Y, p_src_frame->buf_y); 2003 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_U, p_src_frame->buf_cb); 2004 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_V, p_src_frame->buf_cr); 2005 2006 switch (p_open_param->src_format) { 2007 case FORMAT_420: 2008 case FORMAT_422: 2009 case FORMAT_YUYV: 2010 case FORMAT_YVYU: 2011 case FORMAT_UYVY: 2012 case FORMAT_VYUY: 2013 justified = WTL_LEFT_JUSTIFIED; 2014 format_no = WTL_PIXEL_8BIT; 2015 src_stride_c = inst->cbcr_interleave ? p_src_frame->stride : 2016 (p_src_frame->stride / 2); 2017 src_stride_c = (p_open_param->src_format == FORMAT_422) ? src_stride_c * 2 : 2018 src_stride_c; 2019 break; 2020 case FORMAT_420_P10_16BIT_MSB: 2021 case FORMAT_422_P10_16BIT_MSB: 2022 case FORMAT_YUYV_P10_16BIT_MSB: 2023 case FORMAT_YVYU_P10_16BIT_MSB: 2024 case FORMAT_UYVY_P10_16BIT_MSB: 2025 case FORMAT_VYUY_P10_16BIT_MSB: 2026 justified = WTL_RIGHT_JUSTIFIED; 2027 format_no = WTL_PIXEL_16BIT; 2028 src_stride_c = inst->cbcr_interleave ? p_src_frame->stride : 2029 (p_src_frame->stride / 2); 2030 src_stride_c = (p_open_param->src_format == 2031 FORMAT_422_P10_16BIT_MSB) ? src_stride_c * 2 : src_stride_c; 2032 break; 2033 case FORMAT_420_P10_16BIT_LSB: 2034 case FORMAT_422_P10_16BIT_LSB: 2035 case FORMAT_YUYV_P10_16BIT_LSB: 2036 case FORMAT_YVYU_P10_16BIT_LSB: 2037 case FORMAT_UYVY_P10_16BIT_LSB: 2038 case FORMAT_VYUY_P10_16BIT_LSB: 2039 justified = WTL_LEFT_JUSTIFIED; 2040 format_no = WTL_PIXEL_16BIT; 2041 src_stride_c = inst->cbcr_interleave ? p_src_frame->stride : 2042 (p_src_frame->stride / 2); 2043 src_stride_c = (p_open_param->src_format == 2044 FORMAT_422_P10_16BIT_LSB) ? src_stride_c * 2 : src_stride_c; 2045 break; 2046 case FORMAT_420_P10_32BIT_MSB: 2047 case FORMAT_422_P10_32BIT_MSB: 2048 case FORMAT_YUYV_P10_32BIT_MSB: 2049 case FORMAT_YVYU_P10_32BIT_MSB: 2050 case FORMAT_UYVY_P10_32BIT_MSB: 2051 case FORMAT_VYUY_P10_32BIT_MSB: 2052 justified = WTL_RIGHT_JUSTIFIED; 2053 format_no = WTL_PIXEL_32BIT; 2054 src_stride_c = inst->cbcr_interleave ? p_src_frame->stride : 2055 ALIGN(p_src_frame->stride / 2, 16) * BIT(inst->cbcr_interleave); 2056 src_stride_c = (p_open_param->src_format == 2057 FORMAT_422_P10_32BIT_MSB) ? src_stride_c * 2 : src_stride_c; 2058 break; 2059 case FORMAT_420_P10_32BIT_LSB: 2060 case FORMAT_422_P10_32BIT_LSB: 2061 case FORMAT_YUYV_P10_32BIT_LSB: 2062 case FORMAT_YVYU_P10_32BIT_LSB: 2063 case FORMAT_UYVY_P10_32BIT_LSB: 2064 case FORMAT_VYUY_P10_32BIT_LSB: 2065 justified = WTL_LEFT_JUSTIFIED; 2066 format_no = WTL_PIXEL_32BIT; 2067 src_stride_c = inst->cbcr_interleave ? p_src_frame->stride : 2068 ALIGN(p_src_frame->stride / 2, 16) * BIT(inst->cbcr_interleave); 2069 src_stride_c = (p_open_param->src_format == 2070 FORMAT_422_P10_32BIT_LSB) ? src_stride_c * 2 : src_stride_c; 2071 break; 2072 default: 2073 return -EINVAL; 2074 } 2075 2076 src_frame_format = (inst->cbcr_interleave << 1) | (inst->nv21); 2077 switch (p_open_param->packed_format) { 2078 case PACKED_YUYV: 2079 src_frame_format = 4; 2080 break; 2081 case PACKED_YVYU: 2082 src_frame_format = 5; 2083 break; 2084 case PACKED_UYVY: 2085 src_frame_format = 6; 2086 break; 2087 case PACKED_VYUY: 2088 src_frame_format = 7; 2089 break; 2090 default: 2091 break; 2092 } 2093 2094 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_STRIDE, 2095 (p_src_frame->stride << 16) | src_stride_c); 2096 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_FORMAT, src_frame_format | 2097 (format_no << 3) | (justified << 5) | (PIC_SRC_ENDIANNESS_BIG_ENDIAN << 6)); 2098 2099 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CUSTOM_MAP_OPTION_ADDR, 0); 2100 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CUSTOM_MAP_OPTION_PARAM, 0); 2101 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_LONGTERM_PIC, 0); 2102 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_SIGMA_Y, 0); 2103 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_SIGMA_C, 0); 2104 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_MEAN_Y, 0); 2105 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_MEAN_C, 0); 2106 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PREFIX_SEI_INFO, 0); 2107 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PREFIX_SEI_NAL_ADDR, 0); 2108 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SUFFIX_SEI_INFO, 0); 2109 vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SUFFIX_SEI_NAL_ADDR, 0); 2110 2111 ret = send_firmware_command(inst, W5_DEC_ENC_PIC, true, ®_val, fail_res); 2112 if (ret == -ETIMEDOUT) 2113 return ret; 2114 2115 p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff; 2116 p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK); 2117 2118 if (ret) 2119 return ret; 2120 2121 return 0; 2122 } 2123 2124 int wave5_vpu_enc_get_result(struct vpu_instance *inst, struct enc_output_info *result) 2125 { 2126 int ret; 2127 u32 encoding_success; 2128 u32 reg_val; 2129 struct enc_info *p_enc_info = &inst->codec_info->enc_info; 2130 struct vpu_device *vpu_dev = inst->dev; 2131 2132 ret = wave5_send_query(inst->dev, inst, GET_RESULT); 2133 if (ret) 2134 return ret; 2135 2136 dev_dbg(inst->dev->dev, "%s: enc pic complete\n", __func__); 2137 2138 reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS); 2139 2140 p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff; 2141 p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK); 2142 2143 encoding_success = vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_SUCCESS); 2144 if (!encoding_success) { 2145 result->error_reason = vpu_read_reg(inst->dev, W5_RET_ENC_ERR_INFO); 2146 return -EIO; 2147 } 2148 2149 result->warn_info = vpu_read_reg(inst->dev, W5_RET_ENC_WARN_INFO); 2150 2151 reg_val = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_TYPE); 2152 result->pic_type = reg_val & 0xFFFF; 2153 2154 result->enc_vcl_nut = vpu_read_reg(inst->dev, W5_RET_ENC_VCL_NUT); 2155 /* 2156 * To get the reconstructed frame use the following index on 2157 * inst->frame_buf 2158 */ 2159 result->recon_frame_index = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_IDX); 2160 result->enc_pic_byte = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_BYTE); 2161 result->enc_src_idx = vpu_read_reg(inst->dev, W5_RET_ENC_USED_SRC_IDX); 2162 p_enc_info->stream_wr_ptr = vpu_read_reg(inst->dev, W5_RET_ENC_WR_PTR); 2163 p_enc_info->stream_rd_ptr = vpu_read_reg(inst->dev, W5_RET_ENC_RD_PTR); 2164 2165 result->bitstream_buffer = vpu_read_reg(inst->dev, W5_RET_ENC_RD_PTR); 2166 result->rd_ptr = p_enc_info->stream_rd_ptr; 2167 result->wr_ptr = p_enc_info->stream_wr_ptr; 2168 2169 /*result for header only(no vcl) encoding */ 2170 if (result->recon_frame_index == RECON_IDX_FLAG_HEADER_ONLY) 2171 result->bitstream_size = result->enc_pic_byte; 2172 else if (result->recon_frame_index < 0) 2173 result->bitstream_size = 0; 2174 else 2175 result->bitstream_size = result->enc_pic_byte; 2176 2177 result->enc_host_cmd_tick = vpu_read_reg(inst->dev, W5_RET_ENC_HOST_CMD_TICK); 2178 result->enc_encode_end_tick = vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_END_TICK); 2179 2180 if (!p_enc_info->first_cycle_check) { 2181 result->frame_cycle = (result->enc_encode_end_tick - result->enc_host_cmd_tick) * 2182 p_enc_info->cycle_per_tick; 2183 p_enc_info->first_cycle_check = true; 2184 } else { 2185 result->frame_cycle = 2186 (result->enc_encode_end_tick - vpu_dev->last_performance_cycles) * 2187 p_enc_info->cycle_per_tick; 2188 if (vpu_dev->last_performance_cycles < result->enc_host_cmd_tick) 2189 result->frame_cycle = (result->enc_encode_end_tick - 2190 result->enc_host_cmd_tick) * p_enc_info->cycle_per_tick; 2191 } 2192 vpu_dev->last_performance_cycles = result->enc_encode_end_tick; 2193 2194 return 0; 2195 } 2196 2197 int wave5_vpu_enc_finish_seq(struct vpu_instance *inst, u32 *fail_res) 2198 { 2199 return send_firmware_command(inst, W5_DESTROY_INSTANCE, true, NULL, fail_res); 2200 } 2201 2202 static bool wave5_vpu_enc_check_common_param_valid(struct vpu_instance *inst, 2203 struct enc_open_param *open_param) 2204 { 2205 bool low_delay = true; 2206 struct enc_wave_param *param = &open_param->wave_param; 2207 struct vpu_device *vpu_dev = inst->dev; 2208 struct device *dev = vpu_dev->dev; 2209 u32 num_ctu_row = (open_param->pic_height + 64 - 1) / 64; 2210 u32 num_ctu_col = (open_param->pic_width + 64 - 1) / 64; 2211 u32 ctu_sz = num_ctu_col * num_ctu_row; 2212 2213 if (inst->std == W_HEVC_ENC && low_delay && 2214 param->decoding_refresh_type == DEC_REFRESH_TYPE_CRA) { 2215 dev_warn(dev, 2216 "dec_refresh_type(CRA) shouldn't be used together with low delay GOP\n"); 2217 dev_warn(dev, "Suggested configuration parameter: decoding refresh type (IDR)\n"); 2218 param->decoding_refresh_type = 2; 2219 } 2220 2221 if (param->wpp_enable && param->independ_slice_mode) { 2222 unsigned int num_ctb_in_width = ALIGN(open_param->pic_width, 64) >> 6; 2223 2224 if (param->independ_slice_mode_arg % num_ctb_in_width) { 2225 dev_err(dev, "independ_slice_mode_arg %u must be a multiple of %u\n", 2226 param->independ_slice_mode_arg, num_ctb_in_width); 2227 return false; 2228 } 2229 } 2230 2231 /* multi-slice & wpp */ 2232 if (param->wpp_enable && param->depend_slice_mode) { 2233 dev_err(dev, "wpp_enable && depend_slice_mode cannot be used simultaneously\n"); 2234 return false; 2235 } 2236 2237 if (!param->independ_slice_mode && param->depend_slice_mode) { 2238 dev_err(dev, "depend_slice_mode requires independ_slice_mode\n"); 2239 return false; 2240 } else if (param->independ_slice_mode && 2241 param->depend_slice_mode == DEPEND_SLICE_MODE_RECOMMENDED && 2242 param->independ_slice_mode_arg < param->depend_slice_mode_arg) { 2243 dev_err(dev, "independ_slice_mode_arg: %u must be smaller than %u\n", 2244 param->independ_slice_mode_arg, param->depend_slice_mode_arg); 2245 return false; 2246 } 2247 2248 if (param->independ_slice_mode && param->independ_slice_mode_arg > 65535) { 2249 dev_err(dev, "independ_slice_mode_arg: %u must be smaller than 65535\n", 2250 param->independ_slice_mode_arg); 2251 return false; 2252 } 2253 2254 if (param->depend_slice_mode && param->depend_slice_mode_arg > 65535) { 2255 dev_err(dev, "depend_slice_mode_arg: %u must be smaller than 65535\n", 2256 param->depend_slice_mode_arg); 2257 return false; 2258 } 2259 2260 if (param->conf_win_top % 2) { 2261 dev_err(dev, "conf_win_top: %u, must be a multiple of 2\n", param->conf_win_top); 2262 return false; 2263 } 2264 2265 if (param->conf_win_bot % 2) { 2266 dev_err(dev, "conf_win_bot: %u, must be a multiple of 2\n", param->conf_win_bot); 2267 return false; 2268 } 2269 2270 if (param->conf_win_left % 2) { 2271 dev_err(dev, "conf_win_left: %u, must be a multiple of 2\n", param->conf_win_left); 2272 return false; 2273 } 2274 2275 if (param->conf_win_right % 2) { 2276 dev_err(dev, "conf_win_right: %u, Must be a multiple of 2\n", 2277 param->conf_win_right); 2278 return false; 2279 } 2280 2281 if (param->lossless_enable && open_param->rc_enable) { 2282 dev_err(dev, "option rate_control cannot be used with lossless_coding\n"); 2283 return false; 2284 } 2285 2286 if (param->lossless_enable && !param->skip_intra_trans) { 2287 dev_err(dev, "option intra_trans_skip must be enabled with lossless_coding\n"); 2288 return false; 2289 } 2290 2291 /* intra refresh */ 2292 if (param->intra_refresh_mode && param->intra_refresh_arg == 0) { 2293 dev_err(dev, "Invalid refresh argument, mode: %u, refresh: %u must be > 0\n", 2294 param->intra_refresh_mode, param->intra_refresh_arg); 2295 return false; 2296 } 2297 switch (param->intra_refresh_mode) { 2298 case REFRESH_MODE_CTU_ROWS: 2299 if (param->intra_mb_refresh_arg > num_ctu_row) 2300 goto invalid_refresh_argument; 2301 break; 2302 case REFRESH_MODE_CTU_COLUMNS: 2303 if (param->intra_refresh_arg > num_ctu_col) 2304 goto invalid_refresh_argument; 2305 break; 2306 case REFRESH_MODE_CTU_STEP_SIZE: 2307 if (param->intra_refresh_arg > ctu_sz) 2308 goto invalid_refresh_argument; 2309 break; 2310 case REFRESH_MODE_CTUS: 2311 if (param->intra_refresh_arg > ctu_sz) 2312 goto invalid_refresh_argument; 2313 if (param->lossless_enable) { 2314 dev_err(dev, "mode: %u cannot be used lossless_enable", 2315 param->intra_refresh_mode); 2316 return false; 2317 } 2318 } 2319 return true; 2320 2321 invalid_refresh_argument: 2322 dev_err(dev, "Invalid refresh argument, mode: %u, refresh: %u > W(%u)xH(%u)\n", 2323 param->intra_refresh_mode, param->intra_refresh_arg, 2324 num_ctu_row, num_ctu_col); 2325 return false; 2326 } 2327 2328 static bool wave5_vpu_enc_check_param_valid(struct vpu_device *vpu_dev, 2329 struct enc_open_param *open_param) 2330 { 2331 struct enc_wave_param *param = &open_param->wave_param; 2332 2333 if (open_param->rc_enable) { 2334 if (param->min_qp_i > param->max_qp_i || param->min_qp_p > param->max_qp_p || 2335 param->min_qp_b > param->max_qp_b) { 2336 dev_err(vpu_dev->dev, "Configuration failed because min_qp is greater than max_qp\n"); 2337 dev_err(vpu_dev->dev, "Suggested configuration parameters: min_qp = max_qp\n"); 2338 return false; 2339 } 2340 2341 if (open_param->bit_rate <= (int)open_param->frame_rate_info) { 2342 dev_err(vpu_dev->dev, 2343 "enc_bit_rate: %u must be greater than the frame_rate: %u\n", 2344 open_param->bit_rate, (int)open_param->frame_rate_info); 2345 return false; 2346 } 2347 } 2348 2349 return true; 2350 } 2351 2352 int wave5_vpu_enc_check_open_param(struct vpu_instance *inst, struct enc_open_param *open_param) 2353 { 2354 u32 pic_width; 2355 u32 pic_height; 2356 s32 product_id = inst->dev->product; 2357 struct vpu_attr *p_attr = &inst->dev->attr; 2358 struct enc_wave_param *param; 2359 2360 if (!open_param) 2361 return -EINVAL; 2362 2363 param = &open_param->wave_param; 2364 pic_width = open_param->pic_width; 2365 pic_height = open_param->pic_height; 2366 2367 if (inst->id >= MAX_NUM_INSTANCE) { 2368 dev_err(inst->dev->dev, "Too many simultaneous instances: %d (max: %u)\n", 2369 inst->id, MAX_NUM_INSTANCE); 2370 return -EOPNOTSUPP; 2371 } 2372 2373 if (inst->std != W_HEVC_ENC && 2374 !(inst->std == W_AVC_ENC && product_id == PRODUCT_ID_521)) { 2375 dev_err(inst->dev->dev, "Unsupported encoder-codec & product combination\n"); 2376 return -EOPNOTSUPP; 2377 } 2378 2379 if (param->internal_bit_depth == 10) { 2380 if (inst->std == W_HEVC_ENC && !p_attr->support_hevc10bit_enc) { 2381 dev_err(inst->dev->dev, 2382 "Flag support_hevc10bit_enc must be set to encode 10bit HEVC\n"); 2383 return -EOPNOTSUPP; 2384 } else if (inst->std == W_AVC_ENC && !p_attr->support_avc10bit_enc) { 2385 dev_err(inst->dev->dev, 2386 "Flag support_avc10bit_enc must be set to encode 10bit AVC\n"); 2387 return -EOPNOTSUPP; 2388 } 2389 } 2390 2391 if (!open_param->frame_rate_info) { 2392 dev_err(inst->dev->dev, "No frame rate information.\n"); 2393 return -EINVAL; 2394 } 2395 2396 if (open_param->bit_rate > MAX_BIT_RATE) { 2397 dev_err(inst->dev->dev, "Invalid encoding bit-rate: %u (valid: 0-%u)\n", 2398 open_param->bit_rate, MAX_BIT_RATE); 2399 return -EINVAL; 2400 } 2401 2402 if (pic_width < W5_MIN_ENC_PIC_WIDTH || pic_width > W5_MAX_ENC_PIC_WIDTH || 2403 pic_height < W5_MIN_ENC_PIC_HEIGHT || pic_height > W5_MAX_ENC_PIC_HEIGHT) { 2404 dev_err(inst->dev->dev, "Invalid encoding dimension: %ux%u\n", 2405 pic_width, pic_height); 2406 return -EINVAL; 2407 } 2408 2409 if (param->profile) { 2410 if (inst->std == W_HEVC_ENC) { 2411 if ((param->profile != HEVC_PROFILE_MAIN || 2412 (param->profile == HEVC_PROFILE_MAIN && 2413 param->internal_bit_depth > 8)) && 2414 (param->profile != HEVC_PROFILE_MAIN10 || 2415 (param->profile == HEVC_PROFILE_MAIN10 && 2416 param->internal_bit_depth < 10)) && 2417 param->profile != HEVC_PROFILE_STILLPICTURE) { 2418 dev_err(inst->dev->dev, 2419 "Invalid HEVC encoding profile: %u (bit-depth: %u)\n", 2420 param->profile, param->internal_bit_depth); 2421 return -EINVAL; 2422 } 2423 } else if (inst->std == W_AVC_ENC) { 2424 if ((param->internal_bit_depth > 8 && 2425 param->profile != H264_PROFILE_HIGH10)) { 2426 dev_err(inst->dev->dev, 2427 "Invalid AVC encoding profile: %u (bit-depth: %u)\n", 2428 param->profile, param->internal_bit_depth); 2429 return -EINVAL; 2430 } 2431 } 2432 } 2433 2434 if (param->decoding_refresh_type > DEC_REFRESH_TYPE_IDR) { 2435 dev_err(inst->dev->dev, "Invalid decoding refresh type: %u (valid: 0-2)\n", 2436 param->decoding_refresh_type); 2437 return -EINVAL; 2438 } 2439 2440 if (param->intra_refresh_mode > REFRESH_MODE_CTUS) { 2441 dev_err(inst->dev->dev, "Invalid intra refresh mode: %d (valid: 0-4)\n", 2442 param->intra_refresh_mode); 2443 return -EINVAL; 2444 } 2445 2446 if (inst->std == W_HEVC_ENC && param->independ_slice_mode && 2447 param->depend_slice_mode > DEPEND_SLICE_MODE_BOOST) { 2448 dev_err(inst->dev->dev, 2449 "Can't combine slice modes: independent and fast dependent for HEVC\n"); 2450 return -EINVAL; 2451 } 2452 2453 if (!param->disable_deblk) { 2454 if (param->beta_offset_div2 < -6 || param->beta_offset_div2 > 6) { 2455 dev_err(inst->dev->dev, "Invalid beta offset: %d (valid: -6-6)\n", 2456 param->beta_offset_div2); 2457 return -EINVAL; 2458 } 2459 2460 if (param->tc_offset_div2 < -6 || param->tc_offset_div2 > 6) { 2461 dev_err(inst->dev->dev, "Invalid tc offset: %d (valid: -6-6)\n", 2462 param->tc_offset_div2); 2463 return -EINVAL; 2464 } 2465 } 2466 2467 if (param->intra_qp > MAX_INTRA_QP) { 2468 dev_err(inst->dev->dev, 2469 "Invalid intra quantization parameter: %u (valid: 0-%u)\n", 2470 param->intra_qp, MAX_INTRA_QP); 2471 return -EINVAL; 2472 } 2473 2474 if (open_param->rc_enable) { 2475 if (param->min_qp_i > MAX_INTRA_QP || param->max_qp_i > MAX_INTRA_QP || 2476 param->min_qp_p > MAX_INTRA_QP || param->max_qp_p > MAX_INTRA_QP || 2477 param->min_qp_b > MAX_INTRA_QP || param->max_qp_b > MAX_INTRA_QP) { 2478 dev_err(inst->dev->dev, 2479 "Invalid quantization parameter min/max values: " 2480 "I: %u-%u, P: %u-%u, B: %u-%u (valid for each: 0-%u)\n", 2481 param->min_qp_i, param->max_qp_i, param->min_qp_p, param->max_qp_p, 2482 param->min_qp_b, param->max_qp_b, MAX_INTRA_QP); 2483 return -EINVAL; 2484 } 2485 2486 if (param->hvs_qp_enable && param->hvs_max_delta_qp > MAX_HVS_MAX_DELTA_QP) { 2487 dev_err(inst->dev->dev, 2488 "Invalid HVS max delta quantization parameter: %u (valid: 0-%u)\n", 2489 param->hvs_max_delta_qp, MAX_HVS_MAX_DELTA_QP); 2490 return -EINVAL; 2491 } 2492 2493 if (open_param->vbv_buffer_size < MIN_VBV_BUFFER_SIZE || 2494 open_param->vbv_buffer_size > MAX_VBV_BUFFER_SIZE) { 2495 dev_err(inst->dev->dev, "VBV buffer size: %u (valid: %u-%u)\n", 2496 open_param->vbv_buffer_size, MIN_VBV_BUFFER_SIZE, 2497 MAX_VBV_BUFFER_SIZE); 2498 return -EINVAL; 2499 } 2500 } 2501 2502 if (!wave5_vpu_enc_check_common_param_valid(inst, open_param)) 2503 return -EINVAL; 2504 2505 if (!wave5_vpu_enc_check_param_valid(inst->dev, open_param)) 2506 return -EINVAL; 2507 2508 if (param->chroma_cb_qp_offset < -12 || param->chroma_cb_qp_offset > 12) { 2509 dev_err(inst->dev->dev, 2510 "Invalid chroma Cb quantization parameter offset: %d (valid: -12-12)\n", 2511 param->chroma_cb_qp_offset); 2512 return -EINVAL; 2513 } 2514 2515 if (param->chroma_cr_qp_offset < -12 || param->chroma_cr_qp_offset > 12) { 2516 dev_err(inst->dev->dev, 2517 "Invalid chroma Cr quantization parameter offset: %d (valid: -12-12)\n", 2518 param->chroma_cr_qp_offset); 2519 return -EINVAL; 2520 } 2521 2522 if (param->intra_refresh_mode == REFRESH_MODE_CTU_STEP_SIZE && !param->intra_refresh_arg) { 2523 dev_err(inst->dev->dev, 2524 "Intra refresh mode CTU step-size requires an argument\n"); 2525 return -EINVAL; 2526 } 2527 2528 if (inst->std == W_HEVC_ENC) { 2529 if (param->nr_intra_weight_y > MAX_INTRA_WEIGHT || 2530 param->nr_intra_weight_cb > MAX_INTRA_WEIGHT || 2531 param->nr_intra_weight_cr > MAX_INTRA_WEIGHT) { 2532 dev_err(inst->dev->dev, 2533 "Invalid intra weight Y(%u) Cb(%u) Cr(%u) (valid: %u)\n", 2534 param->nr_intra_weight_y, param->nr_intra_weight_cb, 2535 param->nr_intra_weight_cr, MAX_INTRA_WEIGHT); 2536 return -EINVAL; 2537 } 2538 2539 if (param->nr_inter_weight_y > MAX_INTER_WEIGHT || 2540 param->nr_inter_weight_cb > MAX_INTER_WEIGHT || 2541 param->nr_inter_weight_cr > MAX_INTER_WEIGHT) { 2542 dev_err(inst->dev->dev, 2543 "Invalid inter weight Y(%u) Cb(%u) Cr(%u) (valid: %u)\n", 2544 param->nr_inter_weight_y, param->nr_inter_weight_cb, 2545 param->nr_inter_weight_cr, MAX_INTER_WEIGHT); 2546 return -EINVAL; 2547 } 2548 } 2549 2550 return 0; 2551 } 2552