1 // SPDX-License-Identifier: MIT 2 /* 3 * Copyright © 2023 Intel Corporation 4 */ 5 6 #include "xe_gsc_proxy.h" 7 8 #include <linux/component.h> 9 #include <linux/delay.h> 10 11 #include <drm/drm_managed.h> 12 #include <drm/i915_component.h> 13 #include <drm/i915_gsc_proxy_mei_interface.h> 14 15 #include "abi/gsc_proxy_commands_abi.h" 16 #include "regs/xe_gsc_regs.h" 17 #include "xe_bo.h" 18 #include "xe_gsc.h" 19 #include "xe_gsc_submit.h" 20 #include "xe_gt.h" 21 #include "xe_gt_printk.h" 22 #include "xe_map.h" 23 #include "xe_mmio.h" 24 #include "xe_pm.h" 25 26 /* 27 * GSC proxy: 28 * The GSC uC needs to communicate with the CSME to perform certain operations. 29 * Since the GSC can't perform this communication directly on platforms where it 30 * is integrated in GT, the graphics driver needs to transfer the messages from 31 * GSC to CSME and back. The proxy flow must be manually started after the GSC 32 * is loaded to signal to GSC that we're ready to handle its messages and allow 33 * it to query its init data from CSME; GSC will then trigger an HECI2 interrupt 34 * if it needs to send messages to CSME again. 35 * The proxy flow is as follow: 36 * 1 - Xe submits a request to GSC asking for the message to CSME 37 * 2 - GSC replies with the proxy header + payload for CSME 38 * 3 - Xe sends the reply from GSC as-is to CSME via the mei proxy component 39 * 4 - CSME replies with the proxy header + payload for GSC 40 * 5 - Xe submits a request to GSC with the reply from CSME 41 * 6 - GSC replies either with a new header + payload (same as step 2, so we 42 * restart from there) or with an end message. 43 */ 44 45 /* 46 * The component should load quite quickly in most cases, but it could take 47 * a bit. Using a very big timeout just to cover the worst case scenario 48 */ 49 #define GSC_PROXY_INIT_TIMEOUT_MS 20000 50 51 /* shorthand define for code compactness */ 52 #define PROXY_HDR_SIZE (sizeof(struct xe_gsc_proxy_header)) 53 54 /* the protocol supports up to 32K in each direction */ 55 #define GSC_PROXY_BUFFER_SIZE SZ_32K 56 #define GSC_PROXY_CHANNEL_SIZE (GSC_PROXY_BUFFER_SIZE * 2) 57 58 static struct xe_gt * 59 gsc_to_gt(struct xe_gsc *gsc) 60 { 61 return container_of(gsc, struct xe_gt, uc.gsc); 62 } 63 64 static inline struct xe_device *kdev_to_xe(struct device *kdev) 65 { 66 return dev_get_drvdata(kdev); 67 } 68 69 bool xe_gsc_proxy_init_done(struct xe_gsc *gsc) 70 { 71 struct xe_gt *gt = gsc_to_gt(gsc); 72 u32 fwsts1 = xe_mmio_read32(gt, HECI_FWSTS1(MTL_GSC_HECI1_BASE)); 73 74 return REG_FIELD_GET(HECI1_FWSTS1_CURRENT_STATE, fwsts1) == 75 HECI1_FWSTS1_PROXY_STATE_NORMAL; 76 } 77 78 static void __gsc_proxy_irq_rmw(struct xe_gsc *gsc, u32 clr, u32 set) 79 { 80 struct xe_gt *gt = gsc_to_gt(gsc); 81 82 /* make sure we never accidentally write the RST bit */ 83 clr |= HECI_H_CSR_RST; 84 85 xe_mmio_rmw32(gt, HECI_H_CSR(MTL_GSC_HECI2_BASE), clr, set); 86 } 87 88 static void gsc_proxy_irq_clear(struct xe_gsc *gsc) 89 { 90 /* The status bit is cleared by writing to it */ 91 __gsc_proxy_irq_rmw(gsc, 0, HECI_H_CSR_IS); 92 } 93 94 static void gsc_proxy_irq_toggle(struct xe_gsc *gsc, bool enabled) 95 { 96 u32 set = enabled ? HECI_H_CSR_IE : 0; 97 u32 clr = enabled ? 0 : HECI_H_CSR_IE; 98 99 __gsc_proxy_irq_rmw(gsc, clr, set); 100 } 101 102 static int proxy_send_to_csme(struct xe_gsc *gsc, u32 size) 103 { 104 struct xe_gt *gt = gsc_to_gt(gsc); 105 struct i915_gsc_proxy_component *comp = gsc->proxy.component; 106 int ret; 107 108 ret = comp->ops->send(comp->mei_dev, gsc->proxy.to_csme, size); 109 if (ret < 0) { 110 xe_gt_err(gt, "Failed to send CSME proxy message\n"); 111 return ret; 112 } 113 114 ret = comp->ops->recv(comp->mei_dev, gsc->proxy.from_csme, GSC_PROXY_BUFFER_SIZE); 115 if (ret < 0) { 116 xe_gt_err(gt, "Failed to receive CSME proxy message\n"); 117 return ret; 118 } 119 120 return ret; 121 } 122 123 static int proxy_send_to_gsc(struct xe_gsc *gsc, u32 size) 124 { 125 struct xe_gt *gt = gsc_to_gt(gsc); 126 u64 addr_in = xe_bo_ggtt_addr(gsc->proxy.bo); 127 u64 addr_out = addr_in + GSC_PROXY_BUFFER_SIZE; 128 int err; 129 130 /* the message must contain at least the gsc and proxy headers */ 131 if (size > GSC_PROXY_BUFFER_SIZE) { 132 xe_gt_err(gt, "Invalid GSC proxy message size: %u\n", size); 133 return -EINVAL; 134 } 135 136 err = xe_gsc_pkt_submit_kernel(gsc, addr_in, size, 137 addr_out, GSC_PROXY_BUFFER_SIZE); 138 if (err) { 139 xe_gt_err(gt, "Failed to submit gsc proxy rq (%pe)\n", ERR_PTR(err)); 140 return err; 141 } 142 143 return 0; 144 } 145 146 static int validate_proxy_header(struct xe_gsc_proxy_header *header, 147 u32 source, u32 dest, u32 max_size) 148 { 149 u32 type = FIELD_GET(GSC_PROXY_TYPE, header->hdr); 150 u32 length = FIELD_GET(GSC_PROXY_PAYLOAD_LENGTH, header->hdr); 151 152 if (header->destination != dest || header->source != source) 153 return -ENOEXEC; 154 155 if (length + PROXY_HDR_SIZE > max_size) 156 return -E2BIG; 157 158 switch (type) { 159 case GSC_PROXY_MSG_TYPE_PROXY_PAYLOAD: 160 if (length > 0) 161 break; 162 fallthrough; 163 case GSC_PROXY_MSG_TYPE_PROXY_INVALID: 164 return -EIO; 165 default: 166 break; 167 } 168 169 return 0; 170 } 171 172 #define proxy_header_wr(xe_, map_, offset_, field_, val_) \ 173 xe_map_wr_field(xe_, map_, offset_, struct xe_gsc_proxy_header, field_, val_) 174 175 #define proxy_header_rd(xe_, map_, offset_, field_) \ 176 xe_map_rd_field(xe_, map_, offset_, struct xe_gsc_proxy_header, field_) 177 178 static u32 emit_proxy_header(struct xe_device *xe, struct iosys_map *map, u32 offset) 179 { 180 xe_map_memset(xe, map, offset, 0, PROXY_HDR_SIZE); 181 182 proxy_header_wr(xe, map, offset, hdr, 183 FIELD_PREP(GSC_PROXY_TYPE, GSC_PROXY_MSG_TYPE_PROXY_QUERY) | 184 FIELD_PREP(GSC_PROXY_PAYLOAD_LENGTH, 0)); 185 186 proxy_header_wr(xe, map, offset, source, GSC_PROXY_ADDRESSING_KMD); 187 proxy_header_wr(xe, map, offset, destination, GSC_PROXY_ADDRESSING_GSC); 188 proxy_header_wr(xe, map, offset, status, 0); 189 190 return offset + PROXY_HDR_SIZE; 191 } 192 193 static int proxy_query(struct xe_gsc *gsc) 194 { 195 struct xe_gt *gt = gsc_to_gt(gsc); 196 struct xe_device *xe = gt_to_xe(gt); 197 struct xe_gsc_proxy_header *to_csme_hdr = gsc->proxy.to_csme; 198 void *to_csme_payload = gsc->proxy.to_csme + PROXY_HDR_SIZE; 199 u32 wr_offset; 200 u32 reply_offset; 201 u32 size; 202 int ret; 203 204 wr_offset = xe_gsc_emit_header(xe, &gsc->proxy.to_gsc, 0, 205 HECI_MEADDRESS_PROXY, 0, PROXY_HDR_SIZE); 206 wr_offset = emit_proxy_header(xe, &gsc->proxy.to_gsc, wr_offset); 207 208 size = wr_offset; 209 210 while (1) { 211 /* 212 * Poison the GSC response header space to make sure we don't 213 * read a stale reply. 214 */ 215 xe_gsc_poison_header(xe, &gsc->proxy.from_gsc, 0); 216 217 /* send proxy message to GSC */ 218 ret = proxy_send_to_gsc(gsc, size); 219 if (ret) 220 goto proxy_error; 221 222 /* check the reply from GSC */ 223 ret = xe_gsc_read_out_header(xe, &gsc->proxy.from_gsc, 0, 224 PROXY_HDR_SIZE, &reply_offset); 225 if (ret) { 226 xe_gt_err(gt, "Invalid gsc header in proxy reply (%pe)\n", 227 ERR_PTR(ret)); 228 goto proxy_error; 229 } 230 231 /* copy the proxy header reply from GSC */ 232 xe_map_memcpy_from(xe, to_csme_hdr, &gsc->proxy.from_gsc, 233 reply_offset, PROXY_HDR_SIZE); 234 235 /* stop if this was the last message */ 236 if (FIELD_GET(GSC_PROXY_TYPE, to_csme_hdr->hdr) == GSC_PROXY_MSG_TYPE_PROXY_END) 237 break; 238 239 /* make sure the GSC-to-CSME proxy header is sane */ 240 ret = validate_proxy_header(to_csme_hdr, 241 GSC_PROXY_ADDRESSING_GSC, 242 GSC_PROXY_ADDRESSING_CSME, 243 GSC_PROXY_BUFFER_SIZE - reply_offset); 244 if (ret) { 245 xe_gt_err(gt, "invalid GSC to CSME proxy header! (%pe)\n", 246 ERR_PTR(ret)); 247 goto proxy_error; 248 } 249 250 /* copy the rest of the message */ 251 size = FIELD_GET(GSC_PROXY_PAYLOAD_LENGTH, to_csme_hdr->hdr); 252 xe_map_memcpy_from(xe, to_csme_payload, &gsc->proxy.from_gsc, 253 reply_offset + PROXY_HDR_SIZE, size); 254 255 /* send the GSC message to the CSME */ 256 ret = proxy_send_to_csme(gsc, size + PROXY_HDR_SIZE); 257 if (ret < 0) 258 goto proxy_error; 259 260 /* reply size from CSME, including the proxy header */ 261 size = ret; 262 if (size < PROXY_HDR_SIZE) { 263 xe_gt_err(gt, "CSME to GSC proxy msg too small: 0x%x\n", size); 264 ret = -EPROTO; 265 goto proxy_error; 266 } 267 268 /* make sure the CSME-to-GSC proxy header is sane */ 269 ret = validate_proxy_header(gsc->proxy.from_csme, 270 GSC_PROXY_ADDRESSING_CSME, 271 GSC_PROXY_ADDRESSING_GSC, 272 GSC_PROXY_BUFFER_SIZE - reply_offset); 273 if (ret) { 274 xe_gt_err(gt, "invalid CSME to GSC proxy header! %d\n", ret); 275 goto proxy_error; 276 } 277 278 /* Emit a new header for sending the reply to the GSC */ 279 wr_offset = xe_gsc_emit_header(xe, &gsc->proxy.to_gsc, 0, 280 HECI_MEADDRESS_PROXY, 0, size); 281 282 /* copy the CSME reply and update the total msg size to include the GSC header */ 283 xe_map_memcpy_to(xe, &gsc->proxy.to_gsc, wr_offset, gsc->proxy.from_csme, size); 284 285 size += wr_offset; 286 } 287 288 proxy_error: 289 return ret < 0 ? ret : 0; 290 } 291 292 int xe_gsc_proxy_request_handler(struct xe_gsc *gsc) 293 { 294 struct xe_gt *gt = gsc_to_gt(gsc); 295 int slept; 296 int err; 297 298 if (!gsc->proxy.component_added) 299 return -ENODEV; 300 301 /* when GSC is loaded, we can queue this before the component is bound */ 302 for (slept = 0; slept < GSC_PROXY_INIT_TIMEOUT_MS; slept += 100) { 303 if (gsc->proxy.component) 304 break; 305 306 msleep(100); 307 } 308 309 mutex_lock(&gsc->proxy.mutex); 310 if (!gsc->proxy.component) { 311 xe_gt_err(gt, "GSC proxy component not bound!\n"); 312 err = -EIO; 313 } else { 314 /* 315 * clear the pending interrupt and allow new proxy requests to 316 * be generated while we handle the current one 317 */ 318 gsc_proxy_irq_clear(gsc); 319 err = proxy_query(gsc); 320 } 321 mutex_unlock(&gsc->proxy.mutex); 322 return err; 323 } 324 325 void xe_gsc_proxy_irq_handler(struct xe_gsc *gsc, u32 iir) 326 { 327 struct xe_gt *gt = gsc_to_gt(gsc); 328 329 if (unlikely(!iir)) 330 return; 331 332 if (!gsc->proxy.component) { 333 xe_gt_err(gt, "GSC proxy irq received without the component being bound!\n"); 334 return; 335 } 336 337 spin_lock(&gsc->lock); 338 gsc->work_actions |= GSC_ACTION_SW_PROXY; 339 spin_unlock(&gsc->lock); 340 341 queue_work(gsc->wq, &gsc->work); 342 } 343 344 static int xe_gsc_proxy_component_bind(struct device *xe_kdev, 345 struct device *mei_kdev, void *data) 346 { 347 struct xe_device *xe = kdev_to_xe(xe_kdev); 348 struct xe_gt *gt = xe->tiles[0].media_gt; 349 struct xe_gsc *gsc = >->uc.gsc; 350 351 mutex_lock(&gsc->proxy.mutex); 352 gsc->proxy.component = data; 353 gsc->proxy.component->mei_dev = mei_kdev; 354 mutex_unlock(&gsc->proxy.mutex); 355 356 return 0; 357 } 358 359 static void xe_gsc_proxy_component_unbind(struct device *xe_kdev, 360 struct device *mei_kdev, void *data) 361 { 362 struct xe_device *xe = kdev_to_xe(xe_kdev); 363 struct xe_gt *gt = xe->tiles[0].media_gt; 364 struct xe_gsc *gsc = >->uc.gsc; 365 366 xe_gsc_wait_for_worker_completion(gsc); 367 368 mutex_lock(&gsc->proxy.mutex); 369 gsc->proxy.component = NULL; 370 mutex_unlock(&gsc->proxy.mutex); 371 } 372 373 static const struct component_ops xe_gsc_proxy_component_ops = { 374 .bind = xe_gsc_proxy_component_bind, 375 .unbind = xe_gsc_proxy_component_unbind, 376 }; 377 378 static void proxy_channel_free(struct drm_device *drm, void *arg) 379 { 380 struct xe_gsc *gsc = arg; 381 382 if (!gsc->proxy.bo) 383 return; 384 385 if (gsc->proxy.to_csme) { 386 kfree(gsc->proxy.to_csme); 387 gsc->proxy.to_csme = NULL; 388 gsc->proxy.from_csme = NULL; 389 } 390 391 if (gsc->proxy.bo) { 392 iosys_map_clear(&gsc->proxy.to_gsc); 393 iosys_map_clear(&gsc->proxy.from_gsc); 394 xe_bo_unpin_map_no_vm(gsc->proxy.bo); 395 gsc->proxy.bo = NULL; 396 } 397 } 398 399 static int proxy_channel_alloc(struct xe_gsc *gsc) 400 { 401 struct xe_gt *gt = gsc_to_gt(gsc); 402 struct xe_tile *tile = gt_to_tile(gt); 403 struct xe_device *xe = gt_to_xe(gt); 404 struct xe_bo *bo; 405 void *csme; 406 407 csme = kzalloc(GSC_PROXY_CHANNEL_SIZE, GFP_KERNEL); 408 if (!csme) 409 return -ENOMEM; 410 411 bo = xe_bo_create_pin_map(xe, tile, NULL, GSC_PROXY_CHANNEL_SIZE, 412 ttm_bo_type_kernel, 413 XE_BO_FLAG_SYSTEM | 414 XE_BO_FLAG_GGTT); 415 if (IS_ERR(bo)) { 416 kfree(csme); 417 return PTR_ERR(bo); 418 } 419 420 gsc->proxy.bo = bo; 421 gsc->proxy.to_gsc = IOSYS_MAP_INIT_OFFSET(&bo->vmap, 0); 422 gsc->proxy.from_gsc = IOSYS_MAP_INIT_OFFSET(&bo->vmap, GSC_PROXY_BUFFER_SIZE); 423 gsc->proxy.to_csme = csme; 424 gsc->proxy.from_csme = csme + GSC_PROXY_BUFFER_SIZE; 425 426 return drmm_add_action_or_reset(&xe->drm, proxy_channel_free, gsc); 427 } 428 429 /** 430 * xe_gsc_proxy_init() - init objects and MEI component required by GSC proxy 431 * @gsc: the GSC uC 432 * 433 * Return: 0 if the initialization was successful, a negative errno otherwise. 434 */ 435 int xe_gsc_proxy_init(struct xe_gsc *gsc) 436 { 437 int err; 438 struct xe_gt *gt = gsc_to_gt(gsc); 439 struct xe_tile *tile = gt_to_tile(gt); 440 struct xe_device *xe = tile_to_xe(tile); 441 442 mutex_init(&gsc->proxy.mutex); 443 444 if (!IS_ENABLED(CONFIG_INTEL_MEI_GSC_PROXY)) { 445 xe_gt_info(gt, "can't init GSC proxy due to missing mei component\n"); 446 return -ENODEV; 447 } 448 449 /* no multi-tile devices with this feature yet */ 450 if (tile->id > 0) { 451 xe_gt_err(gt, "unexpected GSC proxy init on tile %u\n", tile->id); 452 return -EINVAL; 453 } 454 455 err = proxy_channel_alloc(gsc); 456 if (err) 457 return err; 458 459 err = component_add_typed(xe->drm.dev, &xe_gsc_proxy_component_ops, 460 I915_COMPONENT_GSC_PROXY); 461 if (err < 0) { 462 xe_gt_err(gt, "Failed to add GSC_PROXY component (%pe)\n", ERR_PTR(err)); 463 return err; 464 } 465 466 gsc->proxy.component_added = true; 467 468 /* the component must be removed before unload, so can't use drmm for cleanup */ 469 470 return 0; 471 } 472 473 /** 474 * xe_gsc_proxy_remove() - remove the GSC proxy MEI component 475 * @gsc: the GSC uC 476 */ 477 void xe_gsc_proxy_remove(struct xe_gsc *gsc) 478 { 479 struct xe_gt *gt = gsc_to_gt(gsc); 480 struct xe_device *xe = gt_to_xe(gt); 481 int err = 0; 482 483 if (!gsc->proxy.component_added) 484 return; 485 486 /* disable HECI2 IRQs */ 487 xe_pm_runtime_get(xe); 488 err = xe_force_wake_get(gt_to_fw(gt), XE_FW_GSC); 489 if (err) 490 xe_gt_err(gt, "failed to get forcewake to disable GSC interrupts\n"); 491 492 /* try do disable irq even if forcewake failed */ 493 gsc_proxy_irq_toggle(gsc, false); 494 495 if (!err) 496 xe_force_wake_put(gt_to_fw(gt), XE_FW_GSC); 497 xe_pm_runtime_put(xe); 498 499 xe_gsc_wait_for_worker_completion(gsc); 500 501 component_del(xe->drm.dev, &xe_gsc_proxy_component_ops); 502 gsc->proxy.component_added = false; 503 } 504 505 /** 506 * xe_gsc_proxy_start() - start the proxy by submitting the first request 507 * @gsc: the GSC uC 508 * 509 * Return: 0 if the proxy are now enabled, a negative errno otherwise. 510 */ 511 int xe_gsc_proxy_start(struct xe_gsc *gsc) 512 { 513 int err; 514 515 /* enable the proxy interrupt in the GSC shim layer */ 516 gsc_proxy_irq_toggle(gsc, true); 517 518 /* 519 * The handling of the first proxy request must be manually triggered to 520 * notify the GSC that we're ready to support the proxy flow. 521 */ 522 err = xe_gsc_proxy_request_handler(gsc); 523 if (err) 524 return err; 525 526 if (!xe_gsc_proxy_init_done(gsc)) { 527 xe_gt_err(gsc_to_gt(gsc), "GSC FW reports proxy init not completed\n"); 528 return -EIO; 529 } 530 531 return 0; 532 } 533