1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Copyright 2016-17 IBM Corp. 4 */ 5 6 #define pr_fmt(fmt) "vas: " fmt 7 8 #include <linux/types.h> 9 #include <linux/mutex.h> 10 #include <linux/slab.h> 11 #include <linux/io.h> 12 #include <linux/log2.h> 13 #include <linux/rcupdate.h> 14 #include <linux/cred.h> 15 #include <linux/sched/mm.h> 16 #include <linux/mmu_context.h> 17 #include <asm/switch_to.h> 18 #include <asm/ppc-opcode.h> 19 #include <asm/vas.h> 20 #include "vas.h" 21 #include "copy-paste.h" 22 23 #define CREATE_TRACE_POINTS 24 #include "vas-trace.h" 25 26 /* 27 * Compute the paste address region for the window @window using the 28 * ->paste_base_addr and ->paste_win_id_shift we got from device tree. 29 */ 30 void vas_win_paste_addr(struct pnv_vas_window *window, u64 *addr, int *len) 31 { 32 int winid; 33 u64 base, shift; 34 35 base = window->vinst->paste_base_addr; 36 shift = window->vinst->paste_win_id_shift; 37 winid = window->vas_win.winid; 38 39 *addr = base + (winid << shift); 40 if (len) 41 *len = PAGE_SIZE; 42 43 pr_debug("Txwin #%d: Paste addr 0x%llx\n", winid, *addr); 44 } 45 46 static inline void get_hvwc_mmio_bar(struct pnv_vas_window *window, 47 u64 *start, int *len) 48 { 49 u64 pbaddr; 50 51 pbaddr = window->vinst->hvwc_bar_start; 52 *start = pbaddr + window->vas_win.winid * VAS_HVWC_SIZE; 53 *len = VAS_HVWC_SIZE; 54 } 55 56 static inline void get_uwc_mmio_bar(struct pnv_vas_window *window, 57 u64 *start, int *len) 58 { 59 u64 pbaddr; 60 61 pbaddr = window->vinst->uwc_bar_start; 62 *start = pbaddr + window->vas_win.winid * VAS_UWC_SIZE; 63 *len = VAS_UWC_SIZE; 64 } 65 66 /* 67 * Map the paste bus address of the given send window into kernel address 68 * space. Unlike MMIO regions (map_mmio_region() below), paste region must 69 * be mapped cache-able and is only applicable to send windows. 70 */ 71 static void *map_paste_region(struct pnv_vas_window *txwin) 72 { 73 int len; 74 void *map; 75 char *name; 76 u64 start; 77 78 name = kasprintf(GFP_KERNEL, "window-v%d-w%d", txwin->vinst->vas_id, 79 txwin->vas_win.winid); 80 if (!name) 81 goto free_name; 82 83 txwin->paste_addr_name = name; 84 vas_win_paste_addr(txwin, &start, &len); 85 86 if (!request_mem_region(start, len, name)) { 87 pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n", 88 __func__, start, len); 89 goto free_name; 90 } 91 92 map = ioremap_cache(start, len); 93 if (!map) { 94 pr_devel("%s(): ioremap_cache(0x%llx, %d) failed\n", __func__, 95 start, len); 96 goto free_name; 97 } 98 99 pr_devel("Mapped paste addr 0x%llx to kaddr 0x%p\n", start, map); 100 return map; 101 102 free_name: 103 kfree(name); 104 return ERR_PTR(-ENOMEM); 105 } 106 107 static void *map_mmio_region(char *name, u64 start, int len) 108 { 109 void *map; 110 111 if (!request_mem_region(start, len, name)) { 112 pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n", 113 __func__, start, len); 114 return NULL; 115 } 116 117 map = ioremap(start, len); 118 if (!map) { 119 pr_devel("%s(): ioremap(0x%llx, %d) failed\n", __func__, start, 120 len); 121 return NULL; 122 } 123 124 return map; 125 } 126 127 static void unmap_region(void *addr, u64 start, int len) 128 { 129 iounmap(addr); 130 release_mem_region((phys_addr_t)start, len); 131 } 132 133 /* 134 * Unmap the paste address region for a window. 135 */ 136 static void unmap_paste_region(struct pnv_vas_window *window) 137 { 138 int len; 139 u64 busaddr_start; 140 141 if (window->paste_kaddr) { 142 vas_win_paste_addr(window, &busaddr_start, &len); 143 unmap_region(window->paste_kaddr, busaddr_start, len); 144 window->paste_kaddr = NULL; 145 kfree(window->paste_addr_name); 146 window->paste_addr_name = NULL; 147 } 148 } 149 150 /* 151 * Unmap the MMIO regions for a window. Hold the vas_mutex so we don't 152 * unmap when the window's debugfs dir is in use. This serializes close 153 * of a window even on another VAS instance but since its not a critical 154 * path, just minimize the time we hold the mutex for now. We can add 155 * a per-instance mutex later if necessary. 156 */ 157 static void unmap_winctx_mmio_bars(struct pnv_vas_window *window) 158 { 159 int len; 160 void *uwc_map; 161 void *hvwc_map; 162 u64 busaddr_start; 163 164 mutex_lock(&vas_mutex); 165 166 hvwc_map = window->hvwc_map; 167 window->hvwc_map = NULL; 168 169 uwc_map = window->uwc_map; 170 window->uwc_map = NULL; 171 172 mutex_unlock(&vas_mutex); 173 174 if (hvwc_map) { 175 get_hvwc_mmio_bar(window, &busaddr_start, &len); 176 unmap_region(hvwc_map, busaddr_start, len); 177 } 178 179 if (uwc_map) { 180 get_uwc_mmio_bar(window, &busaddr_start, &len); 181 unmap_region(uwc_map, busaddr_start, len); 182 } 183 } 184 185 /* 186 * Find the Hypervisor Window Context (HVWC) MMIO Base Address Region and the 187 * OS/User Window Context (UWC) MMIO Base Address Region for the given window. 188 * Map these bus addresses and save the mapped kernel addresses in @window. 189 */ 190 static int map_winctx_mmio_bars(struct pnv_vas_window *window) 191 { 192 int len; 193 u64 start; 194 195 get_hvwc_mmio_bar(window, &start, &len); 196 window->hvwc_map = map_mmio_region("HVWCM_Window", start, len); 197 198 get_uwc_mmio_bar(window, &start, &len); 199 window->uwc_map = map_mmio_region("UWCM_Window", start, len); 200 201 if (!window->hvwc_map || !window->uwc_map) { 202 unmap_winctx_mmio_bars(window); 203 return -1; 204 } 205 206 return 0; 207 } 208 209 /* 210 * Reset all valid registers in the HV and OS/User Window Contexts for 211 * the window identified by @window. 212 * 213 * NOTE: We cannot really use a for loop to reset window context. Not all 214 * offsets in a window context are valid registers and the valid 215 * registers are not sequential. And, we can only write to offsets 216 * with valid registers. 217 */ 218 static void reset_window_regs(struct pnv_vas_window *window) 219 { 220 write_hvwc_reg(window, VREG(LPID), 0ULL); 221 write_hvwc_reg(window, VREG(PID), 0ULL); 222 write_hvwc_reg(window, VREG(XLATE_MSR), 0ULL); 223 write_hvwc_reg(window, VREG(XLATE_LPCR), 0ULL); 224 write_hvwc_reg(window, VREG(XLATE_CTL), 0ULL); 225 write_hvwc_reg(window, VREG(AMR), 0ULL); 226 write_hvwc_reg(window, VREG(SEIDR), 0ULL); 227 write_hvwc_reg(window, VREG(FAULT_TX_WIN), 0ULL); 228 write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL); 229 write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), 0ULL); 230 write_hvwc_reg(window, VREG(PSWID), 0ULL); 231 write_hvwc_reg(window, VREG(LFIFO_BAR), 0ULL); 232 write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), 0ULL); 233 write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), 0ULL); 234 write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL); 235 write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL); 236 write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL); 237 write_hvwc_reg(window, VREG(LRX_WCRED), 0ULL); 238 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); 239 write_hvwc_reg(window, VREG(TX_WCRED), 0ULL); 240 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); 241 write_hvwc_reg(window, VREG(LFIFO_SIZE), 0ULL); 242 write_hvwc_reg(window, VREG(WINCTL), 0ULL); 243 write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL); 244 write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), 0ULL); 245 write_hvwc_reg(window, VREG(TX_RSVD_BUF_COUNT), 0ULL); 246 write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), 0ULL); 247 write_hvwc_reg(window, VREG(LNOTIFY_CTL), 0ULL); 248 write_hvwc_reg(window, VREG(LNOTIFY_PID), 0ULL); 249 write_hvwc_reg(window, VREG(LNOTIFY_LPID), 0ULL); 250 write_hvwc_reg(window, VREG(LNOTIFY_TID), 0ULL); 251 write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), 0ULL); 252 write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL); 253 254 /* Skip read-only registers: NX_UTIL and NX_UTIL_SE */ 255 256 /* 257 * The send and receive window credit adder registers are also 258 * accessible from HVWC and have been initialized above. We don't 259 * need to initialize from the OS/User Window Context, so skip 260 * following calls: 261 * 262 * write_uwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); 263 * write_uwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); 264 */ 265 } 266 267 /* 268 * Initialize window context registers related to Address Translation. 269 * These registers are common to send/receive windows although they 270 * differ for user/kernel windows. As we resolve the TODOs we may 271 * want to add fields to vas_winctx and move the initialization to 272 * init_vas_winctx_regs(). 273 */ 274 static void init_xlate_regs(struct pnv_vas_window *window, bool user_win) 275 { 276 u64 lpcr, val; 277 278 /* 279 * MSR_TA, MSR_US are false for both kernel and user. 280 * MSR_DR and MSR_PR are false for kernel. 281 */ 282 val = 0ULL; 283 val = SET_FIELD(VAS_XLATE_MSR_HV, val, 1); 284 val = SET_FIELD(VAS_XLATE_MSR_SF, val, 1); 285 if (user_win) { 286 val = SET_FIELD(VAS_XLATE_MSR_DR, val, 1); 287 val = SET_FIELD(VAS_XLATE_MSR_PR, val, 1); 288 } 289 write_hvwc_reg(window, VREG(XLATE_MSR), val); 290 291 lpcr = mfspr(SPRN_LPCR); 292 val = 0ULL; 293 /* 294 * NOTE: From Section 5.7.8.1 Segment Lookaside Buffer of the 295 * Power ISA, v3.0B, Page size encoding is 0 = 4KB, 5 = 64KB. 296 * 297 * NOTE: From Section 1.3.1, Address Translation Context of the 298 * Nest MMU Workbook, LPCR_SC should be 0 for Power9. 299 */ 300 val = SET_FIELD(VAS_XLATE_LPCR_PAGE_SIZE, val, 5); 301 val = SET_FIELD(VAS_XLATE_LPCR_ISL, val, lpcr & LPCR_ISL); 302 val = SET_FIELD(VAS_XLATE_LPCR_TC, val, lpcr & LPCR_TC); 303 val = SET_FIELD(VAS_XLATE_LPCR_SC, val, 0); 304 write_hvwc_reg(window, VREG(XLATE_LPCR), val); 305 306 /* 307 * Section 1.3.1 (Address translation Context) of NMMU workbook. 308 * 0b00 Hashed Page Table mode 309 * 0b01 Reserved 310 * 0b10 Radix on HPT 311 * 0b11 Radix on Radix 312 */ 313 val = 0ULL; 314 val = SET_FIELD(VAS_XLATE_MODE, val, radix_enabled() ? 3 : 2); 315 write_hvwc_reg(window, VREG(XLATE_CTL), val); 316 317 /* 318 * TODO: Can we mfspr(AMR) even for user windows? 319 */ 320 val = 0ULL; 321 val = SET_FIELD(VAS_AMR, val, mfspr(SPRN_AMR)); 322 write_hvwc_reg(window, VREG(AMR), val); 323 324 val = 0ULL; 325 val = SET_FIELD(VAS_SEIDR, val, 0); 326 write_hvwc_reg(window, VREG(SEIDR), val); 327 } 328 329 /* 330 * Initialize Reserved Send Buffer Count for the send window. It involves 331 * writing to the register, reading it back to confirm that the hardware 332 * has enough buffers to reserve. See section 1.3.1.2.1 of VAS workbook. 333 * 334 * Since we can only make a best-effort attempt to fulfill the request, 335 * we don't return any errors if we cannot. 336 * 337 * TODO: Reserved (aka dedicated) send buffers are not supported yet. 338 */ 339 static void init_rsvd_tx_buf_count(struct pnv_vas_window *txwin, 340 struct vas_winctx *winctx) 341 { 342 write_hvwc_reg(txwin, VREG(TX_RSVD_BUF_COUNT), 0ULL); 343 } 344 345 /* 346 * init_winctx_regs() 347 * Initialize window context registers for a receive window. 348 * Except for caching control and marking window open, the registers 349 * are initialized in the order listed in Section 3.1.4 (Window Context 350 * Cache Register Details) of the VAS workbook although they don't need 351 * to be. 352 * 353 * Design note: For NX receive windows, NX allocates the FIFO buffer in OPAL 354 * (so that it can get a large contiguous area) and passes that buffer 355 * to kernel via device tree. We now write that buffer address to the 356 * FIFO BAR. Would it make sense to do this all in OPAL? i.e have OPAL 357 * write the per-chip RX FIFO addresses to the windows during boot-up 358 * as a one-time task? That could work for NX but what about other 359 * receivers? Let the receivers tell us the rx-fifo buffers for now. 360 */ 361 static void init_winctx_regs(struct pnv_vas_window *window, 362 struct vas_winctx *winctx) 363 { 364 u64 val; 365 int fifo_size; 366 367 reset_window_regs(window); 368 369 val = 0ULL; 370 val = SET_FIELD(VAS_LPID, val, winctx->lpid); 371 write_hvwc_reg(window, VREG(LPID), val); 372 373 val = 0ULL; 374 val = SET_FIELD(VAS_PID_ID, val, winctx->pidr); 375 write_hvwc_reg(window, VREG(PID), val); 376 377 init_xlate_regs(window, winctx->user_win); 378 379 val = 0ULL; 380 val = SET_FIELD(VAS_FAULT_TX_WIN, val, winctx->fault_win_id); 381 write_hvwc_reg(window, VREG(FAULT_TX_WIN), val); 382 383 /* In PowerNV, interrupts go to HV. */ 384 write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL); 385 386 val = 0ULL; 387 val = SET_FIELD(VAS_HV_INTR_SRC_RA, val, winctx->irq_port); 388 write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), val); 389 390 val = 0ULL; 391 val = SET_FIELD(VAS_PSWID_EA_HANDLE, val, winctx->pswid); 392 write_hvwc_reg(window, VREG(PSWID), val); 393 394 write_hvwc_reg(window, VREG(SPARE1), 0ULL); 395 write_hvwc_reg(window, VREG(SPARE2), 0ULL); 396 write_hvwc_reg(window, VREG(SPARE3), 0ULL); 397 398 /* 399 * NOTE: VAS expects the FIFO address to be copied into the LFIFO_BAR 400 * register as is - do NOT shift the address into VAS_LFIFO_BAR 401 * bit fields! Ok to set the page migration select fields - 402 * VAS ignores the lower 10+ bits in the address anyway, because 403 * the minimum FIFO size is 1K? 404 * 405 * See also: Design note in function header. 406 */ 407 val = winctx->rx_fifo; 408 val = SET_FIELD(VAS_PAGE_MIGRATION_SELECT, val, 0); 409 write_hvwc_reg(window, VREG(LFIFO_BAR), val); 410 411 val = 0ULL; 412 val = SET_FIELD(VAS_LDATA_STAMP, val, winctx->data_stamp); 413 write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), val); 414 415 val = 0ULL; 416 val = SET_FIELD(VAS_LDMA_TYPE, val, winctx->dma_type); 417 val = SET_FIELD(VAS_LDMA_FIFO_DISABLE, val, winctx->fifo_disable); 418 write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), val); 419 420 write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL); 421 write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL); 422 write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL); 423 424 val = 0ULL; 425 val = SET_FIELD(VAS_LRX_WCRED, val, winctx->wcreds_max); 426 write_hvwc_reg(window, VREG(LRX_WCRED), val); 427 428 val = 0ULL; 429 val = SET_FIELD(VAS_TX_WCRED, val, winctx->wcreds_max); 430 write_hvwc_reg(window, VREG(TX_WCRED), val); 431 432 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); 433 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); 434 435 fifo_size = winctx->rx_fifo_size / 1024; 436 437 val = 0ULL; 438 val = SET_FIELD(VAS_LFIFO_SIZE, val, ilog2(fifo_size)); 439 write_hvwc_reg(window, VREG(LFIFO_SIZE), val); 440 441 /* Update window control and caching control registers last so 442 * we mark the window open only after fully initializing it and 443 * pushing context to cache. 444 */ 445 446 write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL); 447 448 init_rsvd_tx_buf_count(window, winctx); 449 450 /* for a send window, point to the matching receive window */ 451 val = 0ULL; 452 val = SET_FIELD(VAS_LRX_WIN_ID, val, winctx->rx_win_id); 453 write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), val); 454 455 write_hvwc_reg(window, VREG(SPARE4), 0ULL); 456 457 val = 0ULL; 458 val = SET_FIELD(VAS_NOTIFY_DISABLE, val, winctx->notify_disable); 459 val = SET_FIELD(VAS_INTR_DISABLE, val, winctx->intr_disable); 460 val = SET_FIELD(VAS_NOTIFY_EARLY, val, winctx->notify_early); 461 val = SET_FIELD(VAS_NOTIFY_OSU_INTR, val, winctx->notify_os_intr_reg); 462 write_hvwc_reg(window, VREG(LNOTIFY_CTL), val); 463 464 val = 0ULL; 465 val = SET_FIELD(VAS_LNOTIFY_PID, val, winctx->lnotify_pid); 466 write_hvwc_reg(window, VREG(LNOTIFY_PID), val); 467 468 val = 0ULL; 469 val = SET_FIELD(VAS_LNOTIFY_LPID, val, winctx->lnotify_lpid); 470 write_hvwc_reg(window, VREG(LNOTIFY_LPID), val); 471 472 val = 0ULL; 473 val = SET_FIELD(VAS_LNOTIFY_TID, val, winctx->lnotify_tid); 474 write_hvwc_reg(window, VREG(LNOTIFY_TID), val); 475 476 val = 0ULL; 477 val = SET_FIELD(VAS_LNOTIFY_MIN_SCOPE, val, winctx->min_scope); 478 val = SET_FIELD(VAS_LNOTIFY_MAX_SCOPE, val, winctx->max_scope); 479 write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), val); 480 481 /* Skip read-only registers NX_UTIL and NX_UTIL_SE */ 482 483 write_hvwc_reg(window, VREG(SPARE5), 0ULL); 484 write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL); 485 write_hvwc_reg(window, VREG(SPARE6), 0ULL); 486 487 /* Finally, push window context to memory and... */ 488 val = 0ULL; 489 val = SET_FIELD(VAS_PUSH_TO_MEM, val, 1); 490 write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), val); 491 492 /* ... mark the window open for business */ 493 val = 0ULL; 494 val = SET_FIELD(VAS_WINCTL_REJ_NO_CREDIT, val, winctx->rej_no_credit); 495 val = SET_FIELD(VAS_WINCTL_PIN, val, winctx->pin_win); 496 val = SET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val, winctx->tx_wcred_mode); 497 val = SET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val, winctx->rx_wcred_mode); 498 val = SET_FIELD(VAS_WINCTL_TX_WORD_MODE, val, winctx->tx_word_mode); 499 val = SET_FIELD(VAS_WINCTL_RX_WORD_MODE, val, winctx->rx_word_mode); 500 val = SET_FIELD(VAS_WINCTL_FAULT_WIN, val, winctx->fault_win); 501 val = SET_FIELD(VAS_WINCTL_NX_WIN, val, winctx->nx_win); 502 val = SET_FIELD(VAS_WINCTL_OPEN, val, 1); 503 write_hvwc_reg(window, VREG(WINCTL), val); 504 } 505 506 static void vas_release_window_id(struct ida *ida, int winid) 507 { 508 ida_free(ida, winid); 509 } 510 511 static int vas_assign_window_id(struct ida *ida) 512 { 513 int winid = ida_alloc_max(ida, VAS_WINDOWS_PER_CHIP - 1, GFP_KERNEL); 514 515 if (winid == -ENOSPC) { 516 pr_err("Too many (%d) open windows\n", VAS_WINDOWS_PER_CHIP); 517 return -EAGAIN; 518 } 519 520 return winid; 521 } 522 523 static void vas_window_free(struct pnv_vas_window *window) 524 { 525 struct vas_instance *vinst = window->vinst; 526 int winid = window->vas_win.winid; 527 528 unmap_winctx_mmio_bars(window); 529 530 vas_window_free_dbgdir(window); 531 532 kfree(window); 533 534 vas_release_window_id(&vinst->ida, winid); 535 } 536 537 static struct pnv_vas_window *vas_window_alloc(struct vas_instance *vinst) 538 { 539 int winid; 540 struct pnv_vas_window *window; 541 542 winid = vas_assign_window_id(&vinst->ida); 543 if (winid < 0) 544 return ERR_PTR(winid); 545 546 window = kzalloc(sizeof(*window), GFP_KERNEL); 547 if (!window) 548 goto out_free; 549 550 window->vinst = vinst; 551 window->vas_win.winid = winid; 552 553 if (map_winctx_mmio_bars(window)) 554 goto out_free; 555 556 vas_window_init_dbgdir(window); 557 558 return window; 559 560 out_free: 561 kfree(window); 562 vas_release_window_id(&vinst->ida, winid); 563 return ERR_PTR(-ENOMEM); 564 } 565 566 static void put_rx_win(struct pnv_vas_window *rxwin) 567 { 568 /* Better not be a send window! */ 569 WARN_ON_ONCE(rxwin->tx_win); 570 571 atomic_dec(&rxwin->num_txwins); 572 } 573 574 /* 575 * Find the user space receive window given the @pswid. 576 * - We must have a valid vasid and it must belong to this instance. 577 * (so both send and receive windows are on the same VAS instance) 578 * - The window must refer to an OPEN, FTW, RECEIVE window. 579 * 580 * NOTE: We access ->windows[] table and assume that vinst->mutex is held. 581 */ 582 static struct pnv_vas_window *get_user_rxwin(struct vas_instance *vinst, 583 u32 pswid) 584 { 585 int vasid, winid; 586 struct pnv_vas_window *rxwin; 587 588 decode_pswid(pswid, &vasid, &winid); 589 590 if (vinst->vas_id != vasid) 591 return ERR_PTR(-EINVAL); 592 593 rxwin = vinst->windows[winid]; 594 595 if (!rxwin || rxwin->tx_win || rxwin->vas_win.cop != VAS_COP_TYPE_FTW) 596 return ERR_PTR(-EINVAL); 597 598 return rxwin; 599 } 600 601 /* 602 * Get the VAS receive window associated with NX engine identified 603 * by @cop and if applicable, @pswid. 604 * 605 * See also function header of set_vinst_win(). 606 */ 607 static struct pnv_vas_window *get_vinst_rxwin(struct vas_instance *vinst, 608 enum vas_cop_type cop, u32 pswid) 609 { 610 struct pnv_vas_window *rxwin; 611 612 mutex_lock(&vinst->mutex); 613 614 if (cop == VAS_COP_TYPE_FTW) 615 rxwin = get_user_rxwin(vinst, pswid); 616 else 617 rxwin = vinst->rxwin[cop] ?: ERR_PTR(-EINVAL); 618 619 if (!IS_ERR(rxwin)) 620 atomic_inc(&rxwin->num_txwins); 621 622 mutex_unlock(&vinst->mutex); 623 624 return rxwin; 625 } 626 627 /* 628 * We have two tables of windows in a VAS instance. The first one, 629 * ->windows[], contains all the windows in the instance and allows 630 * looking up a window by its id. It is used to look up send windows 631 * during fault handling and receive windows when pairing user space 632 * send/receive windows. 633 * 634 * The second table, ->rxwin[], contains receive windows that are 635 * associated with NX engines. This table has VAS_COP_TYPE_MAX 636 * entries and is used to look up a receive window by its 637 * coprocessor type. 638 * 639 * Here, we save @window in the ->windows[] table. If it is a receive 640 * window, we also save the window in the ->rxwin[] table. 641 */ 642 static void set_vinst_win(struct vas_instance *vinst, 643 struct pnv_vas_window *window) 644 { 645 int id = window->vas_win.winid; 646 647 mutex_lock(&vinst->mutex); 648 649 /* 650 * There should only be one receive window for a coprocessor type 651 * unless its a user (FTW) window. 652 */ 653 if (!window->user_win && !window->tx_win) { 654 WARN_ON_ONCE(vinst->rxwin[window->vas_win.cop]); 655 vinst->rxwin[window->vas_win.cop] = window; 656 } 657 658 WARN_ON_ONCE(vinst->windows[id] != NULL); 659 vinst->windows[id] = window; 660 661 mutex_unlock(&vinst->mutex); 662 } 663 664 /* 665 * Clear this window from the table(s) of windows for this VAS instance. 666 * See also function header of set_vinst_win(). 667 */ 668 static void clear_vinst_win(struct pnv_vas_window *window) 669 { 670 int id = window->vas_win.winid; 671 struct vas_instance *vinst = window->vinst; 672 673 mutex_lock(&vinst->mutex); 674 675 if (!window->user_win && !window->tx_win) { 676 WARN_ON_ONCE(!vinst->rxwin[window->vas_win.cop]); 677 vinst->rxwin[window->vas_win.cop] = NULL; 678 } 679 680 WARN_ON_ONCE(vinst->windows[id] != window); 681 vinst->windows[id] = NULL; 682 683 mutex_unlock(&vinst->mutex); 684 } 685 686 static void init_winctx_for_rxwin(struct pnv_vas_window *rxwin, 687 struct vas_rx_win_attr *rxattr, 688 struct vas_winctx *winctx) 689 { 690 /* 691 * We first zero (memset()) all fields and only set non-zero fields. 692 * Following fields are 0/false but maybe deserve a comment: 693 * 694 * ->notify_os_intr_reg In powerNV, send intrs to HV 695 * ->notify_disable False for NX windows 696 * ->intr_disable False for Fault Windows 697 * ->xtra_write False for NX windows 698 * ->notify_early NA for NX windows 699 * ->rsvd_txbuf_count NA for Rx windows 700 * ->lpid, ->pid, ->tid NA for Rx windows 701 */ 702 703 memset(winctx, 0, sizeof(struct vas_winctx)); 704 705 winctx->rx_fifo = rxattr->rx_fifo; 706 winctx->rx_fifo_size = rxattr->rx_fifo_size; 707 winctx->wcreds_max = rxwin->vas_win.wcreds_max; 708 winctx->pin_win = rxattr->pin_win; 709 710 winctx->nx_win = rxattr->nx_win; 711 winctx->fault_win = rxattr->fault_win; 712 winctx->user_win = rxattr->user_win; 713 winctx->rej_no_credit = rxattr->rej_no_credit; 714 winctx->rx_word_mode = rxattr->rx_win_ord_mode; 715 winctx->tx_word_mode = rxattr->tx_win_ord_mode; 716 winctx->rx_wcred_mode = rxattr->rx_wcred_mode; 717 winctx->tx_wcred_mode = rxattr->tx_wcred_mode; 718 winctx->notify_early = rxattr->notify_early; 719 720 if (winctx->nx_win) { 721 winctx->data_stamp = true; 722 winctx->intr_disable = true; 723 winctx->pin_win = true; 724 725 WARN_ON_ONCE(winctx->fault_win); 726 WARN_ON_ONCE(!winctx->rx_word_mode); 727 WARN_ON_ONCE(!winctx->tx_word_mode); 728 WARN_ON_ONCE(winctx->notify_after_count); 729 } else if (winctx->fault_win) { 730 winctx->notify_disable = true; 731 } else if (winctx->user_win) { 732 /* 733 * Section 1.8.1 Low Latency Core-Core Wake up of 734 * the VAS workbook: 735 * 736 * - disable credit checks ([tr]x_wcred_mode = false) 737 * - disable FIFO writes 738 * - enable ASB_Notify, disable interrupt 739 */ 740 winctx->fifo_disable = true; 741 winctx->intr_disable = true; 742 winctx->rx_fifo = 0; 743 } 744 745 winctx->lnotify_lpid = rxattr->lnotify_lpid; 746 winctx->lnotify_pid = rxattr->lnotify_pid; 747 winctx->lnotify_tid = rxattr->lnotify_tid; 748 winctx->pswid = rxattr->pswid; 749 winctx->dma_type = VAS_DMA_TYPE_INJECT; 750 winctx->tc_mode = rxattr->tc_mode; 751 752 winctx->min_scope = VAS_SCOPE_LOCAL; 753 winctx->max_scope = VAS_SCOPE_VECTORED_GROUP; 754 if (rxwin->vinst->virq) 755 winctx->irq_port = rxwin->vinst->irq_port; 756 } 757 758 static bool rx_win_args_valid(enum vas_cop_type cop, 759 struct vas_rx_win_attr *attr) 760 { 761 pr_debug("Rxattr: fault %d, notify %d, intr %d, early %d, fifo %d\n", 762 attr->fault_win, attr->notify_disable, 763 attr->intr_disable, attr->notify_early, 764 attr->rx_fifo_size); 765 766 if (cop >= VAS_COP_TYPE_MAX) 767 return false; 768 769 if (cop != VAS_COP_TYPE_FTW && 770 attr->rx_fifo_size < VAS_RX_FIFO_SIZE_MIN) 771 return false; 772 773 if (attr->rx_fifo_size > VAS_RX_FIFO_SIZE_MAX) 774 return false; 775 776 if (!attr->wcreds_max) 777 return false; 778 779 if (attr->nx_win) { 780 /* cannot be fault or user window if it is nx */ 781 if (attr->fault_win || attr->user_win) 782 return false; 783 /* 784 * Section 3.1.4.32: NX Windows must not disable notification, 785 * and must not enable interrupts or early notification. 786 */ 787 if (attr->notify_disable || !attr->intr_disable || 788 attr->notify_early) 789 return false; 790 } else if (attr->fault_win) { 791 /* cannot be both fault and user window */ 792 if (attr->user_win) 793 return false; 794 795 /* 796 * Section 3.1.4.32: Fault windows must disable notification 797 * but not interrupts. 798 */ 799 if (!attr->notify_disable || attr->intr_disable) 800 return false; 801 802 } else if (attr->user_win) { 803 /* 804 * User receive windows are only for fast-thread-wakeup 805 * (FTW). They don't need a FIFO and must disable interrupts 806 */ 807 if (attr->rx_fifo || attr->rx_fifo_size || !attr->intr_disable) 808 return false; 809 } else { 810 /* Rx window must be one of NX or Fault or User window. */ 811 return false; 812 } 813 814 return true; 815 } 816 817 void vas_init_rx_win_attr(struct vas_rx_win_attr *rxattr, enum vas_cop_type cop) 818 { 819 memset(rxattr, 0, sizeof(*rxattr)); 820 821 if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI || 822 cop == VAS_COP_TYPE_GZIP || cop == VAS_COP_TYPE_GZIP_HIPRI) { 823 rxattr->pin_win = true; 824 rxattr->nx_win = true; 825 rxattr->fault_win = false; 826 rxattr->intr_disable = true; 827 rxattr->rx_wcred_mode = true; 828 rxattr->tx_wcred_mode = true; 829 rxattr->rx_win_ord_mode = true; 830 rxattr->tx_win_ord_mode = true; 831 } else if (cop == VAS_COP_TYPE_FAULT) { 832 rxattr->pin_win = true; 833 rxattr->fault_win = true; 834 rxattr->notify_disable = true; 835 rxattr->rx_wcred_mode = true; 836 rxattr->rx_win_ord_mode = true; 837 rxattr->rej_no_credit = true; 838 rxattr->tc_mode = VAS_THRESH_DISABLED; 839 } else if (cop == VAS_COP_TYPE_FTW) { 840 rxattr->user_win = true; 841 rxattr->intr_disable = true; 842 843 /* 844 * As noted in the VAS Workbook we disable credit checks. 845 * If we enable credit checks in the future, we must also 846 * implement a mechanism to return the user credits or new 847 * paste operations will fail. 848 */ 849 } 850 } 851 EXPORT_SYMBOL_GPL(vas_init_rx_win_attr); 852 853 struct vas_window *vas_rx_win_open(int vasid, enum vas_cop_type cop, 854 struct vas_rx_win_attr *rxattr) 855 { 856 struct pnv_vas_window *rxwin; 857 struct vas_winctx winctx; 858 struct vas_instance *vinst; 859 860 trace_vas_rx_win_open(current, vasid, cop, rxattr); 861 862 if (!rx_win_args_valid(cop, rxattr)) 863 return ERR_PTR(-EINVAL); 864 865 vinst = find_vas_instance(vasid); 866 if (!vinst) { 867 pr_devel("vasid %d not found!\n", vasid); 868 return ERR_PTR(-EINVAL); 869 } 870 pr_devel("Found instance %d\n", vasid); 871 872 rxwin = vas_window_alloc(vinst); 873 if (IS_ERR(rxwin)) { 874 pr_devel("Unable to allocate memory for Rx window\n"); 875 return (struct vas_window *)rxwin; 876 } 877 878 rxwin->tx_win = false; 879 rxwin->nx_win = rxattr->nx_win; 880 rxwin->user_win = rxattr->user_win; 881 rxwin->vas_win.cop = cop; 882 rxwin->vas_win.wcreds_max = rxattr->wcreds_max; 883 884 init_winctx_for_rxwin(rxwin, rxattr, &winctx); 885 init_winctx_regs(rxwin, &winctx); 886 887 set_vinst_win(vinst, rxwin); 888 889 return &rxwin->vas_win; 890 } 891 EXPORT_SYMBOL_GPL(vas_rx_win_open); 892 893 void vas_init_tx_win_attr(struct vas_tx_win_attr *txattr, enum vas_cop_type cop) 894 { 895 memset(txattr, 0, sizeof(*txattr)); 896 897 if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI || 898 cop == VAS_COP_TYPE_GZIP || cop == VAS_COP_TYPE_GZIP_HIPRI) { 899 txattr->rej_no_credit = false; 900 txattr->rx_wcred_mode = true; 901 txattr->tx_wcred_mode = true; 902 txattr->rx_win_ord_mode = true; 903 txattr->tx_win_ord_mode = true; 904 } else if (cop == VAS_COP_TYPE_FTW) { 905 txattr->user_win = true; 906 } 907 } 908 EXPORT_SYMBOL_GPL(vas_init_tx_win_attr); 909 910 static void init_winctx_for_txwin(struct pnv_vas_window *txwin, 911 struct vas_tx_win_attr *txattr, 912 struct vas_winctx *winctx) 913 { 914 /* 915 * We first zero all fields and only set non-zero ones. Following 916 * are some fields set to 0/false for the stated reason: 917 * 918 * ->notify_os_intr_reg In powernv, send intrs to HV 919 * ->rsvd_txbuf_count Not supported yet. 920 * ->notify_disable False for NX windows 921 * ->xtra_write False for NX windows 922 * ->notify_early NA for NX windows 923 * ->lnotify_lpid NA for Tx windows 924 * ->lnotify_pid NA for Tx windows 925 * ->lnotify_tid NA for Tx windows 926 * ->tx_win_cred_mode Ignore for now for NX windows 927 * ->rx_win_cred_mode Ignore for now for NX windows 928 */ 929 memset(winctx, 0, sizeof(struct vas_winctx)); 930 931 winctx->wcreds_max = txwin->vas_win.wcreds_max; 932 933 winctx->user_win = txattr->user_win; 934 winctx->nx_win = txwin->rxwin->nx_win; 935 winctx->pin_win = txattr->pin_win; 936 winctx->rej_no_credit = txattr->rej_no_credit; 937 winctx->rsvd_txbuf_enable = txattr->rsvd_txbuf_enable; 938 939 winctx->rx_wcred_mode = txattr->rx_wcred_mode; 940 winctx->tx_wcred_mode = txattr->tx_wcred_mode; 941 winctx->rx_word_mode = txattr->rx_win_ord_mode; 942 winctx->tx_word_mode = txattr->tx_win_ord_mode; 943 winctx->rsvd_txbuf_count = txattr->rsvd_txbuf_count; 944 945 winctx->intr_disable = true; 946 if (winctx->nx_win) 947 winctx->data_stamp = true; 948 949 winctx->lpid = txattr->lpid; 950 winctx->pidr = txattr->pidr; 951 winctx->rx_win_id = txwin->rxwin->vas_win.winid; 952 /* 953 * IRQ and fault window setup is successful. Set fault window 954 * for the send window so that ready to handle faults. 955 */ 956 if (txwin->vinst->virq) 957 winctx->fault_win_id = txwin->vinst->fault_win->vas_win.winid; 958 959 winctx->dma_type = VAS_DMA_TYPE_INJECT; 960 winctx->tc_mode = txattr->tc_mode; 961 winctx->min_scope = VAS_SCOPE_LOCAL; 962 winctx->max_scope = VAS_SCOPE_VECTORED_GROUP; 963 if (txwin->vinst->virq) 964 winctx->irq_port = txwin->vinst->irq_port; 965 966 winctx->pswid = txattr->pswid ? txattr->pswid : 967 encode_pswid(txwin->vinst->vas_id, 968 txwin->vas_win.winid); 969 } 970 971 static bool tx_win_args_valid(enum vas_cop_type cop, 972 struct vas_tx_win_attr *attr) 973 { 974 if (attr->tc_mode != VAS_THRESH_DISABLED) 975 return false; 976 977 if (cop > VAS_COP_TYPE_MAX) 978 return false; 979 980 if (attr->wcreds_max > VAS_TX_WCREDS_MAX) 981 return false; 982 983 if (attr->user_win) { 984 if (attr->rsvd_txbuf_count) 985 return false; 986 987 if (cop != VAS_COP_TYPE_FTW && cop != VAS_COP_TYPE_GZIP && 988 cop != VAS_COP_TYPE_GZIP_HIPRI) 989 return false; 990 } 991 992 return true; 993 } 994 995 struct vas_window *vas_tx_win_open(int vasid, enum vas_cop_type cop, 996 struct vas_tx_win_attr *attr) 997 { 998 int rc; 999 struct pnv_vas_window *txwin; 1000 struct pnv_vas_window *rxwin; 1001 struct vas_winctx winctx; 1002 struct vas_instance *vinst; 1003 1004 trace_vas_tx_win_open(current, vasid, cop, attr); 1005 1006 if (!tx_win_args_valid(cop, attr)) 1007 return ERR_PTR(-EINVAL); 1008 1009 /* 1010 * If caller did not specify a vasid but specified the PSWID of a 1011 * receive window (applicable only to FTW windows), use the vasid 1012 * from that receive window. 1013 */ 1014 if (vasid == -1 && attr->pswid) 1015 decode_pswid(attr->pswid, &vasid, NULL); 1016 1017 vinst = find_vas_instance(vasid); 1018 if (!vinst) { 1019 pr_devel("vasid %d not found!\n", vasid); 1020 return ERR_PTR(-EINVAL); 1021 } 1022 1023 rxwin = get_vinst_rxwin(vinst, cop, attr->pswid); 1024 if (IS_ERR(rxwin)) { 1025 pr_devel("No RxWin for vasid %d, cop %d\n", vasid, cop); 1026 return (struct vas_window *)rxwin; 1027 } 1028 1029 txwin = vas_window_alloc(vinst); 1030 if (IS_ERR(txwin)) { 1031 rc = PTR_ERR(txwin); 1032 goto put_rxwin; 1033 } 1034 1035 txwin->vas_win.cop = cop; 1036 txwin->tx_win = 1; 1037 txwin->rxwin = rxwin; 1038 txwin->nx_win = txwin->rxwin->nx_win; 1039 txwin->user_win = attr->user_win; 1040 txwin->vas_win.wcreds_max = attr->wcreds_max ?: VAS_WCREDS_DEFAULT; 1041 1042 init_winctx_for_txwin(txwin, attr, &winctx); 1043 1044 init_winctx_regs(txwin, &winctx); 1045 1046 /* 1047 * If its a kernel send window, map the window address into the 1048 * kernel's address space. For user windows, user must issue an 1049 * mmap() to map the window into their address space. 1050 * 1051 * NOTE: If kernel ever resubmits a user CRB after handling a page 1052 * fault, we will need to map this into kernel as well. 1053 */ 1054 if (!txwin->user_win) { 1055 txwin->paste_kaddr = map_paste_region(txwin); 1056 if (IS_ERR(txwin->paste_kaddr)) { 1057 rc = PTR_ERR(txwin->paste_kaddr); 1058 goto free_window; 1059 } 1060 } else { 1061 /* 1062 * Interrupt handler or fault window setup failed. Means 1063 * NX can not generate fault for page fault. So not 1064 * opening for user space tx window. 1065 */ 1066 if (!vinst->virq) { 1067 rc = -ENODEV; 1068 goto free_window; 1069 } 1070 rc = get_vas_user_win_ref(&txwin->vas_win.task_ref); 1071 if (rc) 1072 goto free_window; 1073 1074 vas_user_win_add_mm_context(&txwin->vas_win.task_ref); 1075 } 1076 1077 set_vinst_win(vinst, txwin); 1078 1079 return &txwin->vas_win; 1080 1081 free_window: 1082 vas_window_free(txwin); 1083 1084 put_rxwin: 1085 put_rx_win(rxwin); 1086 return ERR_PTR(rc); 1087 1088 } 1089 EXPORT_SYMBOL_GPL(vas_tx_win_open); 1090 1091 int vas_copy_crb(void *crb, int offset) 1092 { 1093 return vas_copy(crb, offset); 1094 } 1095 EXPORT_SYMBOL_GPL(vas_copy_crb); 1096 1097 #define RMA_LSMP_REPORT_ENABLE PPC_BIT(53) 1098 int vas_paste_crb(struct vas_window *vwin, int offset, bool re) 1099 { 1100 struct pnv_vas_window *txwin; 1101 int rc; 1102 void *addr; 1103 uint64_t val; 1104 1105 txwin = container_of(vwin, struct pnv_vas_window, vas_win); 1106 trace_vas_paste_crb(current, txwin); 1107 1108 /* 1109 * Only NX windows are supported for now and hardware assumes 1110 * report-enable flag is set for NX windows. Ensure software 1111 * complies too. 1112 */ 1113 WARN_ON_ONCE(txwin->nx_win && !re); 1114 1115 addr = txwin->paste_kaddr; 1116 if (re) { 1117 /* 1118 * Set the REPORT_ENABLE bit (equivalent to writing 1119 * to 1K offset of the paste address) 1120 */ 1121 val = SET_FIELD(RMA_LSMP_REPORT_ENABLE, 0ULL, 1); 1122 addr += val; 1123 } 1124 1125 /* 1126 * Map the raw CR value from vas_paste() to an error code (there 1127 * is just pass or fail for now though). 1128 */ 1129 rc = vas_paste(addr, offset); 1130 if (rc == 2) 1131 rc = 0; 1132 else 1133 rc = -EINVAL; 1134 1135 pr_debug("Txwin #%d: Msg count %llu\n", txwin->vas_win.winid, 1136 read_hvwc_reg(txwin, VREG(LRFIFO_PUSH))); 1137 1138 return rc; 1139 } 1140 EXPORT_SYMBOL_GPL(vas_paste_crb); 1141 1142 /* 1143 * If credit checking is enabled for this window, poll for the return 1144 * of window credits (i.e for NX engines to process any outstanding CRBs). 1145 * Since NX-842 waits for the CRBs to be processed before closing the 1146 * window, we should not have to wait for too long. 1147 * 1148 * TODO: We retry in 10ms intervals now. We could/should probably peek at 1149 * the VAS_LRFIFO_PUSH_OFFSET register to get an estimate of pending 1150 * CRBs on the FIFO and compute the delay dynamically on each retry. 1151 * But that is not really needed until we support NX-GZIP access from 1152 * user space. (NX-842 driver waits for CSB and Fast thread-wakeup 1153 * doesn't use credit checking). 1154 */ 1155 static void poll_window_credits(struct pnv_vas_window *window) 1156 { 1157 u64 val; 1158 int creds, mode; 1159 int count = 0; 1160 1161 val = read_hvwc_reg(window, VREG(WINCTL)); 1162 if (window->tx_win) 1163 mode = GET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val); 1164 else 1165 mode = GET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val); 1166 1167 if (!mode) 1168 return; 1169 retry: 1170 if (window->tx_win) { 1171 val = read_hvwc_reg(window, VREG(TX_WCRED)); 1172 creds = GET_FIELD(VAS_TX_WCRED, val); 1173 } else { 1174 val = read_hvwc_reg(window, VREG(LRX_WCRED)); 1175 creds = GET_FIELD(VAS_LRX_WCRED, val); 1176 } 1177 1178 /* 1179 * Takes around few milliseconds to complete all pending requests 1180 * and return credits. 1181 * TODO: Scan fault FIFO and invalidate CRBs points to this window 1182 * and issue CRB Kill to stop all pending requests. Need only 1183 * if there is a bug in NX or fault handling in kernel. 1184 */ 1185 if (creds < window->vas_win.wcreds_max) { 1186 val = 0; 1187 set_current_state(TASK_UNINTERRUPTIBLE); 1188 schedule_timeout(msecs_to_jiffies(10)); 1189 count++; 1190 /* 1191 * Process can not close send window until all credits are 1192 * returned. 1193 */ 1194 if (!(count % 1000)) 1195 pr_warn_ratelimited("VAS: pid %d stuck. Waiting for credits returned for Window(%d). creds %d, Retries %d\n", 1196 vas_window_pid(&window->vas_win), 1197 window->vas_win.winid, 1198 creds, count); 1199 1200 goto retry; 1201 } 1202 } 1203 1204 /* 1205 * Wait for the window to go to "not-busy" state. It should only take a 1206 * short time to queue a CRB, so window should not be busy for too long. 1207 * Trying 5ms intervals. 1208 */ 1209 static void poll_window_busy_state(struct pnv_vas_window *window) 1210 { 1211 int busy; 1212 u64 val; 1213 int count = 0; 1214 1215 retry: 1216 val = read_hvwc_reg(window, VREG(WIN_STATUS)); 1217 busy = GET_FIELD(VAS_WIN_BUSY, val); 1218 if (busy) { 1219 val = 0; 1220 set_current_state(TASK_UNINTERRUPTIBLE); 1221 schedule_timeout(msecs_to_jiffies(10)); 1222 count++; 1223 /* 1224 * Takes around few milliseconds to process all pending 1225 * requests. 1226 */ 1227 if (!(count % 1000)) 1228 pr_warn_ratelimited("VAS: pid %d stuck. Window (ID=%d) is in busy state. Retries %d\n", 1229 vas_window_pid(&window->vas_win), 1230 window->vas_win.winid, count); 1231 1232 goto retry; 1233 } 1234 } 1235 1236 /* 1237 * Have the hardware cast a window out of cache and wait for it to 1238 * be completed. 1239 * 1240 * NOTE: It can take a relatively long time to cast the window context 1241 * out of the cache. It is not strictly necessary to cast out if: 1242 * 1243 * - we clear the "Pin Window" bit (so hardware is free to evict) 1244 * 1245 * - we re-initialize the window context when it is reassigned. 1246 * 1247 * We do the former in vas_win_close() and latter in vas_win_open(). 1248 * So, ignoring the cast-out for now. We can add it as needed. If 1249 * casting out becomes necessary we should consider offloading the 1250 * job to a worker thread, so the window close can proceed quickly. 1251 */ 1252 static void poll_window_castout(struct pnv_vas_window *window) 1253 { 1254 /* stub for now */ 1255 } 1256 1257 /* 1258 * Unpin and close a window so no new requests are accepted and the 1259 * hardware can evict this window from cache if necessary. 1260 */ 1261 static void unpin_close_window(struct pnv_vas_window *window) 1262 { 1263 u64 val; 1264 1265 val = read_hvwc_reg(window, VREG(WINCTL)); 1266 val = SET_FIELD(VAS_WINCTL_PIN, val, 0); 1267 val = SET_FIELD(VAS_WINCTL_OPEN, val, 0); 1268 write_hvwc_reg(window, VREG(WINCTL), val); 1269 } 1270 1271 /* 1272 * Close a window. 1273 * 1274 * See Section 1.12.1 of VAS workbook v1.05 for details on closing window: 1275 * - Disable new paste operations (unmap paste address) 1276 * - Poll for the "Window Busy" bit to be cleared 1277 * - Clear the Open/Enable bit for the Window. 1278 * - Poll for return of window Credits (implies FIFO empty for Rx win?) 1279 * - Unpin and cast window context out of cache 1280 * 1281 * Besides the hardware, kernel has some bookkeeping of course. 1282 */ 1283 int vas_win_close(struct vas_window *vwin) 1284 { 1285 struct pnv_vas_window *window; 1286 1287 if (!vwin) 1288 return 0; 1289 1290 window = container_of(vwin, struct pnv_vas_window, vas_win); 1291 1292 if (!window->tx_win && atomic_read(&window->num_txwins) != 0) { 1293 pr_devel("Attempting to close an active Rx window!\n"); 1294 WARN_ON_ONCE(1); 1295 return -EBUSY; 1296 } 1297 1298 unmap_paste_region(window); 1299 1300 poll_window_busy_state(window); 1301 1302 unpin_close_window(window); 1303 1304 poll_window_credits(window); 1305 1306 clear_vinst_win(window); 1307 1308 poll_window_castout(window); 1309 1310 /* if send window, drop reference to matching receive window */ 1311 if (window->tx_win) { 1312 if (window->user_win) { 1313 mm_context_remove_vas_window(vwin->task_ref.mm); 1314 put_vas_user_win_ref(&vwin->task_ref); 1315 } 1316 put_rx_win(window->rxwin); 1317 } 1318 1319 vas_window_free(window); 1320 1321 return 0; 1322 } 1323 EXPORT_SYMBOL_GPL(vas_win_close); 1324 1325 /* 1326 * Return credit for the given window. 1327 * Send windows and fault window uses credit mechanism as follows: 1328 * 1329 * Send windows: 1330 * - The default number of credits available for each send window is 1331 * 1024. It means 1024 requests can be issued asynchronously at the 1332 * same time. If the credit is not available, that request will be 1333 * returned with RMA_Busy. 1334 * - One credit is taken when NX request is issued. 1335 * - This credit is returned after NX processed that request. 1336 * - If NX encounters translation error, kernel will return the 1337 * credit on the specific send window after processing the fault CRB. 1338 * 1339 * Fault window: 1340 * - The total number credits available is FIFO_SIZE/CRB_SIZE. 1341 * Means 4MB/128 in the current implementation. If credit is not 1342 * available, RMA_Reject is returned. 1343 * - A credit is taken when NX pastes CRB in fault FIFO. 1344 * - The kernel with return credit on fault window after reading entry 1345 * from fault FIFO. 1346 */ 1347 void vas_return_credit(struct pnv_vas_window *window, bool tx) 1348 { 1349 uint64_t val; 1350 1351 val = 0ULL; 1352 if (tx) { /* send window */ 1353 val = SET_FIELD(VAS_TX_WCRED, val, 1); 1354 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), val); 1355 } else { 1356 val = SET_FIELD(VAS_LRX_WCRED, val, 1); 1357 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), val); 1358 } 1359 } 1360 1361 struct pnv_vas_window *vas_pswid_to_window(struct vas_instance *vinst, 1362 uint32_t pswid) 1363 { 1364 struct pnv_vas_window *window; 1365 int winid; 1366 1367 if (!pswid) { 1368 pr_devel("%s: called for pswid 0!\n", __func__); 1369 return ERR_PTR(-ESRCH); 1370 } 1371 1372 decode_pswid(pswid, NULL, &winid); 1373 1374 if (winid >= VAS_WINDOWS_PER_CHIP) 1375 return ERR_PTR(-ESRCH); 1376 1377 /* 1378 * If application closes the window before the hardware 1379 * returns the fault CRB, we should wait in vas_win_close() 1380 * for the pending requests. so the window must be active 1381 * and the process alive. 1382 * 1383 * If its a kernel process, we should not get any faults and 1384 * should not get here. 1385 */ 1386 window = vinst->windows[winid]; 1387 1388 if (!window) { 1389 pr_err("PSWID decode: Could not find window for winid %d pswid %d vinst 0x%p\n", 1390 winid, pswid, vinst); 1391 return NULL; 1392 } 1393 1394 /* 1395 * Do some sanity checks on the decoded window. Window should be 1396 * NX GZIP user send window. FTW windows should not incur faults 1397 * since their CRBs are ignored (not queued on FIFO or processed 1398 * by NX). 1399 */ 1400 if (!window->tx_win || !window->user_win || !window->nx_win || 1401 window->vas_win.cop == VAS_COP_TYPE_FAULT || 1402 window->vas_win.cop == VAS_COP_TYPE_FTW) { 1403 pr_err("PSWID decode: id %d, tx %d, user %d, nx %d, cop %d\n", 1404 winid, window->tx_win, window->user_win, 1405 window->nx_win, window->vas_win.cop); 1406 WARN_ON(1); 1407 } 1408 1409 return window; 1410 } 1411 1412 static struct vas_window *vas_user_win_open(int vas_id, u64 flags, 1413 enum vas_cop_type cop_type) 1414 { 1415 struct vas_tx_win_attr txattr = {}; 1416 1417 vas_init_tx_win_attr(&txattr, cop_type); 1418 1419 txattr.lpid = mfspr(SPRN_LPID); 1420 txattr.pidr = mfspr(SPRN_PID); 1421 txattr.user_win = true; 1422 txattr.rsvd_txbuf_count = false; 1423 txattr.pswid = false; 1424 1425 pr_devel("Pid %d: Opening txwin, PIDR %ld\n", txattr.pidr, 1426 mfspr(SPRN_PID)); 1427 1428 return vas_tx_win_open(vas_id, cop_type, &txattr); 1429 } 1430 1431 static u64 vas_user_win_paste_addr(struct vas_window *txwin) 1432 { 1433 struct pnv_vas_window *win; 1434 u64 paste_addr; 1435 1436 win = container_of(txwin, struct pnv_vas_window, vas_win); 1437 vas_win_paste_addr(win, &paste_addr, NULL); 1438 1439 return paste_addr; 1440 } 1441 1442 static int vas_user_win_close(struct vas_window *txwin) 1443 { 1444 vas_win_close(txwin); 1445 1446 return 0; 1447 } 1448 1449 static const struct vas_user_win_ops vops = { 1450 .open_win = vas_user_win_open, 1451 .paste_addr = vas_user_win_paste_addr, 1452 .close_win = vas_user_win_close, 1453 }; 1454 1455 /* 1456 * Supporting only nx-gzip coprocessor type now, but this API code 1457 * extended to other coprocessor types later. 1458 */ 1459 int vas_register_api_powernv(struct module *mod, enum vas_cop_type cop_type, 1460 const char *name) 1461 { 1462 1463 return vas_register_coproc_api(mod, cop_type, name, &vops); 1464 } 1465 EXPORT_SYMBOL_GPL(vas_register_api_powernv); 1466 1467 void vas_unregister_api_powernv(void) 1468 { 1469 vas_unregister_coproc_api(); 1470 } 1471 EXPORT_SYMBOL_GPL(vas_unregister_api_powernv); 1472