1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Copyright 2016-17 IBM Corp.
4 */
5
6 #define pr_fmt(fmt) "vas: " fmt
7
8 #include <linux/types.h>
9 #include <linux/mutex.h>
10 #include <linux/slab.h>
11 #include <linux/io.h>
12 #include <linux/log2.h>
13 #include <linux/rcupdate.h>
14 #include <linux/cred.h>
15 #include <linux/sched/mm.h>
16 #include <linux/mmu_context.h>
17 #include <asm/switch_to.h>
18 #include <asm/ppc-opcode.h>
19 #include <asm/vas.h>
20 #include "vas.h"
21 #include "copy-paste.h"
22
23 #define CREATE_TRACE_POINTS
24 #include "vas-trace.h"
25
26 /*
27 * Compute the paste address region for the window @window using the
28 * ->paste_base_addr and ->paste_win_id_shift we got from device tree.
29 */
vas_win_paste_addr(struct pnv_vas_window * window,u64 * addr,int * len)30 void vas_win_paste_addr(struct pnv_vas_window *window, u64 *addr, int *len)
31 {
32 int winid;
33 u64 base, shift;
34
35 base = window->vinst->paste_base_addr;
36 shift = window->vinst->paste_win_id_shift;
37 winid = window->vas_win.winid;
38
39 *addr = base + (winid << shift);
40 if (len)
41 *len = PAGE_SIZE;
42
43 pr_debug("Txwin #%d: Paste addr 0x%llx\n", winid, *addr);
44 }
45
get_hvwc_mmio_bar(struct pnv_vas_window * window,u64 * start,int * len)46 static inline void get_hvwc_mmio_bar(struct pnv_vas_window *window,
47 u64 *start, int *len)
48 {
49 u64 pbaddr;
50
51 pbaddr = window->vinst->hvwc_bar_start;
52 *start = pbaddr + window->vas_win.winid * VAS_HVWC_SIZE;
53 *len = VAS_HVWC_SIZE;
54 }
55
get_uwc_mmio_bar(struct pnv_vas_window * window,u64 * start,int * len)56 static inline void get_uwc_mmio_bar(struct pnv_vas_window *window,
57 u64 *start, int *len)
58 {
59 u64 pbaddr;
60
61 pbaddr = window->vinst->uwc_bar_start;
62 *start = pbaddr + window->vas_win.winid * VAS_UWC_SIZE;
63 *len = VAS_UWC_SIZE;
64 }
65
66 /*
67 * Map the paste bus address of the given send window into kernel address
68 * space. Unlike MMIO regions (map_mmio_region() below), paste region must
69 * be mapped cache-able and is only applicable to send windows.
70 */
map_paste_region(struct pnv_vas_window * txwin)71 static void *map_paste_region(struct pnv_vas_window *txwin)
72 {
73 int len;
74 void *map;
75 char *name;
76 u64 start;
77
78 name = kasprintf(GFP_KERNEL, "window-v%d-w%d", txwin->vinst->vas_id,
79 txwin->vas_win.winid);
80 if (!name)
81 goto free_name;
82
83 txwin->paste_addr_name = name;
84 vas_win_paste_addr(txwin, &start, &len);
85
86 if (!request_mem_region(start, len, name)) {
87 pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n",
88 __func__, start, len);
89 goto free_name;
90 }
91
92 map = ioremap_cache(start, len);
93 if (!map) {
94 pr_devel("%s(): ioremap_cache(0x%llx, %d) failed\n", __func__,
95 start, len);
96 goto free_name;
97 }
98
99 pr_devel("Mapped paste addr 0x%llx to kaddr 0x%p\n", start, map);
100 return map;
101
102 free_name:
103 kfree(name);
104 return ERR_PTR(-ENOMEM);
105 }
106
map_mmio_region(char * name,u64 start,int len)107 static void *map_mmio_region(char *name, u64 start, int len)
108 {
109 void *map;
110
111 if (!request_mem_region(start, len, name)) {
112 pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n",
113 __func__, start, len);
114 return NULL;
115 }
116
117 map = ioremap(start, len);
118 if (!map) {
119 pr_devel("%s(): ioremap(0x%llx, %d) failed\n", __func__, start,
120 len);
121 return NULL;
122 }
123
124 return map;
125 }
126
unmap_region(void * addr,u64 start,int len)127 static void unmap_region(void *addr, u64 start, int len)
128 {
129 iounmap(addr);
130 release_mem_region((phys_addr_t)start, len);
131 }
132
133 /*
134 * Unmap the paste address region for a window.
135 */
unmap_paste_region(struct pnv_vas_window * window)136 static void unmap_paste_region(struct pnv_vas_window *window)
137 {
138 int len;
139 u64 busaddr_start;
140
141 if (window->paste_kaddr) {
142 vas_win_paste_addr(window, &busaddr_start, &len);
143 unmap_region(window->paste_kaddr, busaddr_start, len);
144 window->paste_kaddr = NULL;
145 kfree(window->paste_addr_name);
146 window->paste_addr_name = NULL;
147 }
148 }
149
150 /*
151 * Unmap the MMIO regions for a window. Hold the vas_mutex so we don't
152 * unmap when the window's debugfs dir is in use. This serializes close
153 * of a window even on another VAS instance but since its not a critical
154 * path, just minimize the time we hold the mutex for now. We can add
155 * a per-instance mutex later if necessary.
156 */
unmap_winctx_mmio_bars(struct pnv_vas_window * window)157 static void unmap_winctx_mmio_bars(struct pnv_vas_window *window)
158 {
159 int len;
160 void *uwc_map;
161 void *hvwc_map;
162 u64 busaddr_start;
163
164 mutex_lock(&vas_mutex);
165
166 hvwc_map = window->hvwc_map;
167 window->hvwc_map = NULL;
168
169 uwc_map = window->uwc_map;
170 window->uwc_map = NULL;
171
172 mutex_unlock(&vas_mutex);
173
174 if (hvwc_map) {
175 get_hvwc_mmio_bar(window, &busaddr_start, &len);
176 unmap_region(hvwc_map, busaddr_start, len);
177 }
178
179 if (uwc_map) {
180 get_uwc_mmio_bar(window, &busaddr_start, &len);
181 unmap_region(uwc_map, busaddr_start, len);
182 }
183 }
184
185 /*
186 * Find the Hypervisor Window Context (HVWC) MMIO Base Address Region and the
187 * OS/User Window Context (UWC) MMIO Base Address Region for the given window.
188 * Map these bus addresses and save the mapped kernel addresses in @window.
189 */
map_winctx_mmio_bars(struct pnv_vas_window * window)190 static int map_winctx_mmio_bars(struct pnv_vas_window *window)
191 {
192 int len;
193 u64 start;
194
195 get_hvwc_mmio_bar(window, &start, &len);
196 window->hvwc_map = map_mmio_region("HVWCM_Window", start, len);
197
198 get_uwc_mmio_bar(window, &start, &len);
199 window->uwc_map = map_mmio_region("UWCM_Window", start, len);
200
201 if (!window->hvwc_map || !window->uwc_map) {
202 unmap_winctx_mmio_bars(window);
203 return -1;
204 }
205
206 return 0;
207 }
208
209 /*
210 * Reset all valid registers in the HV and OS/User Window Contexts for
211 * the window identified by @window.
212 *
213 * NOTE: We cannot really use a for loop to reset window context. Not all
214 * offsets in a window context are valid registers and the valid
215 * registers are not sequential. And, we can only write to offsets
216 * with valid registers.
217 */
reset_window_regs(struct pnv_vas_window * window)218 static void reset_window_regs(struct pnv_vas_window *window)
219 {
220 write_hvwc_reg(window, VREG(LPID), 0ULL);
221 write_hvwc_reg(window, VREG(PID), 0ULL);
222 write_hvwc_reg(window, VREG(XLATE_MSR), 0ULL);
223 write_hvwc_reg(window, VREG(XLATE_LPCR), 0ULL);
224 write_hvwc_reg(window, VREG(XLATE_CTL), 0ULL);
225 write_hvwc_reg(window, VREG(AMR), 0ULL);
226 write_hvwc_reg(window, VREG(SEIDR), 0ULL);
227 write_hvwc_reg(window, VREG(FAULT_TX_WIN), 0ULL);
228 write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL);
229 write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), 0ULL);
230 write_hvwc_reg(window, VREG(PSWID), 0ULL);
231 write_hvwc_reg(window, VREG(LFIFO_BAR), 0ULL);
232 write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), 0ULL);
233 write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), 0ULL);
234 write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL);
235 write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL);
236 write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL);
237 write_hvwc_reg(window, VREG(LRX_WCRED), 0ULL);
238 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL);
239 write_hvwc_reg(window, VREG(TX_WCRED), 0ULL);
240 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL);
241 write_hvwc_reg(window, VREG(LFIFO_SIZE), 0ULL);
242 write_hvwc_reg(window, VREG(WINCTL), 0ULL);
243 write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL);
244 write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), 0ULL);
245 write_hvwc_reg(window, VREG(TX_RSVD_BUF_COUNT), 0ULL);
246 write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), 0ULL);
247 write_hvwc_reg(window, VREG(LNOTIFY_CTL), 0ULL);
248 write_hvwc_reg(window, VREG(LNOTIFY_PID), 0ULL);
249 write_hvwc_reg(window, VREG(LNOTIFY_LPID), 0ULL);
250 write_hvwc_reg(window, VREG(LNOTIFY_TID), 0ULL);
251 write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), 0ULL);
252 write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL);
253
254 /* Skip read-only registers: NX_UTIL and NX_UTIL_SE */
255
256 /*
257 * The send and receive window credit adder registers are also
258 * accessible from HVWC and have been initialized above. We don't
259 * need to initialize from the OS/User Window Context, so skip
260 * following calls:
261 *
262 * write_uwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL);
263 * write_uwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL);
264 */
265 }
266
267 /*
268 * Initialize window context registers related to Address Translation.
269 * These registers are common to send/receive windows although they
270 * differ for user/kernel windows. As we resolve the TODOs we may
271 * want to add fields to vas_winctx and move the initialization to
272 * init_vas_winctx_regs().
273 */
init_xlate_regs(struct pnv_vas_window * window,bool user_win)274 static void init_xlate_regs(struct pnv_vas_window *window, bool user_win)
275 {
276 u64 lpcr, val;
277
278 /*
279 * MSR_TA, MSR_US are false for both kernel and user.
280 * MSR_DR and MSR_PR are false for kernel.
281 */
282 val = 0ULL;
283 val = SET_FIELD(VAS_XLATE_MSR_HV, val, 1);
284 val = SET_FIELD(VAS_XLATE_MSR_SF, val, 1);
285 if (user_win) {
286 val = SET_FIELD(VAS_XLATE_MSR_DR, val, 1);
287 val = SET_FIELD(VAS_XLATE_MSR_PR, val, 1);
288 }
289 write_hvwc_reg(window, VREG(XLATE_MSR), val);
290
291 lpcr = mfspr(SPRN_LPCR);
292 val = 0ULL;
293 /*
294 * NOTE: From Section 5.7.8.1 Segment Lookaside Buffer of the
295 * Power ISA, v3.0B, Page size encoding is 0 = 4KB, 5 = 64KB.
296 *
297 * NOTE: From Section 1.3.1, Address Translation Context of the
298 * Nest MMU Workbook, LPCR_SC should be 0 for Power9.
299 */
300 val = SET_FIELD(VAS_XLATE_LPCR_PAGE_SIZE, val, 5);
301 val = SET_FIELD(VAS_XLATE_LPCR_ISL, val, lpcr & LPCR_ISL);
302 val = SET_FIELD(VAS_XLATE_LPCR_TC, val, lpcr & LPCR_TC);
303 val = SET_FIELD(VAS_XLATE_LPCR_SC, val, 0);
304 write_hvwc_reg(window, VREG(XLATE_LPCR), val);
305
306 /*
307 * Section 1.3.1 (Address translation Context) of NMMU workbook.
308 * 0b00 Hashed Page Table mode
309 * 0b01 Reserved
310 * 0b10 Radix on HPT
311 * 0b11 Radix on Radix
312 */
313 val = 0ULL;
314 val = SET_FIELD(VAS_XLATE_MODE, val, radix_enabled() ? 3 : 2);
315 write_hvwc_reg(window, VREG(XLATE_CTL), val);
316
317 /*
318 * TODO: Can we mfspr(AMR) even for user windows?
319 */
320 val = 0ULL;
321 val = SET_FIELD(VAS_AMR, val, mfspr(SPRN_AMR));
322 write_hvwc_reg(window, VREG(AMR), val);
323
324 val = 0ULL;
325 val = SET_FIELD(VAS_SEIDR, val, 0);
326 write_hvwc_reg(window, VREG(SEIDR), val);
327 }
328
329 /*
330 * Initialize Reserved Send Buffer Count for the send window. It involves
331 * writing to the register, reading it back to confirm that the hardware
332 * has enough buffers to reserve. See section 1.3.1.2.1 of VAS workbook.
333 *
334 * Since we can only make a best-effort attempt to fulfill the request,
335 * we don't return any errors if we cannot.
336 *
337 * TODO: Reserved (aka dedicated) send buffers are not supported yet.
338 */
init_rsvd_tx_buf_count(struct pnv_vas_window * txwin,struct vas_winctx * winctx)339 static void init_rsvd_tx_buf_count(struct pnv_vas_window *txwin,
340 struct vas_winctx *winctx)
341 {
342 write_hvwc_reg(txwin, VREG(TX_RSVD_BUF_COUNT), 0ULL);
343 }
344
345 /*
346 * init_winctx_regs()
347 * Initialize window context registers for a receive window.
348 * Except for caching control and marking window open, the registers
349 * are initialized in the order listed in Section 3.1.4 (Window Context
350 * Cache Register Details) of the VAS workbook although they don't need
351 * to be.
352 *
353 * Design note: For NX receive windows, NX allocates the FIFO buffer in OPAL
354 * (so that it can get a large contiguous area) and passes that buffer
355 * to kernel via device tree. We now write that buffer address to the
356 * FIFO BAR. Would it make sense to do this all in OPAL? i.e have OPAL
357 * write the per-chip RX FIFO addresses to the windows during boot-up
358 * as a one-time task? That could work for NX but what about other
359 * receivers? Let the receivers tell us the rx-fifo buffers for now.
360 */
init_winctx_regs(struct pnv_vas_window * window,struct vas_winctx * winctx)361 static void init_winctx_regs(struct pnv_vas_window *window,
362 struct vas_winctx *winctx)
363 {
364 u64 val;
365 int fifo_size;
366
367 reset_window_regs(window);
368
369 val = 0ULL;
370 val = SET_FIELD(VAS_LPID, val, winctx->lpid);
371 write_hvwc_reg(window, VREG(LPID), val);
372
373 val = 0ULL;
374 val = SET_FIELD(VAS_PID_ID, val, winctx->pidr);
375 write_hvwc_reg(window, VREG(PID), val);
376
377 init_xlate_regs(window, winctx->user_win);
378
379 val = 0ULL;
380 val = SET_FIELD(VAS_FAULT_TX_WIN, val, winctx->fault_win_id);
381 write_hvwc_reg(window, VREG(FAULT_TX_WIN), val);
382
383 /* In PowerNV, interrupts go to HV. */
384 write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL);
385
386 val = 0ULL;
387 val = SET_FIELD(VAS_HV_INTR_SRC_RA, val, winctx->irq_port);
388 write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), val);
389
390 val = 0ULL;
391 val = SET_FIELD(VAS_PSWID_EA_HANDLE, val, winctx->pswid);
392 write_hvwc_reg(window, VREG(PSWID), val);
393
394 write_hvwc_reg(window, VREG(SPARE1), 0ULL);
395 write_hvwc_reg(window, VREG(SPARE2), 0ULL);
396 write_hvwc_reg(window, VREG(SPARE3), 0ULL);
397
398 /*
399 * NOTE: VAS expects the FIFO address to be copied into the LFIFO_BAR
400 * register as is - do NOT shift the address into VAS_LFIFO_BAR
401 * bit fields! Ok to set the page migration select fields -
402 * VAS ignores the lower 10+ bits in the address anyway, because
403 * the minimum FIFO size is 1K?
404 *
405 * See also: Design note in function header.
406 */
407 val = winctx->rx_fifo;
408 val = SET_FIELD(VAS_PAGE_MIGRATION_SELECT, val, 0);
409 write_hvwc_reg(window, VREG(LFIFO_BAR), val);
410
411 val = 0ULL;
412 val = SET_FIELD(VAS_LDATA_STAMP, val, winctx->data_stamp);
413 write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), val);
414
415 val = 0ULL;
416 val = SET_FIELD(VAS_LDMA_TYPE, val, winctx->dma_type);
417 val = SET_FIELD(VAS_LDMA_FIFO_DISABLE, val, winctx->fifo_disable);
418 write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), val);
419
420 write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL);
421 write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL);
422 write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL);
423
424 val = 0ULL;
425 val = SET_FIELD(VAS_LRX_WCRED, val, winctx->wcreds_max);
426 write_hvwc_reg(window, VREG(LRX_WCRED), val);
427
428 val = 0ULL;
429 val = SET_FIELD(VAS_TX_WCRED, val, winctx->wcreds_max);
430 write_hvwc_reg(window, VREG(TX_WCRED), val);
431
432 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL);
433 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL);
434
435 fifo_size = winctx->rx_fifo_size / 1024;
436
437 val = 0ULL;
438 val = SET_FIELD(VAS_LFIFO_SIZE, val, ilog2(fifo_size));
439 write_hvwc_reg(window, VREG(LFIFO_SIZE), val);
440
441 /* Update window control and caching control registers last so
442 * we mark the window open only after fully initializing it and
443 * pushing context to cache.
444 */
445
446 write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL);
447
448 init_rsvd_tx_buf_count(window, winctx);
449
450 /* for a send window, point to the matching receive window */
451 val = 0ULL;
452 val = SET_FIELD(VAS_LRX_WIN_ID, val, winctx->rx_win_id);
453 write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), val);
454
455 write_hvwc_reg(window, VREG(SPARE4), 0ULL);
456
457 val = 0ULL;
458 val = SET_FIELD(VAS_NOTIFY_DISABLE, val, winctx->notify_disable);
459 val = SET_FIELD(VAS_INTR_DISABLE, val, winctx->intr_disable);
460 val = SET_FIELD(VAS_NOTIFY_EARLY, val, winctx->notify_early);
461 val = SET_FIELD(VAS_NOTIFY_OSU_INTR, val, winctx->notify_os_intr_reg);
462 write_hvwc_reg(window, VREG(LNOTIFY_CTL), val);
463
464 val = 0ULL;
465 val = SET_FIELD(VAS_LNOTIFY_PID, val, winctx->lnotify_pid);
466 write_hvwc_reg(window, VREG(LNOTIFY_PID), val);
467
468 val = 0ULL;
469 val = SET_FIELD(VAS_LNOTIFY_LPID, val, winctx->lnotify_lpid);
470 write_hvwc_reg(window, VREG(LNOTIFY_LPID), val);
471
472 val = 0ULL;
473 val = SET_FIELD(VAS_LNOTIFY_TID, val, winctx->lnotify_tid);
474 write_hvwc_reg(window, VREG(LNOTIFY_TID), val);
475
476 val = 0ULL;
477 val = SET_FIELD(VAS_LNOTIFY_MIN_SCOPE, val, winctx->min_scope);
478 val = SET_FIELD(VAS_LNOTIFY_MAX_SCOPE, val, winctx->max_scope);
479 write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), val);
480
481 /* Skip read-only registers NX_UTIL and NX_UTIL_SE */
482
483 write_hvwc_reg(window, VREG(SPARE5), 0ULL);
484 write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL);
485 write_hvwc_reg(window, VREG(SPARE6), 0ULL);
486
487 /* Finally, push window context to memory and... */
488 val = 0ULL;
489 val = SET_FIELD(VAS_PUSH_TO_MEM, val, 1);
490 write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), val);
491
492 /* ... mark the window open for business */
493 val = 0ULL;
494 val = SET_FIELD(VAS_WINCTL_REJ_NO_CREDIT, val, winctx->rej_no_credit);
495 val = SET_FIELD(VAS_WINCTL_PIN, val, winctx->pin_win);
496 val = SET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val, winctx->tx_wcred_mode);
497 val = SET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val, winctx->rx_wcred_mode);
498 val = SET_FIELD(VAS_WINCTL_TX_WORD_MODE, val, winctx->tx_word_mode);
499 val = SET_FIELD(VAS_WINCTL_RX_WORD_MODE, val, winctx->rx_word_mode);
500 val = SET_FIELD(VAS_WINCTL_FAULT_WIN, val, winctx->fault_win);
501 val = SET_FIELD(VAS_WINCTL_NX_WIN, val, winctx->nx_win);
502 val = SET_FIELD(VAS_WINCTL_OPEN, val, 1);
503 write_hvwc_reg(window, VREG(WINCTL), val);
504 }
505
vas_release_window_id(struct ida * ida,int winid)506 static void vas_release_window_id(struct ida *ida, int winid)
507 {
508 ida_free(ida, winid);
509 }
510
vas_assign_window_id(struct ida * ida)511 static int vas_assign_window_id(struct ida *ida)
512 {
513 int winid = ida_alloc_max(ida, VAS_WINDOWS_PER_CHIP - 1, GFP_KERNEL);
514
515 if (winid == -ENOSPC) {
516 pr_err("Too many (%d) open windows\n", VAS_WINDOWS_PER_CHIP);
517 return -EAGAIN;
518 }
519
520 return winid;
521 }
522
vas_window_free(struct pnv_vas_window * window)523 static void vas_window_free(struct pnv_vas_window *window)
524 {
525 struct vas_instance *vinst = window->vinst;
526 int winid = window->vas_win.winid;
527
528 unmap_winctx_mmio_bars(window);
529
530 vas_window_free_dbgdir(window);
531
532 kfree(window);
533
534 vas_release_window_id(&vinst->ida, winid);
535 }
536
vas_window_alloc(struct vas_instance * vinst)537 static struct pnv_vas_window *vas_window_alloc(struct vas_instance *vinst)
538 {
539 int winid;
540 struct pnv_vas_window *window;
541
542 winid = vas_assign_window_id(&vinst->ida);
543 if (winid < 0)
544 return ERR_PTR(winid);
545
546 window = kzalloc(sizeof(*window), GFP_KERNEL);
547 if (!window)
548 goto out_free;
549
550 window->vinst = vinst;
551 window->vas_win.winid = winid;
552
553 if (map_winctx_mmio_bars(window))
554 goto out_free;
555
556 vas_window_init_dbgdir(window);
557
558 return window;
559
560 out_free:
561 kfree(window);
562 vas_release_window_id(&vinst->ida, winid);
563 return ERR_PTR(-ENOMEM);
564 }
565
put_rx_win(struct pnv_vas_window * rxwin)566 static void put_rx_win(struct pnv_vas_window *rxwin)
567 {
568 /* Better not be a send window! */
569 WARN_ON_ONCE(rxwin->tx_win);
570
571 atomic_dec(&rxwin->num_txwins);
572 }
573
574 /*
575 * Find the user space receive window given the @pswid.
576 * - We must have a valid vasid and it must belong to this instance.
577 * (so both send and receive windows are on the same VAS instance)
578 * - The window must refer to an OPEN, FTW, RECEIVE window.
579 *
580 * NOTE: We access ->windows[] table and assume that vinst->mutex is held.
581 */
get_user_rxwin(struct vas_instance * vinst,u32 pswid)582 static struct pnv_vas_window *get_user_rxwin(struct vas_instance *vinst,
583 u32 pswid)
584 {
585 int vasid, winid;
586 struct pnv_vas_window *rxwin;
587
588 decode_pswid(pswid, &vasid, &winid);
589
590 if (vinst->vas_id != vasid)
591 return ERR_PTR(-EINVAL);
592
593 rxwin = vinst->windows[winid];
594
595 if (!rxwin || rxwin->tx_win || rxwin->vas_win.cop != VAS_COP_TYPE_FTW)
596 return ERR_PTR(-EINVAL);
597
598 return rxwin;
599 }
600
601 /*
602 * Get the VAS receive window associated with NX engine identified
603 * by @cop and if applicable, @pswid.
604 *
605 * See also function header of set_vinst_win().
606 */
get_vinst_rxwin(struct vas_instance * vinst,enum vas_cop_type cop,u32 pswid)607 static struct pnv_vas_window *get_vinst_rxwin(struct vas_instance *vinst,
608 enum vas_cop_type cop, u32 pswid)
609 {
610 struct pnv_vas_window *rxwin;
611
612 mutex_lock(&vinst->mutex);
613
614 if (cop == VAS_COP_TYPE_FTW)
615 rxwin = get_user_rxwin(vinst, pswid);
616 else
617 rxwin = vinst->rxwin[cop] ?: ERR_PTR(-EINVAL);
618
619 if (!IS_ERR(rxwin))
620 atomic_inc(&rxwin->num_txwins);
621
622 mutex_unlock(&vinst->mutex);
623
624 return rxwin;
625 }
626
627 /*
628 * We have two tables of windows in a VAS instance. The first one,
629 * ->windows[], contains all the windows in the instance and allows
630 * looking up a window by its id. It is used to look up send windows
631 * during fault handling and receive windows when pairing user space
632 * send/receive windows.
633 *
634 * The second table, ->rxwin[], contains receive windows that are
635 * associated with NX engines. This table has VAS_COP_TYPE_MAX
636 * entries and is used to look up a receive window by its
637 * coprocessor type.
638 *
639 * Here, we save @window in the ->windows[] table. If it is a receive
640 * window, we also save the window in the ->rxwin[] table.
641 */
set_vinst_win(struct vas_instance * vinst,struct pnv_vas_window * window)642 static void set_vinst_win(struct vas_instance *vinst,
643 struct pnv_vas_window *window)
644 {
645 int id = window->vas_win.winid;
646
647 mutex_lock(&vinst->mutex);
648
649 /*
650 * There should only be one receive window for a coprocessor type
651 * unless its a user (FTW) window.
652 */
653 if (!window->user_win && !window->tx_win) {
654 WARN_ON_ONCE(vinst->rxwin[window->vas_win.cop]);
655 vinst->rxwin[window->vas_win.cop] = window;
656 }
657
658 WARN_ON_ONCE(vinst->windows[id] != NULL);
659 vinst->windows[id] = window;
660
661 mutex_unlock(&vinst->mutex);
662 }
663
664 /*
665 * Clear this window from the table(s) of windows for this VAS instance.
666 * See also function header of set_vinst_win().
667 */
clear_vinst_win(struct pnv_vas_window * window)668 static void clear_vinst_win(struct pnv_vas_window *window)
669 {
670 int id = window->vas_win.winid;
671 struct vas_instance *vinst = window->vinst;
672
673 mutex_lock(&vinst->mutex);
674
675 if (!window->user_win && !window->tx_win) {
676 WARN_ON_ONCE(!vinst->rxwin[window->vas_win.cop]);
677 vinst->rxwin[window->vas_win.cop] = NULL;
678 }
679
680 WARN_ON_ONCE(vinst->windows[id] != window);
681 vinst->windows[id] = NULL;
682
683 mutex_unlock(&vinst->mutex);
684 }
685
init_winctx_for_rxwin(struct pnv_vas_window * rxwin,struct vas_rx_win_attr * rxattr,struct vas_winctx * winctx)686 static void init_winctx_for_rxwin(struct pnv_vas_window *rxwin,
687 struct vas_rx_win_attr *rxattr,
688 struct vas_winctx *winctx)
689 {
690 /*
691 * We first zero (memset()) all fields and only set non-zero fields.
692 * Following fields are 0/false but maybe deserve a comment:
693 *
694 * ->notify_os_intr_reg In powerNV, send intrs to HV
695 * ->notify_disable False for NX windows
696 * ->intr_disable False for Fault Windows
697 * ->xtra_write False for NX windows
698 * ->notify_early NA for NX windows
699 * ->rsvd_txbuf_count NA for Rx windows
700 * ->lpid, ->pid, ->tid NA for Rx windows
701 */
702
703 memset(winctx, 0, sizeof(struct vas_winctx));
704
705 winctx->rx_fifo = rxattr->rx_fifo;
706 winctx->rx_fifo_size = rxattr->rx_fifo_size;
707 winctx->wcreds_max = rxwin->vas_win.wcreds_max;
708 winctx->pin_win = rxattr->pin_win;
709
710 winctx->nx_win = rxattr->nx_win;
711 winctx->fault_win = rxattr->fault_win;
712 winctx->user_win = rxattr->user_win;
713 winctx->rej_no_credit = rxattr->rej_no_credit;
714 winctx->rx_word_mode = rxattr->rx_win_ord_mode;
715 winctx->tx_word_mode = rxattr->tx_win_ord_mode;
716 winctx->rx_wcred_mode = rxattr->rx_wcred_mode;
717 winctx->tx_wcred_mode = rxattr->tx_wcred_mode;
718 winctx->notify_early = rxattr->notify_early;
719
720 if (winctx->nx_win) {
721 winctx->data_stamp = true;
722 winctx->intr_disable = true;
723 winctx->pin_win = true;
724
725 WARN_ON_ONCE(winctx->fault_win);
726 WARN_ON_ONCE(!winctx->rx_word_mode);
727 WARN_ON_ONCE(!winctx->tx_word_mode);
728 WARN_ON_ONCE(winctx->notify_after_count);
729 } else if (winctx->fault_win) {
730 winctx->notify_disable = true;
731 } else if (winctx->user_win) {
732 /*
733 * Section 1.8.1 Low Latency Core-Core Wake up of
734 * the VAS workbook:
735 *
736 * - disable credit checks ([tr]x_wcred_mode = false)
737 * - disable FIFO writes
738 * - enable ASB_Notify, disable interrupt
739 */
740 winctx->fifo_disable = true;
741 winctx->intr_disable = true;
742 winctx->rx_fifo = 0;
743 }
744
745 winctx->lnotify_lpid = rxattr->lnotify_lpid;
746 winctx->lnotify_pid = rxattr->lnotify_pid;
747 winctx->lnotify_tid = rxattr->lnotify_tid;
748 winctx->pswid = rxattr->pswid;
749 winctx->dma_type = VAS_DMA_TYPE_INJECT;
750 winctx->tc_mode = rxattr->tc_mode;
751
752 winctx->min_scope = VAS_SCOPE_LOCAL;
753 winctx->max_scope = VAS_SCOPE_VECTORED_GROUP;
754 if (rxwin->vinst->virq)
755 winctx->irq_port = rxwin->vinst->irq_port;
756 }
757
rx_win_args_valid(enum vas_cop_type cop,struct vas_rx_win_attr * attr)758 static bool rx_win_args_valid(enum vas_cop_type cop,
759 struct vas_rx_win_attr *attr)
760 {
761 pr_debug("Rxattr: fault %d, notify %d, intr %d, early %d, fifo %d\n",
762 attr->fault_win, attr->notify_disable,
763 attr->intr_disable, attr->notify_early,
764 attr->rx_fifo_size);
765
766 if (cop >= VAS_COP_TYPE_MAX)
767 return false;
768
769 if (cop != VAS_COP_TYPE_FTW &&
770 attr->rx_fifo_size < VAS_RX_FIFO_SIZE_MIN)
771 return false;
772
773 if (attr->rx_fifo_size > VAS_RX_FIFO_SIZE_MAX)
774 return false;
775
776 if (!attr->wcreds_max)
777 return false;
778
779 if (attr->nx_win) {
780 /* cannot be fault or user window if it is nx */
781 if (attr->fault_win || attr->user_win)
782 return false;
783 /*
784 * Section 3.1.4.32: NX Windows must not disable notification,
785 * and must not enable interrupts or early notification.
786 */
787 if (attr->notify_disable || !attr->intr_disable ||
788 attr->notify_early)
789 return false;
790 } else if (attr->fault_win) {
791 /* cannot be both fault and user window */
792 if (attr->user_win)
793 return false;
794
795 /*
796 * Section 3.1.4.32: Fault windows must disable notification
797 * but not interrupts.
798 */
799 if (!attr->notify_disable || attr->intr_disable)
800 return false;
801
802 } else if (attr->user_win) {
803 /*
804 * User receive windows are only for fast-thread-wakeup
805 * (FTW). They don't need a FIFO and must disable interrupts
806 */
807 if (attr->rx_fifo || attr->rx_fifo_size || !attr->intr_disable)
808 return false;
809 } else {
810 /* Rx window must be one of NX or Fault or User window. */
811 return false;
812 }
813
814 return true;
815 }
816
vas_init_rx_win_attr(struct vas_rx_win_attr * rxattr,enum vas_cop_type cop)817 void vas_init_rx_win_attr(struct vas_rx_win_attr *rxattr, enum vas_cop_type cop)
818 {
819 memset(rxattr, 0, sizeof(*rxattr));
820
821 if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI ||
822 cop == VAS_COP_TYPE_GZIP || cop == VAS_COP_TYPE_GZIP_HIPRI) {
823 rxattr->pin_win = true;
824 rxattr->nx_win = true;
825 rxattr->fault_win = false;
826 rxattr->intr_disable = true;
827 rxattr->rx_wcred_mode = true;
828 rxattr->tx_wcred_mode = true;
829 rxattr->rx_win_ord_mode = true;
830 rxattr->tx_win_ord_mode = true;
831 } else if (cop == VAS_COP_TYPE_FAULT) {
832 rxattr->pin_win = true;
833 rxattr->fault_win = true;
834 rxattr->notify_disable = true;
835 rxattr->rx_wcred_mode = true;
836 rxattr->rx_win_ord_mode = true;
837 rxattr->rej_no_credit = true;
838 rxattr->tc_mode = VAS_THRESH_DISABLED;
839 } else if (cop == VAS_COP_TYPE_FTW) {
840 rxattr->user_win = true;
841 rxattr->intr_disable = true;
842
843 /*
844 * As noted in the VAS Workbook we disable credit checks.
845 * If we enable credit checks in the future, we must also
846 * implement a mechanism to return the user credits or new
847 * paste operations will fail.
848 */
849 }
850 }
851 EXPORT_SYMBOL_GPL(vas_init_rx_win_attr);
852
vas_rx_win_open(int vasid,enum vas_cop_type cop,struct vas_rx_win_attr * rxattr)853 struct vas_window *vas_rx_win_open(int vasid, enum vas_cop_type cop,
854 struct vas_rx_win_attr *rxattr)
855 {
856 struct pnv_vas_window *rxwin;
857 struct vas_winctx winctx;
858 struct vas_instance *vinst;
859
860 trace_vas_rx_win_open(current, vasid, cop, rxattr);
861
862 if (!rx_win_args_valid(cop, rxattr))
863 return ERR_PTR(-EINVAL);
864
865 vinst = find_vas_instance(vasid);
866 if (!vinst) {
867 pr_devel("vasid %d not found!\n", vasid);
868 return ERR_PTR(-EINVAL);
869 }
870 pr_devel("Found instance %d\n", vasid);
871
872 rxwin = vas_window_alloc(vinst);
873 if (IS_ERR(rxwin)) {
874 pr_devel("Unable to allocate memory for Rx window\n");
875 return (struct vas_window *)rxwin;
876 }
877
878 rxwin->tx_win = false;
879 rxwin->nx_win = rxattr->nx_win;
880 rxwin->user_win = rxattr->user_win;
881 rxwin->vas_win.cop = cop;
882 rxwin->vas_win.wcreds_max = rxattr->wcreds_max;
883
884 init_winctx_for_rxwin(rxwin, rxattr, &winctx);
885 init_winctx_regs(rxwin, &winctx);
886
887 set_vinst_win(vinst, rxwin);
888
889 return &rxwin->vas_win;
890 }
891 EXPORT_SYMBOL_GPL(vas_rx_win_open);
892
vas_init_tx_win_attr(struct vas_tx_win_attr * txattr,enum vas_cop_type cop)893 void vas_init_tx_win_attr(struct vas_tx_win_attr *txattr, enum vas_cop_type cop)
894 {
895 memset(txattr, 0, sizeof(*txattr));
896
897 if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI ||
898 cop == VAS_COP_TYPE_GZIP || cop == VAS_COP_TYPE_GZIP_HIPRI) {
899 txattr->rej_no_credit = false;
900 txattr->rx_wcred_mode = true;
901 txattr->tx_wcred_mode = true;
902 txattr->rx_win_ord_mode = true;
903 txattr->tx_win_ord_mode = true;
904 } else if (cop == VAS_COP_TYPE_FTW) {
905 txattr->user_win = true;
906 }
907 }
908 EXPORT_SYMBOL_GPL(vas_init_tx_win_attr);
909
init_winctx_for_txwin(struct pnv_vas_window * txwin,struct vas_tx_win_attr * txattr,struct vas_winctx * winctx)910 static void init_winctx_for_txwin(struct pnv_vas_window *txwin,
911 struct vas_tx_win_attr *txattr,
912 struct vas_winctx *winctx)
913 {
914 /*
915 * We first zero all fields and only set non-zero ones. Following
916 * are some fields set to 0/false for the stated reason:
917 *
918 * ->notify_os_intr_reg In powernv, send intrs to HV
919 * ->rsvd_txbuf_count Not supported yet.
920 * ->notify_disable False for NX windows
921 * ->xtra_write False for NX windows
922 * ->notify_early NA for NX windows
923 * ->lnotify_lpid NA for Tx windows
924 * ->lnotify_pid NA for Tx windows
925 * ->lnotify_tid NA for Tx windows
926 * ->tx_win_cred_mode Ignore for now for NX windows
927 * ->rx_win_cred_mode Ignore for now for NX windows
928 */
929 memset(winctx, 0, sizeof(struct vas_winctx));
930
931 winctx->wcreds_max = txwin->vas_win.wcreds_max;
932
933 winctx->user_win = txattr->user_win;
934 winctx->nx_win = txwin->rxwin->nx_win;
935 winctx->pin_win = txattr->pin_win;
936 winctx->rej_no_credit = txattr->rej_no_credit;
937 winctx->rsvd_txbuf_enable = txattr->rsvd_txbuf_enable;
938
939 winctx->rx_wcred_mode = txattr->rx_wcred_mode;
940 winctx->tx_wcred_mode = txattr->tx_wcred_mode;
941 winctx->rx_word_mode = txattr->rx_win_ord_mode;
942 winctx->tx_word_mode = txattr->tx_win_ord_mode;
943 winctx->rsvd_txbuf_count = txattr->rsvd_txbuf_count;
944
945 winctx->intr_disable = true;
946 if (winctx->nx_win)
947 winctx->data_stamp = true;
948
949 winctx->lpid = txattr->lpid;
950 winctx->pidr = txattr->pidr;
951 winctx->rx_win_id = txwin->rxwin->vas_win.winid;
952 /*
953 * IRQ and fault window setup is successful. Set fault window
954 * for the send window so that ready to handle faults.
955 */
956 if (txwin->vinst->virq)
957 winctx->fault_win_id = txwin->vinst->fault_win->vas_win.winid;
958
959 winctx->dma_type = VAS_DMA_TYPE_INJECT;
960 winctx->tc_mode = txattr->tc_mode;
961 winctx->min_scope = VAS_SCOPE_LOCAL;
962 winctx->max_scope = VAS_SCOPE_VECTORED_GROUP;
963 if (txwin->vinst->virq)
964 winctx->irq_port = txwin->vinst->irq_port;
965
966 winctx->pswid = txattr->pswid ? txattr->pswid :
967 encode_pswid(txwin->vinst->vas_id,
968 txwin->vas_win.winid);
969 }
970
tx_win_args_valid(enum vas_cop_type cop,struct vas_tx_win_attr * attr)971 static bool tx_win_args_valid(enum vas_cop_type cop,
972 struct vas_tx_win_attr *attr)
973 {
974 if (attr->tc_mode != VAS_THRESH_DISABLED)
975 return false;
976
977 if (cop > VAS_COP_TYPE_MAX)
978 return false;
979
980 if (attr->wcreds_max > VAS_TX_WCREDS_MAX)
981 return false;
982
983 if (attr->user_win) {
984 if (attr->rsvd_txbuf_count)
985 return false;
986
987 if (cop != VAS_COP_TYPE_FTW && cop != VAS_COP_TYPE_GZIP &&
988 cop != VAS_COP_TYPE_GZIP_HIPRI)
989 return false;
990 }
991
992 return true;
993 }
994
vas_tx_win_open(int vasid,enum vas_cop_type cop,struct vas_tx_win_attr * attr)995 struct vas_window *vas_tx_win_open(int vasid, enum vas_cop_type cop,
996 struct vas_tx_win_attr *attr)
997 {
998 int rc;
999 struct pnv_vas_window *txwin;
1000 struct pnv_vas_window *rxwin;
1001 struct vas_winctx winctx;
1002 struct vas_instance *vinst;
1003
1004 trace_vas_tx_win_open(current, vasid, cop, attr);
1005
1006 if (!tx_win_args_valid(cop, attr))
1007 return ERR_PTR(-EINVAL);
1008
1009 /*
1010 * If caller did not specify a vasid but specified the PSWID of a
1011 * receive window (applicable only to FTW windows), use the vasid
1012 * from that receive window.
1013 */
1014 if (vasid == -1 && attr->pswid)
1015 decode_pswid(attr->pswid, &vasid, NULL);
1016
1017 vinst = find_vas_instance(vasid);
1018 if (!vinst) {
1019 pr_devel("vasid %d not found!\n", vasid);
1020 return ERR_PTR(-EINVAL);
1021 }
1022
1023 rxwin = get_vinst_rxwin(vinst, cop, attr->pswid);
1024 if (IS_ERR(rxwin)) {
1025 pr_devel("No RxWin for vasid %d, cop %d\n", vasid, cop);
1026 return (struct vas_window *)rxwin;
1027 }
1028
1029 txwin = vas_window_alloc(vinst);
1030 if (IS_ERR(txwin)) {
1031 rc = PTR_ERR(txwin);
1032 goto put_rxwin;
1033 }
1034
1035 txwin->vas_win.cop = cop;
1036 txwin->tx_win = 1;
1037 txwin->rxwin = rxwin;
1038 txwin->nx_win = txwin->rxwin->nx_win;
1039 txwin->user_win = attr->user_win;
1040 txwin->vas_win.wcreds_max = attr->wcreds_max ?: VAS_WCREDS_DEFAULT;
1041
1042 init_winctx_for_txwin(txwin, attr, &winctx);
1043
1044 init_winctx_regs(txwin, &winctx);
1045
1046 /*
1047 * If its a kernel send window, map the window address into the
1048 * kernel's address space. For user windows, user must issue an
1049 * mmap() to map the window into their address space.
1050 *
1051 * NOTE: If kernel ever resubmits a user CRB after handling a page
1052 * fault, we will need to map this into kernel as well.
1053 */
1054 if (!txwin->user_win) {
1055 txwin->paste_kaddr = map_paste_region(txwin);
1056 if (IS_ERR(txwin->paste_kaddr)) {
1057 rc = PTR_ERR(txwin->paste_kaddr);
1058 goto free_window;
1059 }
1060 } else {
1061 /*
1062 * Interrupt handler or fault window setup failed. Means
1063 * NX can not generate fault for page fault. So not
1064 * opening for user space tx window.
1065 */
1066 if (!vinst->virq) {
1067 rc = -ENODEV;
1068 goto free_window;
1069 }
1070 rc = get_vas_user_win_ref(&txwin->vas_win.task_ref);
1071 if (rc)
1072 goto free_window;
1073
1074 vas_user_win_add_mm_context(&txwin->vas_win.task_ref);
1075 }
1076
1077 set_vinst_win(vinst, txwin);
1078
1079 return &txwin->vas_win;
1080
1081 free_window:
1082 vas_window_free(txwin);
1083
1084 put_rxwin:
1085 put_rx_win(rxwin);
1086 return ERR_PTR(rc);
1087
1088 }
1089 EXPORT_SYMBOL_GPL(vas_tx_win_open);
1090
vas_copy_crb(void * crb,int offset)1091 int vas_copy_crb(void *crb, int offset)
1092 {
1093 return vas_copy(crb, offset);
1094 }
1095 EXPORT_SYMBOL_GPL(vas_copy_crb);
1096
1097 #define RMA_LSMP_REPORT_ENABLE PPC_BIT(53)
vas_paste_crb(struct vas_window * vwin,int offset,bool re)1098 int vas_paste_crb(struct vas_window *vwin, int offset, bool re)
1099 {
1100 struct pnv_vas_window *txwin;
1101 int rc;
1102 void *addr;
1103 uint64_t val;
1104
1105 txwin = container_of(vwin, struct pnv_vas_window, vas_win);
1106 trace_vas_paste_crb(current, txwin);
1107
1108 /*
1109 * Only NX windows are supported for now and hardware assumes
1110 * report-enable flag is set for NX windows. Ensure software
1111 * complies too.
1112 */
1113 WARN_ON_ONCE(txwin->nx_win && !re);
1114
1115 addr = txwin->paste_kaddr;
1116 if (re) {
1117 /*
1118 * Set the REPORT_ENABLE bit (equivalent to writing
1119 * to 1K offset of the paste address)
1120 */
1121 val = SET_FIELD(RMA_LSMP_REPORT_ENABLE, 0ULL, 1);
1122 addr += val;
1123 }
1124
1125 /*
1126 * Map the raw CR value from vas_paste() to an error code (there
1127 * is just pass or fail for now though).
1128 */
1129 rc = vas_paste(addr, offset);
1130 if (rc == 2)
1131 rc = 0;
1132 else
1133 rc = -EINVAL;
1134
1135 pr_debug("Txwin #%d: Msg count %llu\n", txwin->vas_win.winid,
1136 read_hvwc_reg(txwin, VREG(LRFIFO_PUSH)));
1137
1138 return rc;
1139 }
1140 EXPORT_SYMBOL_GPL(vas_paste_crb);
1141
1142 /*
1143 * If credit checking is enabled for this window, poll for the return
1144 * of window credits (i.e for NX engines to process any outstanding CRBs).
1145 * Since NX-842 waits for the CRBs to be processed before closing the
1146 * window, we should not have to wait for too long.
1147 *
1148 * TODO: We retry in 10ms intervals now. We could/should probably peek at
1149 * the VAS_LRFIFO_PUSH_OFFSET register to get an estimate of pending
1150 * CRBs on the FIFO and compute the delay dynamically on each retry.
1151 * But that is not really needed until we support NX-GZIP access from
1152 * user space. (NX-842 driver waits for CSB and Fast thread-wakeup
1153 * doesn't use credit checking).
1154 */
poll_window_credits(struct pnv_vas_window * window)1155 static void poll_window_credits(struct pnv_vas_window *window)
1156 {
1157 u64 val;
1158 int creds, mode;
1159 int count = 0;
1160
1161 val = read_hvwc_reg(window, VREG(WINCTL));
1162 if (window->tx_win)
1163 mode = GET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val);
1164 else
1165 mode = GET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val);
1166
1167 if (!mode)
1168 return;
1169 retry:
1170 if (window->tx_win) {
1171 val = read_hvwc_reg(window, VREG(TX_WCRED));
1172 creds = GET_FIELD(VAS_TX_WCRED, val);
1173 } else {
1174 val = read_hvwc_reg(window, VREG(LRX_WCRED));
1175 creds = GET_FIELD(VAS_LRX_WCRED, val);
1176 }
1177
1178 /*
1179 * Takes around few milliseconds to complete all pending requests
1180 * and return credits.
1181 * TODO: Scan fault FIFO and invalidate CRBs points to this window
1182 * and issue CRB Kill to stop all pending requests. Need only
1183 * if there is a bug in NX or fault handling in kernel.
1184 */
1185 if (creds < window->vas_win.wcreds_max) {
1186 val = 0;
1187 set_current_state(TASK_UNINTERRUPTIBLE);
1188 schedule_timeout(msecs_to_jiffies(10));
1189 count++;
1190 /*
1191 * Process can not close send window until all credits are
1192 * returned.
1193 */
1194 if (!(count % 1000))
1195 pr_warn_ratelimited("VAS: pid %d stuck. Waiting for credits returned for Window(%d). creds %d, Retries %d\n",
1196 vas_window_pid(&window->vas_win),
1197 window->vas_win.winid,
1198 creds, count);
1199
1200 goto retry;
1201 }
1202 }
1203
1204 /*
1205 * Wait for the window to go to "not-busy" state. It should only take a
1206 * short time to queue a CRB, so window should not be busy for too long.
1207 * Trying 5ms intervals.
1208 */
poll_window_busy_state(struct pnv_vas_window * window)1209 static void poll_window_busy_state(struct pnv_vas_window *window)
1210 {
1211 int busy;
1212 u64 val;
1213 int count = 0;
1214
1215 retry:
1216 val = read_hvwc_reg(window, VREG(WIN_STATUS));
1217 busy = GET_FIELD(VAS_WIN_BUSY, val);
1218 if (busy) {
1219 val = 0;
1220 set_current_state(TASK_UNINTERRUPTIBLE);
1221 schedule_timeout(msecs_to_jiffies(10));
1222 count++;
1223 /*
1224 * Takes around few milliseconds to process all pending
1225 * requests.
1226 */
1227 if (!(count % 1000))
1228 pr_warn_ratelimited("VAS: pid %d stuck. Window (ID=%d) is in busy state. Retries %d\n",
1229 vas_window_pid(&window->vas_win),
1230 window->vas_win.winid, count);
1231
1232 goto retry;
1233 }
1234 }
1235
1236 /*
1237 * Have the hardware cast a window out of cache and wait for it to
1238 * be completed.
1239 *
1240 * NOTE: It can take a relatively long time to cast the window context
1241 * out of the cache. It is not strictly necessary to cast out if:
1242 *
1243 * - we clear the "Pin Window" bit (so hardware is free to evict)
1244 *
1245 * - we re-initialize the window context when it is reassigned.
1246 *
1247 * We do the former in vas_win_close() and latter in vas_win_open().
1248 * So, ignoring the cast-out for now. We can add it as needed. If
1249 * casting out becomes necessary we should consider offloading the
1250 * job to a worker thread, so the window close can proceed quickly.
1251 */
poll_window_castout(struct pnv_vas_window * window)1252 static void poll_window_castout(struct pnv_vas_window *window)
1253 {
1254 /* stub for now */
1255 }
1256
1257 /*
1258 * Unpin and close a window so no new requests are accepted and the
1259 * hardware can evict this window from cache if necessary.
1260 */
unpin_close_window(struct pnv_vas_window * window)1261 static void unpin_close_window(struct pnv_vas_window *window)
1262 {
1263 u64 val;
1264
1265 val = read_hvwc_reg(window, VREG(WINCTL));
1266 val = SET_FIELD(VAS_WINCTL_PIN, val, 0);
1267 val = SET_FIELD(VAS_WINCTL_OPEN, val, 0);
1268 write_hvwc_reg(window, VREG(WINCTL), val);
1269 }
1270
1271 /*
1272 * Close a window.
1273 *
1274 * See Section 1.12.1 of VAS workbook v1.05 for details on closing window:
1275 * - Disable new paste operations (unmap paste address)
1276 * - Poll for the "Window Busy" bit to be cleared
1277 * - Clear the Open/Enable bit for the Window.
1278 * - Poll for return of window Credits (implies FIFO empty for Rx win?)
1279 * - Unpin and cast window context out of cache
1280 *
1281 * Besides the hardware, kernel has some bookkeeping of course.
1282 */
vas_win_close(struct vas_window * vwin)1283 int vas_win_close(struct vas_window *vwin)
1284 {
1285 struct pnv_vas_window *window;
1286
1287 if (!vwin)
1288 return 0;
1289
1290 window = container_of(vwin, struct pnv_vas_window, vas_win);
1291
1292 if (!window->tx_win && atomic_read(&window->num_txwins) != 0) {
1293 pr_devel("Attempting to close an active Rx window!\n");
1294 WARN_ON_ONCE(1);
1295 return -EBUSY;
1296 }
1297
1298 unmap_paste_region(window);
1299
1300 poll_window_busy_state(window);
1301
1302 unpin_close_window(window);
1303
1304 poll_window_credits(window);
1305
1306 clear_vinst_win(window);
1307
1308 poll_window_castout(window);
1309
1310 /* if send window, drop reference to matching receive window */
1311 if (window->tx_win) {
1312 if (window->user_win) {
1313 mm_context_remove_vas_window(vwin->task_ref.mm);
1314 put_vas_user_win_ref(&vwin->task_ref);
1315 }
1316 put_rx_win(window->rxwin);
1317 }
1318
1319 vas_window_free(window);
1320
1321 return 0;
1322 }
1323 EXPORT_SYMBOL_GPL(vas_win_close);
1324
1325 /*
1326 * Return credit for the given window.
1327 * Send windows and fault window uses credit mechanism as follows:
1328 *
1329 * Send windows:
1330 * - The default number of credits available for each send window is
1331 * 1024. It means 1024 requests can be issued asynchronously at the
1332 * same time. If the credit is not available, that request will be
1333 * returned with RMA_Busy.
1334 * - One credit is taken when NX request is issued.
1335 * - This credit is returned after NX processed that request.
1336 * - If NX encounters translation error, kernel will return the
1337 * credit on the specific send window after processing the fault CRB.
1338 *
1339 * Fault window:
1340 * - The total number credits available is FIFO_SIZE/CRB_SIZE.
1341 * Means 4MB/128 in the current implementation. If credit is not
1342 * available, RMA_Reject is returned.
1343 * - A credit is taken when NX pastes CRB in fault FIFO.
1344 * - The kernel with return credit on fault window after reading entry
1345 * from fault FIFO.
1346 */
vas_return_credit(struct pnv_vas_window * window,bool tx)1347 void vas_return_credit(struct pnv_vas_window *window, bool tx)
1348 {
1349 uint64_t val;
1350
1351 val = 0ULL;
1352 if (tx) { /* send window */
1353 val = SET_FIELD(VAS_TX_WCRED, val, 1);
1354 write_hvwc_reg(window, VREG(TX_WCRED_ADDER), val);
1355 } else {
1356 val = SET_FIELD(VAS_LRX_WCRED, val, 1);
1357 write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), val);
1358 }
1359 }
1360
vas_pswid_to_window(struct vas_instance * vinst,uint32_t pswid)1361 struct pnv_vas_window *vas_pswid_to_window(struct vas_instance *vinst,
1362 uint32_t pswid)
1363 {
1364 struct pnv_vas_window *window;
1365 int winid;
1366
1367 if (!pswid) {
1368 pr_devel("%s: called for pswid 0!\n", __func__);
1369 return ERR_PTR(-ESRCH);
1370 }
1371
1372 decode_pswid(pswid, NULL, &winid);
1373
1374 if (winid >= VAS_WINDOWS_PER_CHIP)
1375 return ERR_PTR(-ESRCH);
1376
1377 /*
1378 * If application closes the window before the hardware
1379 * returns the fault CRB, we should wait in vas_win_close()
1380 * for the pending requests. so the window must be active
1381 * and the process alive.
1382 *
1383 * If its a kernel process, we should not get any faults and
1384 * should not get here.
1385 */
1386 window = vinst->windows[winid];
1387
1388 if (!window) {
1389 pr_err("PSWID decode: Could not find window for winid %d pswid %d vinst 0x%p\n",
1390 winid, pswid, vinst);
1391 return NULL;
1392 }
1393
1394 /*
1395 * Do some sanity checks on the decoded window. Window should be
1396 * NX GZIP user send window. FTW windows should not incur faults
1397 * since their CRBs are ignored (not queued on FIFO or processed
1398 * by NX).
1399 */
1400 if (!window->tx_win || !window->user_win || !window->nx_win ||
1401 window->vas_win.cop == VAS_COP_TYPE_FAULT ||
1402 window->vas_win.cop == VAS_COP_TYPE_FTW) {
1403 pr_err("PSWID decode: id %d, tx %d, user %d, nx %d, cop %d\n",
1404 winid, window->tx_win, window->user_win,
1405 window->nx_win, window->vas_win.cop);
1406 WARN_ON(1);
1407 }
1408
1409 return window;
1410 }
1411
vas_user_win_open(int vas_id,u64 flags,enum vas_cop_type cop_type)1412 static struct vas_window *vas_user_win_open(int vas_id, u64 flags,
1413 enum vas_cop_type cop_type)
1414 {
1415 struct vas_tx_win_attr txattr = {};
1416
1417 vas_init_tx_win_attr(&txattr, cop_type);
1418
1419 txattr.lpid = mfspr(SPRN_LPID);
1420 txattr.pidr = mfspr(SPRN_PID);
1421 txattr.user_win = true;
1422 txattr.rsvd_txbuf_count = false;
1423 txattr.pswid = false;
1424
1425 pr_devel("Pid %d: Opening txwin, PIDR %ld\n", txattr.pidr,
1426 mfspr(SPRN_PID));
1427
1428 return vas_tx_win_open(vas_id, cop_type, &txattr);
1429 }
1430
vas_user_win_paste_addr(struct vas_window * txwin)1431 static u64 vas_user_win_paste_addr(struct vas_window *txwin)
1432 {
1433 struct pnv_vas_window *win;
1434 u64 paste_addr;
1435
1436 win = container_of(txwin, struct pnv_vas_window, vas_win);
1437 vas_win_paste_addr(win, &paste_addr, NULL);
1438
1439 return paste_addr;
1440 }
1441
vas_user_win_close(struct vas_window * txwin)1442 static int vas_user_win_close(struct vas_window *txwin)
1443 {
1444 vas_win_close(txwin);
1445
1446 return 0;
1447 }
1448
1449 static const struct vas_user_win_ops vops = {
1450 .open_win = vas_user_win_open,
1451 .paste_addr = vas_user_win_paste_addr,
1452 .close_win = vas_user_win_close,
1453 };
1454
1455 /*
1456 * Supporting only nx-gzip coprocessor type now, but this API code
1457 * extended to other coprocessor types later.
1458 */
vas_register_api_powernv(struct module * mod,enum vas_cop_type cop_type,const char * name)1459 int vas_register_api_powernv(struct module *mod, enum vas_cop_type cop_type,
1460 const char *name)
1461 {
1462
1463 return vas_register_coproc_api(mod, cop_type, name, &vops);
1464 }
1465 EXPORT_SYMBOL_GPL(vas_register_api_powernv);
1466
vas_unregister_api_powernv(void)1467 void vas_unregister_api_powernv(void)
1468 {
1469 vas_unregister_coproc_api();
1470 }
1471 EXPORT_SYMBOL_GPL(vas_unregister_api_powernv);
1472