17c478bd9Sstevel@tonic-gate /* 27c478bd9Sstevel@tonic-gate * CDDL HEADER START 37c478bd9Sstevel@tonic-gate * 47c478bd9Sstevel@tonic-gate * The contents of this file are subject to the terms of the 5*17169044Sbrutus * Common Development and Distribution License (the "License"). 6*17169044Sbrutus * You may not use this file except in compliance with the License. 77c478bd9Sstevel@tonic-gate * 87c478bd9Sstevel@tonic-gate * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 97c478bd9Sstevel@tonic-gate * or http://www.opensolaris.org/os/licensing. 107c478bd9Sstevel@tonic-gate * See the License for the specific language governing permissions 117c478bd9Sstevel@tonic-gate * and limitations under the License. 127c478bd9Sstevel@tonic-gate * 137c478bd9Sstevel@tonic-gate * When distributing Covered Code, include this CDDL HEADER in each 147c478bd9Sstevel@tonic-gate * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 157c478bd9Sstevel@tonic-gate * If applicable, add the following below this CDDL HEADER, with the 167c478bd9Sstevel@tonic-gate * fields enclosed by brackets "[]" replaced with your own identifying 177c478bd9Sstevel@tonic-gate * information: Portions Copyright [yyyy] [name of copyright owner] 187c478bd9Sstevel@tonic-gate * 197c478bd9Sstevel@tonic-gate * CDDL HEADER END 207c478bd9Sstevel@tonic-gate */ 217c478bd9Sstevel@tonic-gate /* 22*17169044Sbrutus * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 237c478bd9Sstevel@tonic-gate * Use is subject to license terms. 247c478bd9Sstevel@tonic-gate */ 257c478bd9Sstevel@tonic-gate 267c478bd9Sstevel@tonic-gate /* Copyright (c) 1984, 1986, 1987, 1988, 1989 AT&T */ 277c478bd9Sstevel@tonic-gate /* All Rights Reserved */ 287c478bd9Sstevel@tonic-gate 297c478bd9Sstevel@tonic-gate /* 307c478bd9Sstevel@tonic-gate * University Copyright- Copyright (c) 1982, 1986, 1988 317c478bd9Sstevel@tonic-gate * The Regents of the University of California 327c478bd9Sstevel@tonic-gate * All Rights Reserved 337c478bd9Sstevel@tonic-gate * 347c478bd9Sstevel@tonic-gate * University Acknowledgment- Portions of this document are derived from 357c478bd9Sstevel@tonic-gate * software developed by the University of California, Berkeley, and its 367c478bd9Sstevel@tonic-gate * contributors. 377c478bd9Sstevel@tonic-gate */ 387c478bd9Sstevel@tonic-gate 397c478bd9Sstevel@tonic-gate #pragma ident "%Z%%M% %I% %E% SMI" 407c478bd9Sstevel@tonic-gate 417c478bd9Sstevel@tonic-gate #include <sys/types.h> 427c478bd9Sstevel@tonic-gate #include <sys/sysmacros.h> 437c478bd9Sstevel@tonic-gate #include <sys/param.h> 447c478bd9Sstevel@tonic-gate #include <sys/systm.h> 457c478bd9Sstevel@tonic-gate #include <sys/uio.h> 467c478bd9Sstevel@tonic-gate #include <sys/errno.h> 47*17169044Sbrutus #include <sys/vmsystm.h> 48*17169044Sbrutus #include <sys/cmn_err.h> 49*17169044Sbrutus #include <vm/as.h> 50*17169044Sbrutus #include <vm/page.h> 51*17169044Sbrutus 52*17169044Sbrutus #include <sys/dcopy.h> 53*17169044Sbrutus 54*17169044Sbrutus int64_t uioa_maxpoll = -1; /* <0 = noblock, 0 = block, >0 = block after */ 55*17169044Sbrutus #define UIO_DCOPY_CHANNEL 0 56*17169044Sbrutus #define UIO_DCOPY_CMD 1 577c478bd9Sstevel@tonic-gate 587c478bd9Sstevel@tonic-gate /* 597c478bd9Sstevel@tonic-gate * Move "n" bytes at byte address "p"; "rw" indicates the direction 607c478bd9Sstevel@tonic-gate * of the move, and the I/O parameters are provided in "uio", which is 617c478bd9Sstevel@tonic-gate * update to reflect the data which was moved. Returns 0 on success or 627c478bd9Sstevel@tonic-gate * a non-zero errno on failure. 637c478bd9Sstevel@tonic-gate */ 647c478bd9Sstevel@tonic-gate int 657c478bd9Sstevel@tonic-gate uiomove(void *p, size_t n, enum uio_rw rw, struct uio *uio) 667c478bd9Sstevel@tonic-gate { 677c478bd9Sstevel@tonic-gate struct iovec *iov; 687c478bd9Sstevel@tonic-gate ulong_t cnt; 697c478bd9Sstevel@tonic-gate int error; 707c478bd9Sstevel@tonic-gate 717c478bd9Sstevel@tonic-gate while (n && uio->uio_resid) { 727c478bd9Sstevel@tonic-gate iov = uio->uio_iov; 737c478bd9Sstevel@tonic-gate cnt = MIN(iov->iov_len, n); 747c478bd9Sstevel@tonic-gate if (cnt == 0l) { 757c478bd9Sstevel@tonic-gate uio->uio_iov++; 767c478bd9Sstevel@tonic-gate uio->uio_iovcnt--; 777c478bd9Sstevel@tonic-gate continue; 787c478bd9Sstevel@tonic-gate } 797c478bd9Sstevel@tonic-gate switch (uio->uio_segflg) { 807c478bd9Sstevel@tonic-gate 817c478bd9Sstevel@tonic-gate case UIO_USERSPACE: 827c478bd9Sstevel@tonic-gate case UIO_USERISPACE: 837c478bd9Sstevel@tonic-gate if (rw == UIO_READ) { 847c478bd9Sstevel@tonic-gate error = xcopyout_nta(p, iov->iov_base, cnt, 857c478bd9Sstevel@tonic-gate (uio->uio_extflg & UIO_COPY_CACHED)); 867c478bd9Sstevel@tonic-gate } else { 877c478bd9Sstevel@tonic-gate error = xcopyin_nta(iov->iov_base, p, cnt, 887c478bd9Sstevel@tonic-gate (uio->uio_extflg & UIO_COPY_CACHED)); 897c478bd9Sstevel@tonic-gate } 907c478bd9Sstevel@tonic-gate 917c478bd9Sstevel@tonic-gate if (error) 927c478bd9Sstevel@tonic-gate return (error); 937c478bd9Sstevel@tonic-gate break; 947c478bd9Sstevel@tonic-gate 957c478bd9Sstevel@tonic-gate case UIO_SYSSPACE: 967c478bd9Sstevel@tonic-gate if (rw == UIO_READ) 977c478bd9Sstevel@tonic-gate error = kcopy_nta(p, iov->iov_base, cnt, 987c478bd9Sstevel@tonic-gate (uio->uio_extflg & UIO_COPY_CACHED)); 997c478bd9Sstevel@tonic-gate else 1007c478bd9Sstevel@tonic-gate error = kcopy_nta(iov->iov_base, p, cnt, 1017c478bd9Sstevel@tonic-gate (uio->uio_extflg & UIO_COPY_CACHED)); 1027c478bd9Sstevel@tonic-gate if (error) 1037c478bd9Sstevel@tonic-gate return (error); 1047c478bd9Sstevel@tonic-gate break; 1057c478bd9Sstevel@tonic-gate } 1067c478bd9Sstevel@tonic-gate iov->iov_base += cnt; 1077c478bd9Sstevel@tonic-gate iov->iov_len -= cnt; 1087c478bd9Sstevel@tonic-gate uio->uio_resid -= cnt; 1097c478bd9Sstevel@tonic-gate uio->uio_loffset += cnt; 1107c478bd9Sstevel@tonic-gate p = (caddr_t)p + cnt; 1117c478bd9Sstevel@tonic-gate n -= cnt; 1127c478bd9Sstevel@tonic-gate } 1137c478bd9Sstevel@tonic-gate return (0); 1147c478bd9Sstevel@tonic-gate } 1157c478bd9Sstevel@tonic-gate 1167c478bd9Sstevel@tonic-gate /* 1177c478bd9Sstevel@tonic-gate * transfer a character value into the address space 1187c478bd9Sstevel@tonic-gate * delineated by a uio and update fields within the 1197c478bd9Sstevel@tonic-gate * uio for next character. Return 0 for success, EFAULT 1207c478bd9Sstevel@tonic-gate * for error. 1217c478bd9Sstevel@tonic-gate */ 1227c478bd9Sstevel@tonic-gate int 1237c478bd9Sstevel@tonic-gate ureadc(int val, struct uio *uiop) 1247c478bd9Sstevel@tonic-gate { 1257c478bd9Sstevel@tonic-gate struct iovec *iovp; 1267c478bd9Sstevel@tonic-gate unsigned char c; 1277c478bd9Sstevel@tonic-gate 1287c478bd9Sstevel@tonic-gate /* 1297c478bd9Sstevel@tonic-gate * first determine if uio is valid. uiop should be 1307c478bd9Sstevel@tonic-gate * non-NULL and the resid count > 0. 1317c478bd9Sstevel@tonic-gate */ 1327c478bd9Sstevel@tonic-gate if (!(uiop && uiop->uio_resid > 0)) 1337c478bd9Sstevel@tonic-gate return (EFAULT); 1347c478bd9Sstevel@tonic-gate 1357c478bd9Sstevel@tonic-gate /* 1367c478bd9Sstevel@tonic-gate * scan through iovecs until one is found that is non-empty. 1377c478bd9Sstevel@tonic-gate * Return EFAULT if none found. 1387c478bd9Sstevel@tonic-gate */ 1397c478bd9Sstevel@tonic-gate while (uiop->uio_iovcnt > 0) { 1407c478bd9Sstevel@tonic-gate iovp = uiop->uio_iov; 1417c478bd9Sstevel@tonic-gate if (iovp->iov_len <= 0) { 1427c478bd9Sstevel@tonic-gate uiop->uio_iovcnt--; 1437c478bd9Sstevel@tonic-gate uiop->uio_iov++; 1447c478bd9Sstevel@tonic-gate } else 1457c478bd9Sstevel@tonic-gate break; 1467c478bd9Sstevel@tonic-gate } 1477c478bd9Sstevel@tonic-gate 1487c478bd9Sstevel@tonic-gate if (uiop->uio_iovcnt <= 0) 1497c478bd9Sstevel@tonic-gate return (EFAULT); 1507c478bd9Sstevel@tonic-gate 1517c478bd9Sstevel@tonic-gate /* 1527c478bd9Sstevel@tonic-gate * Transfer character to uio space. 1537c478bd9Sstevel@tonic-gate */ 1547c478bd9Sstevel@tonic-gate 1557c478bd9Sstevel@tonic-gate c = (unsigned char) (val & 0xFF); 1567c478bd9Sstevel@tonic-gate 1577c478bd9Sstevel@tonic-gate switch (uiop->uio_segflg) { 1587c478bd9Sstevel@tonic-gate 1597c478bd9Sstevel@tonic-gate case UIO_USERISPACE: 1607c478bd9Sstevel@tonic-gate case UIO_USERSPACE: 1617c478bd9Sstevel@tonic-gate if (copyout(&c, iovp->iov_base, sizeof (unsigned char))) 1627c478bd9Sstevel@tonic-gate return (EFAULT); 1637c478bd9Sstevel@tonic-gate break; 1647c478bd9Sstevel@tonic-gate 1657c478bd9Sstevel@tonic-gate case UIO_SYSSPACE: /* can do direct copy since kernel-kernel */ 1667c478bd9Sstevel@tonic-gate *iovp->iov_base = c; 1677c478bd9Sstevel@tonic-gate break; 1687c478bd9Sstevel@tonic-gate 1697c478bd9Sstevel@tonic-gate default: 1707c478bd9Sstevel@tonic-gate return (EFAULT); /* invalid segflg value */ 1717c478bd9Sstevel@tonic-gate } 1727c478bd9Sstevel@tonic-gate 1737c478bd9Sstevel@tonic-gate /* 1747c478bd9Sstevel@tonic-gate * bump up/down iovec and uio members to reflect transfer. 1757c478bd9Sstevel@tonic-gate */ 1767c478bd9Sstevel@tonic-gate iovp->iov_base++; 1777c478bd9Sstevel@tonic-gate iovp->iov_len--; 1787c478bd9Sstevel@tonic-gate uiop->uio_resid--; 1797c478bd9Sstevel@tonic-gate uiop->uio_loffset++; 1807c478bd9Sstevel@tonic-gate return (0); /* success */ 1817c478bd9Sstevel@tonic-gate } 1827c478bd9Sstevel@tonic-gate 1837c478bd9Sstevel@tonic-gate /* 1847c478bd9Sstevel@tonic-gate * return a character value from the address space 1857c478bd9Sstevel@tonic-gate * delineated by a uio and update fields within the 1867c478bd9Sstevel@tonic-gate * uio for next character. Return the character for success, 1877c478bd9Sstevel@tonic-gate * -1 for error. 1887c478bd9Sstevel@tonic-gate */ 1897c478bd9Sstevel@tonic-gate int 1907c478bd9Sstevel@tonic-gate uwritec(struct uio *uiop) 1917c478bd9Sstevel@tonic-gate { 1927c478bd9Sstevel@tonic-gate struct iovec *iovp; 1937c478bd9Sstevel@tonic-gate unsigned char c; 1947c478bd9Sstevel@tonic-gate 1957c478bd9Sstevel@tonic-gate /* 1967c478bd9Sstevel@tonic-gate * verify we were passed a valid uio structure. 1977c478bd9Sstevel@tonic-gate * (1) non-NULL uiop, (2) positive resid count 1987c478bd9Sstevel@tonic-gate * (3) there is an iovec with positive length 1997c478bd9Sstevel@tonic-gate */ 2007c478bd9Sstevel@tonic-gate 2017c478bd9Sstevel@tonic-gate if (!(uiop && uiop->uio_resid > 0)) 2027c478bd9Sstevel@tonic-gate return (-1); 2037c478bd9Sstevel@tonic-gate 2047c478bd9Sstevel@tonic-gate while (uiop->uio_iovcnt > 0) { 2057c478bd9Sstevel@tonic-gate iovp = uiop->uio_iov; 2067c478bd9Sstevel@tonic-gate if (iovp->iov_len <= 0) { 2077c478bd9Sstevel@tonic-gate uiop->uio_iovcnt--; 2087c478bd9Sstevel@tonic-gate uiop->uio_iov++; 2097c478bd9Sstevel@tonic-gate } else 2107c478bd9Sstevel@tonic-gate break; 2117c478bd9Sstevel@tonic-gate } 2127c478bd9Sstevel@tonic-gate 2137c478bd9Sstevel@tonic-gate if (uiop->uio_iovcnt <= 0) 2147c478bd9Sstevel@tonic-gate return (-1); 2157c478bd9Sstevel@tonic-gate 2167c478bd9Sstevel@tonic-gate /* 2177c478bd9Sstevel@tonic-gate * Get the character from the uio address space. 2187c478bd9Sstevel@tonic-gate */ 2197c478bd9Sstevel@tonic-gate switch (uiop->uio_segflg) { 2207c478bd9Sstevel@tonic-gate 2217c478bd9Sstevel@tonic-gate case UIO_USERISPACE: 2227c478bd9Sstevel@tonic-gate case UIO_USERSPACE: 2237c478bd9Sstevel@tonic-gate if (copyin(iovp->iov_base, &c, sizeof (unsigned char))) 2247c478bd9Sstevel@tonic-gate return (-1); 2257c478bd9Sstevel@tonic-gate break; 2267c478bd9Sstevel@tonic-gate 2277c478bd9Sstevel@tonic-gate case UIO_SYSSPACE: 2287c478bd9Sstevel@tonic-gate c = *iovp->iov_base; 2297c478bd9Sstevel@tonic-gate break; 2307c478bd9Sstevel@tonic-gate 2317c478bd9Sstevel@tonic-gate default: 2327c478bd9Sstevel@tonic-gate return (-1); /* invalid segflg */ 2337c478bd9Sstevel@tonic-gate } 2347c478bd9Sstevel@tonic-gate 2357c478bd9Sstevel@tonic-gate /* 2367c478bd9Sstevel@tonic-gate * Adjust fields of iovec and uio appropriately. 2377c478bd9Sstevel@tonic-gate */ 2387c478bd9Sstevel@tonic-gate iovp->iov_base++; 2397c478bd9Sstevel@tonic-gate iovp->iov_len--; 2407c478bd9Sstevel@tonic-gate uiop->uio_resid--; 2417c478bd9Sstevel@tonic-gate uiop->uio_loffset++; 2427c478bd9Sstevel@tonic-gate return ((int)c & 0xFF); /* success */ 2437c478bd9Sstevel@tonic-gate } 2447c478bd9Sstevel@tonic-gate 2457c478bd9Sstevel@tonic-gate /* 2467c478bd9Sstevel@tonic-gate * Drop the next n chars out of *uiop. 2477c478bd9Sstevel@tonic-gate */ 2487c478bd9Sstevel@tonic-gate void 2497c478bd9Sstevel@tonic-gate uioskip(uio_t *uiop, size_t n) 2507c478bd9Sstevel@tonic-gate { 2517c478bd9Sstevel@tonic-gate if (n > uiop->uio_resid) 2527c478bd9Sstevel@tonic-gate return; 2537c478bd9Sstevel@tonic-gate while (n != 0) { 2547c478bd9Sstevel@tonic-gate register iovec_t *iovp = uiop->uio_iov; 2557c478bd9Sstevel@tonic-gate register size_t niovb = MIN(iovp->iov_len, n); 2567c478bd9Sstevel@tonic-gate 2577c478bd9Sstevel@tonic-gate if (niovb == 0) { 2587c478bd9Sstevel@tonic-gate uiop->uio_iov++; 2597c478bd9Sstevel@tonic-gate uiop->uio_iovcnt--; 2607c478bd9Sstevel@tonic-gate continue; 2617c478bd9Sstevel@tonic-gate } 2627c478bd9Sstevel@tonic-gate iovp->iov_base += niovb; 2637c478bd9Sstevel@tonic-gate uiop->uio_loffset += niovb; 2647c478bd9Sstevel@tonic-gate iovp->iov_len -= niovb; 2657c478bd9Sstevel@tonic-gate uiop->uio_resid -= niovb; 2667c478bd9Sstevel@tonic-gate n -= niovb; 2677c478bd9Sstevel@tonic-gate } 2687c478bd9Sstevel@tonic-gate } 2697c478bd9Sstevel@tonic-gate 2707c478bd9Sstevel@tonic-gate /* 2717c478bd9Sstevel@tonic-gate * Dup the suio into the duio and diovec of size diov_cnt. If diov 2727c478bd9Sstevel@tonic-gate * is too small to dup suio then an error will be returned, else 0. 2737c478bd9Sstevel@tonic-gate */ 2747c478bd9Sstevel@tonic-gate int 2757c478bd9Sstevel@tonic-gate uiodup(uio_t *suio, uio_t *duio, iovec_t *diov, int diov_cnt) 2767c478bd9Sstevel@tonic-gate { 2777c478bd9Sstevel@tonic-gate int ix; 2787c478bd9Sstevel@tonic-gate iovec_t *siov = suio->uio_iov; 2797c478bd9Sstevel@tonic-gate 2807c478bd9Sstevel@tonic-gate *duio = *suio; 2817c478bd9Sstevel@tonic-gate for (ix = 0; ix < suio->uio_iovcnt; ix++) { 2827c478bd9Sstevel@tonic-gate diov[ix] = siov[ix]; 2837c478bd9Sstevel@tonic-gate if (ix >= diov_cnt) 2847c478bd9Sstevel@tonic-gate return (1); 2857c478bd9Sstevel@tonic-gate } 2867c478bd9Sstevel@tonic-gate duio->uio_iov = diov; 2877c478bd9Sstevel@tonic-gate return (0); 2887c478bd9Sstevel@tonic-gate } 289*17169044Sbrutus 290*17169044Sbrutus /* 291*17169044Sbrutus * Shadow state for checking if a platform has hardware asynchronous 292*17169044Sbrutus * copy capability and minimum copy size, e.g. Intel's I/OAT dma engine, 293*17169044Sbrutus * 294*17169044Sbrutus * Dcopy does a call-back to uioa_dcopy_enable() when a dma device calls 295*17169044Sbrutus * into dcopy to register and uioa_dcopy_disable() when the device calls 296*17169044Sbrutus * into dcopy to unregister. 297*17169044Sbrutus */ 298*17169044Sbrutus uioasync_t uioasync = {B_FALSE, 1024}; 299*17169044Sbrutus 300*17169044Sbrutus void 301*17169044Sbrutus uioa_dcopy_enable() 302*17169044Sbrutus { 303*17169044Sbrutus uioasync.enabled = B_TRUE; 304*17169044Sbrutus } 305*17169044Sbrutus 306*17169044Sbrutus void 307*17169044Sbrutus uioa_dcopy_disable() 308*17169044Sbrutus { 309*17169044Sbrutus uioasync.enabled = B_FALSE; 310*17169044Sbrutus } 311*17169044Sbrutus 312*17169044Sbrutus /* 313*17169044Sbrutus * Schedule an asynchronous move of "n" bytes at byte address "p", 314*17169044Sbrutus * "rw" indicates the direction of the move, I/O parameters and 315*17169044Sbrutus * async state are provided in "uioa" which is update to reflect 316*17169044Sbrutus * the data which is to be moved. 317*17169044Sbrutus * 318*17169044Sbrutus * Returns 0 on success or a non-zero errno on failure. 319*17169044Sbrutus * 320*17169044Sbrutus * Note, while the uioasync APIs are general purpose in design 321*17169044Sbrutus * the current implementation is Intel I/OAT specific. 322*17169044Sbrutus */ 323*17169044Sbrutus int 324*17169044Sbrutus uioamove(void *p, size_t n, enum uio_rw rw, uioa_t *uioa) 325*17169044Sbrutus { 326*17169044Sbrutus int soff, doff; 327*17169044Sbrutus uint64_t pa; 328*17169044Sbrutus int cnt; 329*17169044Sbrutus iovec_t *iov; 330*17169044Sbrutus dcopy_handle_t channel; 331*17169044Sbrutus dcopy_cmd_t cmd; 332*17169044Sbrutus int ret = 0; 333*17169044Sbrutus int dcopy_flags; 334*17169044Sbrutus 335*17169044Sbrutus if (!(uioa->uioa_state & UIOA_ENABLED)) { 336*17169044Sbrutus /* The uioa_t isn't enabled */ 337*17169044Sbrutus return (ENXIO); 338*17169044Sbrutus } 339*17169044Sbrutus 340*17169044Sbrutus if (uioa->uio_segflg != UIO_USERSPACE || rw != UIO_READ) { 341*17169044Sbrutus /* Only support to user-land from kernel */ 342*17169044Sbrutus return (ENOTSUP); 343*17169044Sbrutus } 344*17169044Sbrutus 345*17169044Sbrutus 346*17169044Sbrutus channel = uioa->uioa_hwst[UIO_DCOPY_CHANNEL]; 347*17169044Sbrutus cmd = uioa->uioa_hwst[UIO_DCOPY_CMD]; 348*17169044Sbrutus dcopy_flags = DCOPY_NOSLEEP; 349*17169044Sbrutus 350*17169044Sbrutus /* 351*17169044Sbrutus * While source bytes and destination bytes. 352*17169044Sbrutus */ 353*17169044Sbrutus while (n > 0 && uioa->uio_resid > 0) { 354*17169044Sbrutus iov = uioa->uio_iov; 355*17169044Sbrutus if (iov->iov_len == 0l) { 356*17169044Sbrutus uioa->uio_iov++; 357*17169044Sbrutus uioa->uio_iovcnt--; 358*17169044Sbrutus uioa->uioa_lcur++; 359*17169044Sbrutus uioa->uioa_lppp = uioa->uioa_lcur->uioa_ppp; 360*17169044Sbrutus continue; 361*17169044Sbrutus } 362*17169044Sbrutus /* 363*17169044Sbrutus * While source bytes schedule an async 364*17169044Sbrutus * dma for destination page by page. 365*17169044Sbrutus */ 366*17169044Sbrutus while (n > 0) { 367*17169044Sbrutus /* Addr offset in page src/dst */ 368*17169044Sbrutus soff = (uintptr_t)p & PAGEOFFSET; 369*17169044Sbrutus doff = (uintptr_t)iov->iov_base & PAGEOFFSET; 370*17169044Sbrutus /* Min copy count src and dst and page sized */ 371*17169044Sbrutus cnt = MIN(n, iov->iov_len); 372*17169044Sbrutus cnt = MIN(cnt, PAGESIZE - soff); 373*17169044Sbrutus cnt = MIN(cnt, PAGESIZE - doff); 374*17169044Sbrutus /* XXX if next page(s) contiguous could use multipage */ 375*17169044Sbrutus 376*17169044Sbrutus /* 377*17169044Sbrutus * if we have an old command, we want to link all 378*17169044Sbrutus * other commands to the next command we alloced so 379*17169044Sbrutus * we only need to track the last command but can 380*17169044Sbrutus * still free them all. 381*17169044Sbrutus */ 382*17169044Sbrutus if (cmd != NULL) { 383*17169044Sbrutus dcopy_flags |= DCOPY_ALLOC_LINK; 384*17169044Sbrutus } 385*17169044Sbrutus ret = dcopy_cmd_alloc(channel, dcopy_flags, &cmd); 386*17169044Sbrutus if (ret != DCOPY_SUCCESS) { 387*17169044Sbrutus /* Error of some sort */ 388*17169044Sbrutus return (EIO); 389*17169044Sbrutus } 390*17169044Sbrutus uioa->uioa_hwst[UIO_DCOPY_CMD] = cmd; 391*17169044Sbrutus 392*17169044Sbrutus ASSERT(cmd->dp_version == DCOPY_CMD_V0); 393*17169044Sbrutus if (uioa_maxpoll >= 0) { 394*17169044Sbrutus /* Blocking (>0 may be) used in uioafini() */ 395*17169044Sbrutus cmd->dp_flags = DCOPY_CMD_INTR; 396*17169044Sbrutus } else { 397*17169044Sbrutus /* Non blocking uioafini() so no intr */ 398*17169044Sbrutus cmd->dp_flags = DCOPY_CMD_NOFLAGS; 399*17169044Sbrutus } 400*17169044Sbrutus cmd->dp_cmd = DCOPY_CMD_COPY; 401*17169044Sbrutus pa = ptob((uint64_t)hat_getpfnum(kas.a_hat, p)); 402*17169044Sbrutus cmd->dp.copy.cc_source = pa + soff; 403*17169044Sbrutus if (uioa->uioa_lcur->uioa_pfncnt == 0) { 404*17169044Sbrutus /* Have a (page_t **) */ 405*17169044Sbrutus pa = ptob((uint64_t)( 406*17169044Sbrutus *(page_t **)uioa->uioa_lppp)->p_pagenum); 407*17169044Sbrutus } else { 408*17169044Sbrutus /* Have a (pfn_t *) */ 409*17169044Sbrutus pa = ptob((uint64_t)( 410*17169044Sbrutus *(pfn_t *)uioa->uioa_lppp)); 411*17169044Sbrutus } 412*17169044Sbrutus cmd->dp.copy.cc_dest = pa + doff; 413*17169044Sbrutus cmd->dp.copy.cc_size = cnt; 414*17169044Sbrutus ret = dcopy_cmd_post(cmd); 415*17169044Sbrutus if (ret != DCOPY_SUCCESS) { 416*17169044Sbrutus /* Error of some sort */ 417*17169044Sbrutus return (EIO); 418*17169044Sbrutus } 419*17169044Sbrutus ret = 0; 420*17169044Sbrutus 421*17169044Sbrutus /* If UIOA_POLL not set, set it */ 422*17169044Sbrutus if (!(uioa->uioa_state & UIOA_POLL)) 423*17169044Sbrutus uioa->uioa_state |= UIOA_POLL; 424*17169044Sbrutus 425*17169044Sbrutus /* Update iov, uio, and local pointers/counters */ 426*17169044Sbrutus iov->iov_base += cnt; 427*17169044Sbrutus iov->iov_len -= cnt; 428*17169044Sbrutus uioa->uio_resid -= cnt; 429*17169044Sbrutus uioa->uio_loffset += cnt; 430*17169044Sbrutus p = (caddr_t)p + cnt; 431*17169044Sbrutus n -= cnt; 432*17169044Sbrutus 433*17169044Sbrutus /* End of iovec? */ 434*17169044Sbrutus if (iov->iov_len == 0) { 435*17169044Sbrutus /* Yup, next iovec */ 436*17169044Sbrutus break; 437*17169044Sbrutus } 438*17169044Sbrutus 439*17169044Sbrutus /* Next dst addr page? */ 440*17169044Sbrutus if (doff + cnt == PAGESIZE) { 441*17169044Sbrutus /* Yup, next page_t */ 442*17169044Sbrutus uioa->uioa_lppp++; 443*17169044Sbrutus } 444*17169044Sbrutus } 445*17169044Sbrutus } 446*17169044Sbrutus 447*17169044Sbrutus return (ret); 448*17169044Sbrutus } 449*17169044Sbrutus 450*17169044Sbrutus /* 451*17169044Sbrutus * Initialize a uioa_t for a given uio_t for the current user context, 452*17169044Sbrutus * copy the common uio_t to the uioa_t, walk the shared iovec_t and 453*17169044Sbrutus * lock down the user-land page(s) containing iovec_t data, then mapin 454*17169044Sbrutus * user-land pages using segkpm. 455*17169044Sbrutus */ 456*17169044Sbrutus int 457*17169044Sbrutus uioainit(uio_t *uiop, uioa_t *uioap) 458*17169044Sbrutus { 459*17169044Sbrutus caddr_t addr; 460*17169044Sbrutus page_t **pages; 461*17169044Sbrutus int off; 462*17169044Sbrutus int len; 463*17169044Sbrutus proc_t *procp = ttoproc(curthread); 464*17169044Sbrutus struct as *as = procp->p_as; 465*17169044Sbrutus iovec_t *iov = uiop->uio_iov; 466*17169044Sbrutus int32_t iovcnt = uiop->uio_iovcnt; 467*17169044Sbrutus uioa_page_t *locked = uioap->uioa_locked; 468*17169044Sbrutus dcopy_handle_t channel; 469*17169044Sbrutus int error; 470*17169044Sbrutus 471*17169044Sbrutus if (! (uioap->uioa_state & UIOA_ALLOC)) { 472*17169044Sbrutus /* Can only init() a freshly allocated uioa_t */ 473*17169044Sbrutus return (EINVAL); 474*17169044Sbrutus } 475*17169044Sbrutus 476*17169044Sbrutus error = dcopy_alloc(DCOPY_NOSLEEP, &channel); 477*17169044Sbrutus if (error == DCOPY_NORESOURCES) { 478*17169044Sbrutus /* Turn off uioa */ 479*17169044Sbrutus uioasync.enabled = B_FALSE; 480*17169044Sbrutus return (ENODEV); 481*17169044Sbrutus } 482*17169044Sbrutus if (error != DCOPY_SUCCESS) { 483*17169044Sbrutus /* Alloc failed */ 484*17169044Sbrutus return (EIO); 485*17169044Sbrutus } 486*17169044Sbrutus 487*17169044Sbrutus uioap->uioa_hwst[UIO_DCOPY_CHANNEL] = channel; 488*17169044Sbrutus uioap->uioa_hwst[UIO_DCOPY_CMD] = NULL; 489*17169044Sbrutus 490*17169044Sbrutus /* Indicate uioa_t (will be) initialized */ 491*17169044Sbrutus uioap->uioa_state = UIOA_INIT; 492*17169044Sbrutus 493*17169044Sbrutus /* uio_t/uioa_t uio_t common struct copy */ 494*17169044Sbrutus *((uio_t *)uioap) = *uiop; 495*17169044Sbrutus 496*17169044Sbrutus /* initialize *uiop->uio_iov */ 497*17169044Sbrutus if (iovcnt > UIOA_IOV_MAX) { 498*17169044Sbrutus /* Too big? */ 499*17169044Sbrutus return (E2BIG); 500*17169044Sbrutus } 501*17169044Sbrutus uioap->uio_iov = iov; 502*17169044Sbrutus uioap->uio_iovcnt = iovcnt; 503*17169044Sbrutus 504*17169044Sbrutus /* Mark the uioap as such */ 505*17169044Sbrutus uioap->uio_extflg |= UIO_ASYNC; 506*17169044Sbrutus 507*17169044Sbrutus /* 508*17169044Sbrutus * For each iovec_t, lock-down the page(s) backing the iovec_t 509*17169044Sbrutus * and save the page_t list for phys addr use in uioamove(). 510*17169044Sbrutus */ 511*17169044Sbrutus iov = uiop->uio_iov; 512*17169044Sbrutus iovcnt = uiop->uio_iovcnt; 513*17169044Sbrutus while (iovcnt > 0) { 514*17169044Sbrutus addr = iov->iov_base; 515*17169044Sbrutus off = (uintptr_t)addr & PAGEOFFSET; 516*17169044Sbrutus addr = (caddr_t)((uintptr_t)addr & (uintptr_t)PAGEMASK); 517*17169044Sbrutus len = iov->iov_len + off; 518*17169044Sbrutus 519*17169044Sbrutus /* Lock down page(s) for the iov span */ 520*17169044Sbrutus if ((error = as_pagelock(as, &pages, 521*17169044Sbrutus iov->iov_base, iov->iov_len, S_WRITE)) != 0) { 522*17169044Sbrutus /* Error */ 523*17169044Sbrutus goto cleanup; 524*17169044Sbrutus } 525*17169044Sbrutus 526*17169044Sbrutus if (pages == NULL) { 527*17169044Sbrutus /* 528*17169044Sbrutus * Need page_t list, really only need 529*17169044Sbrutus * a pfn list so build one. 530*17169044Sbrutus */ 531*17169044Sbrutus pfn_t *pfnp; 532*17169044Sbrutus int pcnt = len >> PAGESHIFT; 533*17169044Sbrutus 534*17169044Sbrutus if (off) 535*17169044Sbrutus pcnt++; 536*17169044Sbrutus if ((pfnp = kmem_alloc(pcnt * sizeof (pfnp), 537*17169044Sbrutus KM_NOSLEEP)) == NULL) { 538*17169044Sbrutus error = ENOMEM; 539*17169044Sbrutus goto cleanup; 540*17169044Sbrutus } 541*17169044Sbrutus locked->uioa_ppp = (void **)pfnp; 542*17169044Sbrutus locked->uioa_pfncnt = pcnt; 543*17169044Sbrutus AS_LOCK_ENTER(as, &as->a_lock, RW_READER); 544*17169044Sbrutus while (pcnt-- > 0) { 545*17169044Sbrutus *pfnp++ = hat_getpfnum(as->a_hat, addr); 546*17169044Sbrutus addr += PAGESIZE; 547*17169044Sbrutus } 548*17169044Sbrutus AS_LOCK_EXIT(as, &as->a_lock); 549*17169044Sbrutus } else { 550*17169044Sbrutus /* Have a page_t list, save it */ 551*17169044Sbrutus locked->uioa_ppp = (void **)pages; 552*17169044Sbrutus locked->uioa_pfncnt = 0; 553*17169044Sbrutus } 554*17169044Sbrutus /* Save for as_pageunlock() in uioafini() */ 555*17169044Sbrutus locked->uioa_base = iov->iov_base; 556*17169044Sbrutus locked->uioa_len = iov->iov_len; 557*17169044Sbrutus locked++; 558*17169044Sbrutus 559*17169044Sbrutus /* Next iovec_t */ 560*17169044Sbrutus iov++; 561*17169044Sbrutus iovcnt--; 562*17169044Sbrutus } 563*17169044Sbrutus /* Initialize curret pointer into uioa_locked[] and it's uioa_ppp */ 564*17169044Sbrutus uioap->uioa_lcur = uioap->uioa_locked; 565*17169044Sbrutus uioap->uioa_lppp = uioap->uioa_lcur->uioa_ppp; 566*17169044Sbrutus return (0); 567*17169044Sbrutus 568*17169044Sbrutus cleanup: 569*17169044Sbrutus /* Unlock any previously locked page_t(s) */ 570*17169044Sbrutus while (locked > uioap->uioa_locked) { 571*17169044Sbrutus locked--; 572*17169044Sbrutus as_pageunlock(as, (page_t **)locked->uioa_ppp, 573*17169044Sbrutus locked->uioa_base, locked->uioa_len, S_WRITE); 574*17169044Sbrutus } 575*17169044Sbrutus 576*17169044Sbrutus /* Last indicate uioa_t still in alloc state */ 577*17169044Sbrutus uioap->uioa_state = UIOA_ALLOC; 578*17169044Sbrutus 579*17169044Sbrutus return (error); 580*17169044Sbrutus } 581*17169044Sbrutus 582*17169044Sbrutus /* 583*17169044Sbrutus * Finish processing of a uioa_t by cleanup any pending "uioap" actions. 584*17169044Sbrutus */ 585*17169044Sbrutus int 586*17169044Sbrutus uioafini(uio_t *uiop, uioa_t *uioap) 587*17169044Sbrutus { 588*17169044Sbrutus int32_t iovcnt = uiop->uio_iovcnt; 589*17169044Sbrutus uioa_page_t *locked = uioap->uioa_locked; 590*17169044Sbrutus struct as *as = ttoproc(curthread)->p_as; 591*17169044Sbrutus dcopy_handle_t channel; 592*17169044Sbrutus dcopy_cmd_t cmd; 593*17169044Sbrutus int ret = 0; 594*17169044Sbrutus 595*17169044Sbrutus ASSERT(uioap->uio_extflg & UIO_ASYNC); 596*17169044Sbrutus 597*17169044Sbrutus if (!(uioap->uioa_state & (UIOA_ENABLED|UIOA_FINI))) { 598*17169044Sbrutus /* Must be an active uioa_t */ 599*17169044Sbrutus return (EINVAL); 600*17169044Sbrutus } 601*17169044Sbrutus 602*17169044Sbrutus channel = uioap->uioa_hwst[UIO_DCOPY_CHANNEL]; 603*17169044Sbrutus cmd = uioap->uioa_hwst[UIO_DCOPY_CMD]; 604*17169044Sbrutus 605*17169044Sbrutus /* XXX - why do we get cmd == NULL sometimes? */ 606*17169044Sbrutus if (cmd != NULL) { 607*17169044Sbrutus if (uioap->uioa_state & UIOA_POLL) { 608*17169044Sbrutus /* Wait for last dcopy() to finish */ 609*17169044Sbrutus int64_t poll = 1; 610*17169044Sbrutus int poll_flag = DCOPY_POLL_NOFLAGS; 611*17169044Sbrutus 612*17169044Sbrutus do { 613*17169044Sbrutus if (uioa_maxpoll == 0 || 614*17169044Sbrutus (uioa_maxpoll > 0 && 615*17169044Sbrutus poll >= uioa_maxpoll)) { 616*17169044Sbrutus /* Always block or after maxpoll */ 617*17169044Sbrutus poll_flag = DCOPY_POLL_BLOCK; 618*17169044Sbrutus } else { 619*17169044Sbrutus /* No block, poll */ 620*17169044Sbrutus poll++; 621*17169044Sbrutus } 622*17169044Sbrutus ret = dcopy_cmd_poll(cmd, poll_flag); 623*17169044Sbrutus } while (ret == DCOPY_PENDING); 624*17169044Sbrutus 625*17169044Sbrutus if (ret == DCOPY_COMPLETED) { 626*17169044Sbrutus /* Poll/block succeeded */ 627*17169044Sbrutus ret = 0; 628*17169044Sbrutus } else { 629*17169044Sbrutus /* Poll/block failed */ 630*17169044Sbrutus ret = EIO; 631*17169044Sbrutus } 632*17169044Sbrutus } 633*17169044Sbrutus dcopy_cmd_free(&cmd); 634*17169044Sbrutus } 635*17169044Sbrutus 636*17169044Sbrutus dcopy_free(&channel); 637*17169044Sbrutus 638*17169044Sbrutus /* Unlock all page(s) iovec_t by iovec_t */ 639*17169044Sbrutus while (iovcnt-- > 0) { 640*17169044Sbrutus page_t **pages; 641*17169044Sbrutus 642*17169044Sbrutus if (locked->uioa_pfncnt == 0) { 643*17169044Sbrutus /* A as_pagelock() returned (page_t **) */ 644*17169044Sbrutus pages = (page_t **)locked->uioa_ppp; 645*17169044Sbrutus } else { 646*17169044Sbrutus /* Our pfn_t array */ 647*17169044Sbrutus pages = NULL; 648*17169044Sbrutus kmem_free(locked->uioa_ppp, locked->uioa_pfncnt * 649*17169044Sbrutus sizeof (pfn_t *)); 650*17169044Sbrutus } 651*17169044Sbrutus as_pageunlock(as, pages, locked->uioa_base, locked->uioa_len, 652*17169044Sbrutus S_WRITE); 653*17169044Sbrutus 654*17169044Sbrutus locked++; 655*17169044Sbrutus } 656*17169044Sbrutus /* uioa_t->uio_t common struct copy */ 657*17169044Sbrutus *uiop = *((uio_t *)uioap); 658*17169044Sbrutus 659*17169044Sbrutus /* 660*17169044Sbrutus * Last, reset uioa state to alloc. 661*17169044Sbrutus * 662*17169044Sbrutus * Note, we only initialize the state here, all other members 663*17169044Sbrutus * will be initialized in a subsequent uioainit(). 664*17169044Sbrutus */ 665*17169044Sbrutus uioap->uioa_state = UIOA_ALLOC; 666*17169044Sbrutus 667*17169044Sbrutus uioap->uioa_hwst[UIO_DCOPY_CMD] = NULL; 668*17169044Sbrutus uioap->uioa_hwst[UIO_DCOPY_CHANNEL] = NULL; 669*17169044Sbrutus 670*17169044Sbrutus return (ret); 671*17169044Sbrutus } 672