1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2010 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 * Copyright (c) 2015, Joyent, Inc. All rights reserved. 26 */ 27 28 /* Copyright (c) 2013, OmniTI Computer Consulting, Inc. All rights reserved. */ 29 30 #include "lint.h" 31 #include "thr_uberdata.h" 32 #include <stdarg.h> 33 #include <poll.h> 34 #include <stropts.h> 35 #include <dlfcn.h> 36 #include <wait.h> 37 #include <sys/socket.h> 38 #include <sys/uio.h> 39 #include <sys/file.h> 40 #include <sys/door.h> 41 42 /* 43 * These leading-underbar symbols exist because mistakes were made 44 * in the past that put them into non-SUNWprivate versions of 45 * the libc mapfiles. They should be eliminated, but oh well... 46 */ 47 #pragma weak _fork = fork 48 #pragma weak _read = read 49 #pragma weak _write = write 50 #pragma weak _getmsg = getmsg 51 #pragma weak _getpmsg = getpmsg 52 #pragma weak _putmsg = putmsg 53 #pragma weak _putpmsg = putpmsg 54 #pragma weak _sleep = sleep 55 #pragma weak _close = close 56 #pragma weak _creat = creat 57 #pragma weak _fcntl = fcntl 58 #pragma weak _fsync = fsync 59 #pragma weak _lockf = lockf 60 #pragma weak _msgrcv = msgrcv 61 #pragma weak _msgsnd = msgsnd 62 #pragma weak _msync = msync 63 #pragma weak _open = open 64 #pragma weak _openat = openat 65 #pragma weak _pause = pause 66 #pragma weak _readv = readv 67 #pragma weak _sigpause = sigpause 68 #pragma weak _sigsuspend = sigsuspend 69 #pragma weak _tcdrain = tcdrain 70 #pragma weak _waitid = waitid 71 #pragma weak _writev = writev 72 73 #if !defined(_LP64) 74 #pragma weak _creat64 = creat64 75 #pragma weak _lockf64 = lockf64 76 #pragma weak _open64 = open64 77 #pragma weak _openat64 = openat64 78 #pragma weak _pread64 = pread64 79 #pragma weak _pwrite64 = pwrite64 80 #endif 81 82 /* 83 * These are SUNWprivate, but they are being used by Sun Studio libcollector. 84 */ 85 #pragma weak _fork1 = fork1 86 #pragma weak _forkall = forkall 87 88 /* 89 * atfork_lock protects the pthread_atfork() data structures. 90 * 91 * fork_lock does double-duty. Not only does it (and atfork_lock) 92 * serialize calls to fork() and forkall(), but it also serializes calls 93 * to thr_suspend() and thr_continue() (because fork() and forkall() also 94 * suspend and continue other threads and they want no competition). 95 * 96 * Functions called in dlopen()ed L10N objects can do anything, including 97 * call malloc() and free(). Such calls are not fork-safe when protected 98 * by an ordinary mutex that is acquired in libc's prefork processing 99 * because, with an interposed malloc library present, there would be a 100 * lock ordering violation due to the pthread_atfork() prefork function 101 * in the interposition library acquiring its malloc lock(s) before the 102 * ordinary mutex in libc being acquired by libc's prefork functions. 103 * 104 * Within libc, calls to malloc() and free() are fork-safe if the calls 105 * are made while holding no other libc locks. This covers almost all 106 * of libc's malloc() and free() calls. For those libc code paths, such 107 * as the above-mentioned L10N calls, that require serialization and that 108 * may call malloc() or free(), libc uses callout_lock_enter() to perform 109 * the serialization. This works because callout_lock is not acquired as 110 * part of running the pthread_atfork() prefork handlers (to avoid the 111 * lock ordering violation described above). Rather, it is simply 112 * reinitialized in postfork1_child() to cover the case that some 113 * now-defunct thread might have been suspended while holding it. 114 */ 115 116 void 117 fork_lock_enter(void) 118 { 119 ASSERT(curthread->ul_critical == 0); 120 (void) mutex_lock(&curthread->ul_uberdata->fork_lock); 121 } 122 123 void 124 fork_lock_exit(void) 125 { 126 ASSERT(curthread->ul_critical == 0); 127 (void) mutex_unlock(&curthread->ul_uberdata->fork_lock); 128 } 129 130 /* 131 * Use cancel_safe_mutex_lock() to protect against being cancelled while 132 * holding callout_lock and calling outside of libc (via L10N plugins). 133 * We will honor a pending cancellation request when callout_lock_exit() 134 * is called, by calling cancel_safe_mutex_unlock(). 135 */ 136 void 137 callout_lock_enter(void) 138 { 139 ASSERT(curthread->ul_critical == 0); 140 cancel_safe_mutex_lock(&curthread->ul_uberdata->callout_lock); 141 } 142 143 void 144 callout_lock_exit(void) 145 { 146 ASSERT(curthread->ul_critical == 0); 147 cancel_safe_mutex_unlock(&curthread->ul_uberdata->callout_lock); 148 } 149 150 pid_t 151 forkx(int flags) 152 { 153 ulwp_t *self = curthread; 154 uberdata_t *udp = self->ul_uberdata; 155 pid_t pid; 156 157 if (self->ul_vfork) { 158 /* 159 * We are a child of vfork(); omit all of the fork 160 * logic and go straight to the system call trap. 161 * A vfork() child of a multithreaded parent 162 * must never call fork(). 163 */ 164 if (udp->uberflags.uf_mt) { 165 errno = ENOTSUP; 166 return (-1); 167 } 168 pid = __forkx(flags); 169 if (pid == 0) { /* child */ 170 udp->pid = getpid(); 171 self->ul_vfork = 0; 172 } 173 return (pid); 174 } 175 176 sigoff(self); 177 if (self->ul_fork) { 178 /* 179 * Cannot call fork() from a fork handler. 180 */ 181 sigon(self); 182 errno = EDEADLK; 183 return (-1); 184 } 185 self->ul_fork = 1; 186 187 /* 188 * The functions registered by pthread_atfork() are defined by 189 * the application and its libraries and we must not hold any 190 * internal lmutex_lock()-acquired locks while invoking them. 191 * We hold only udp->atfork_lock to protect the atfork linkages. 192 * If one of these pthread_atfork() functions attempts to fork 193 * or to call pthread_atfork(), libc will detect the error and 194 * fail the call with EDEADLK. Otherwise, the pthread_atfork() 195 * functions are free to do anything they please (except they 196 * will not receive any signals). 197 */ 198 (void) mutex_lock(&udp->atfork_lock); 199 200 /* 201 * Posix (SUSv3) requires fork() to be async-signal-safe. 202 * This cannot be made to happen with fork handlers in place 203 * (they grab locks). To be in nominal compliance, don't run 204 * any fork handlers if we are called within a signal context. 205 * This leaves the child process in a questionable state with 206 * respect to its locks, but at least the parent process does 207 * not become deadlocked due to the calling thread attempting 208 * to acquire a lock that it already owns. 209 */ 210 if (self->ul_siglink == NULL) 211 _prefork_handler(); 212 213 /* 214 * Block every other thread attempting thr_suspend() or thr_continue(). 215 */ 216 (void) mutex_lock(&udp->fork_lock); 217 218 /* 219 * Block all signals. 220 * Just deferring them via sigoff() is not enough. 221 * We have to avoid taking a deferred signal in the child 222 * that was actually sent to the parent before __forkx(). 223 */ 224 block_all_signals(self); 225 226 /* 227 * This suspends all threads but this one, leaving them 228 * suspended outside of any critical regions in the library. 229 * Thus, we are assured that no lmutex_lock()-acquired library 230 * locks are held while we invoke fork() from the current thread. 231 */ 232 suspend_fork(); 233 234 pid = __forkx(flags); 235 236 if (pid == 0) { /* child */ 237 /* 238 * Clear our schedctl pointer. 239 * Discard any deferred signal that was sent to the parent. 240 * Because we blocked all signals before __forkx(), a 241 * deferred signal cannot have been taken by the child. 242 */ 243 self->ul_schedctl_called = NULL; 244 self->ul_schedctl = NULL; 245 self->ul_cursig = 0; 246 self->ul_siginfo.si_signo = 0; 247 udp->pid = getpid(); 248 /* reset the library's data structures to reflect one thread */ 249 unregister_locks(); 250 postfork1_child(); 251 restore_signals(self); 252 (void) mutex_unlock(&udp->fork_lock); 253 if (self->ul_siglink == NULL) 254 _postfork_child_handler(); 255 } else { 256 /* restart all threads that were suspended for fork() */ 257 continue_fork(0); 258 restore_signals(self); 259 (void) mutex_unlock(&udp->fork_lock); 260 if (self->ul_siglink == NULL) 261 _postfork_parent_handler(); 262 } 263 264 (void) mutex_unlock(&udp->atfork_lock); 265 self->ul_fork = 0; 266 sigon(self); 267 268 return (pid); 269 } 270 271 /* 272 * fork() is fork1() for both Posix threads and Solaris threads. 273 * The forkall() interface exists for applications that require 274 * the semantics of replicating all threads. 275 */ 276 #pragma weak fork1 = fork 277 pid_t 278 fork(void) 279 { 280 return (forkx(0)); 281 } 282 283 /* 284 * Much of the logic here is the same as in forkx(). 285 * See the comments in forkx(), above. 286 */ 287 pid_t 288 forkallx(int flags) 289 { 290 ulwp_t *self = curthread; 291 uberdata_t *udp = self->ul_uberdata; 292 pid_t pid; 293 294 if (self->ul_vfork) { 295 if (udp->uberflags.uf_mt) { 296 errno = ENOTSUP; 297 return (-1); 298 } 299 pid = __forkallx(flags); 300 if (pid == 0) { /* child */ 301 udp->pid = getpid(); 302 self->ul_vfork = 0; 303 } 304 return (pid); 305 } 306 307 sigoff(self); 308 if (self->ul_fork) { 309 sigon(self); 310 errno = EDEADLK; 311 return (-1); 312 } 313 self->ul_fork = 1; 314 (void) mutex_lock(&udp->atfork_lock); 315 (void) mutex_lock(&udp->fork_lock); 316 block_all_signals(self); 317 suspend_fork(); 318 319 pid = __forkallx(flags); 320 321 if (pid == 0) { 322 self->ul_schedctl_called = NULL; 323 self->ul_schedctl = NULL; 324 self->ul_cursig = 0; 325 self->ul_siginfo.si_signo = 0; 326 udp->pid = getpid(); 327 unregister_locks(); 328 continue_fork(1); 329 } else { 330 continue_fork(0); 331 } 332 restore_signals(self); 333 (void) mutex_unlock(&udp->fork_lock); 334 (void) mutex_unlock(&udp->atfork_lock); 335 self->ul_fork = 0; 336 sigon(self); 337 338 return (pid); 339 } 340 341 pid_t 342 forkall(void) 343 { 344 return (forkallx(0)); 345 } 346 347 /* 348 * For the implementation of cancellation at cancellation points. 349 */ 350 #define PROLOGUE \ 351 { \ 352 ulwp_t *self = curthread; \ 353 int nocancel = \ 354 (self->ul_vfork | self->ul_nocancel | self->ul_libc_locks | \ 355 self->ul_critical | self->ul_sigdefer); \ 356 int abort = 0; \ 357 if (nocancel == 0) { \ 358 self->ul_save_async = self->ul_cancel_async; \ 359 if (!self->ul_cancel_disabled) { \ 360 self->ul_cancel_async = 1; \ 361 if (self->ul_cancel_pending) \ 362 pthread_exit(PTHREAD_CANCELED); \ 363 } \ 364 self->ul_sp = stkptr(); \ 365 } else if (self->ul_cancel_pending && \ 366 !self->ul_cancel_disabled) { \ 367 set_cancel_eintr_flag(self); \ 368 abort = 1; \ 369 } 370 371 #define EPILOGUE \ 372 if (nocancel == 0) { \ 373 self->ul_sp = 0; \ 374 self->ul_cancel_async = self->ul_save_async; \ 375 } \ 376 } 377 378 /* 379 * Perform the body of the action required by most of the cancelable 380 * function calls. The return(function_call) part is to allow the 381 * compiler to make the call be executed with tail recursion, which 382 * saves a register window on sparc and slightly (not much) improves 383 * the code for x86/x64 compilations. 384 */ 385 #define PERFORM(function_call) \ 386 PROLOGUE \ 387 if (abort) { \ 388 *self->ul_errnop = EINTR; \ 389 return (-1); \ 390 } \ 391 if (nocancel) \ 392 return (function_call); \ 393 rv = function_call; \ 394 EPILOGUE \ 395 return (rv); 396 397 /* 398 * Specialized prologue for sigsuspend() and pollsys(). 399 * These system calls pass a signal mask to the kernel. 400 * The kernel replaces the thread's signal mask with the 401 * temporary mask before the thread goes to sleep. If 402 * a signal is received, the signal handler will execute 403 * with the temporary mask, as modified by the sigaction 404 * for the particular signal. 405 * 406 * We block all signals until we reach the kernel with the 407 * temporary mask. This eliminates race conditions with 408 * setting the signal mask while signals are being posted. 409 */ 410 #define PROLOGUE_MASK(sigmask) \ 411 { \ 412 ulwp_t *self = curthread; \ 413 int nocancel = \ 414 (self->ul_vfork | self->ul_nocancel | self->ul_libc_locks | \ 415 self->ul_critical | self->ul_sigdefer); \ 416 if (!self->ul_vfork) { \ 417 if (sigmask) { \ 418 block_all_signals(self); \ 419 self->ul_tmpmask = *sigmask; \ 420 delete_reserved_signals(&self->ul_tmpmask); \ 421 self->ul_sigsuspend = 1; \ 422 } \ 423 if (nocancel == 0) { \ 424 self->ul_save_async = self->ul_cancel_async; \ 425 if (!self->ul_cancel_disabled) { \ 426 self->ul_cancel_async = 1; \ 427 if (self->ul_cancel_pending) { \ 428 if (self->ul_sigsuspend) { \ 429 self->ul_sigsuspend = 0;\ 430 restore_signals(self); \ 431 } \ 432 pthread_exit(PTHREAD_CANCELED); \ 433 } \ 434 } \ 435 self->ul_sp = stkptr(); \ 436 } \ 437 } 438 439 /* 440 * If a signal is taken, we return from the system call wrapper with 441 * our original signal mask restored (see code in call_user_handler()). 442 * If not (self->ul_sigsuspend is still non-zero), we must restore our 443 * original signal mask ourself. 444 */ 445 #define EPILOGUE_MASK \ 446 if (nocancel == 0) { \ 447 self->ul_sp = 0; \ 448 self->ul_cancel_async = self->ul_save_async; \ 449 } \ 450 if (self->ul_sigsuspend) { \ 451 self->ul_sigsuspend = 0; \ 452 restore_signals(self); \ 453 } \ 454 } 455 456 /* 457 * Cancellation prologue and epilogue functions, 458 * for cancellation points too complex to include here. 459 */ 460 void 461 _cancel_prologue(void) 462 { 463 ulwp_t *self = curthread; 464 465 self->ul_cancel_prologue = 466 (self->ul_vfork | self->ul_nocancel | self->ul_libc_locks | 467 self->ul_critical | self->ul_sigdefer) != 0; 468 if (self->ul_cancel_prologue == 0) { 469 self->ul_save_async = self->ul_cancel_async; 470 if (!self->ul_cancel_disabled) { 471 self->ul_cancel_async = 1; 472 if (self->ul_cancel_pending) 473 pthread_exit(PTHREAD_CANCELED); 474 } 475 self->ul_sp = stkptr(); 476 } else if (self->ul_cancel_pending && 477 !self->ul_cancel_disabled) { 478 set_cancel_eintr_flag(self); 479 } 480 } 481 482 void 483 _cancel_epilogue(void) 484 { 485 ulwp_t *self = curthread; 486 487 if (self->ul_cancel_prologue == 0) { 488 self->ul_sp = 0; 489 self->ul_cancel_async = self->ul_save_async; 490 } 491 } 492 493 /* 494 * Called from _thrp_join() (thr_join() is a cancellation point) 495 */ 496 int 497 lwp_wait(thread_t tid, thread_t *found) 498 { 499 int error; 500 501 PROLOGUE 502 if (abort) 503 return (EINTR); 504 while ((error = __lwp_wait(tid, found)) == EINTR && !cancel_active()) 505 continue; 506 EPILOGUE 507 return (error); 508 } 509 510 ssize_t 511 read(int fd, void *buf, size_t size) 512 { 513 extern ssize_t __read(int, void *, size_t); 514 ssize_t rv; 515 516 PERFORM(__read(fd, buf, size)) 517 } 518 519 ssize_t 520 write(int fd, const void *buf, size_t size) 521 { 522 extern ssize_t __write(int, const void *, size_t); 523 ssize_t rv; 524 525 PERFORM(__write(fd, buf, size)) 526 } 527 528 int 529 getmsg(int fd, struct strbuf *ctlptr, struct strbuf *dataptr, 530 int *flagsp) 531 { 532 extern int __getmsg(int, struct strbuf *, struct strbuf *, int *); 533 int rv; 534 535 PERFORM(__getmsg(fd, ctlptr, dataptr, flagsp)) 536 } 537 538 int 539 getpmsg(int fd, struct strbuf *ctlptr, struct strbuf *dataptr, 540 int *bandp, int *flagsp) 541 { 542 extern int __getpmsg(int, struct strbuf *, struct strbuf *, 543 int *, int *); 544 int rv; 545 546 PERFORM(__getpmsg(fd, ctlptr, dataptr, bandp, flagsp)) 547 } 548 549 int 550 putmsg(int fd, const struct strbuf *ctlptr, 551 const struct strbuf *dataptr, int flags) 552 { 553 extern int __putmsg(int, const struct strbuf *, 554 const struct strbuf *, int); 555 int rv; 556 557 PERFORM(__putmsg(fd, ctlptr, dataptr, flags)) 558 } 559 560 int 561 __xpg4_putmsg(int fd, const struct strbuf *ctlptr, 562 const struct strbuf *dataptr, int flags) 563 { 564 extern int __putmsg(int, const struct strbuf *, 565 const struct strbuf *, int); 566 int rv; 567 568 PERFORM(__putmsg(fd, ctlptr, dataptr, flags|MSG_XPG4)) 569 } 570 571 int 572 putpmsg(int fd, const struct strbuf *ctlptr, 573 const struct strbuf *dataptr, int band, int flags) 574 { 575 extern int __putpmsg(int, const struct strbuf *, 576 const struct strbuf *, int, int); 577 int rv; 578 579 PERFORM(__putpmsg(fd, ctlptr, dataptr, band, flags)) 580 } 581 582 int 583 __xpg4_putpmsg(int fd, const struct strbuf *ctlptr, 584 const struct strbuf *dataptr, int band, int flags) 585 { 586 extern int __putpmsg(int, const struct strbuf *, 587 const struct strbuf *, int, int); 588 int rv; 589 590 PERFORM(__putpmsg(fd, ctlptr, dataptr, band, flags|MSG_XPG4)) 591 } 592 593 int 594 nanosleep(const timespec_t *rqtp, timespec_t *rmtp) 595 { 596 int error; 597 598 PROLOGUE 599 error = abort? EINTR : __nanosleep(rqtp, rmtp); 600 EPILOGUE 601 if (error) { 602 errno = error; 603 return (-1); 604 } 605 return (0); 606 } 607 608 int 609 clock_nanosleep(clockid_t clock_id, int flags, 610 const timespec_t *rqtp, timespec_t *rmtp) 611 { 612 timespec_t reltime; 613 hrtime_t start; 614 hrtime_t rqlapse; 615 hrtime_t lapse; 616 int error; 617 618 switch (clock_id) { 619 case CLOCK_VIRTUAL: 620 case CLOCK_PROCESS_CPUTIME_ID: 621 case CLOCK_THREAD_CPUTIME_ID: 622 return (ENOTSUP); 623 case CLOCK_REALTIME: 624 case CLOCK_HIGHRES: 625 break; 626 default: 627 return (EINVAL); 628 } 629 if (flags & TIMER_ABSTIME) { 630 abstime_to_reltime(clock_id, rqtp, &reltime); 631 rmtp = NULL; 632 } else { 633 reltime = *rqtp; 634 if (clock_id == CLOCK_HIGHRES) 635 start = gethrtime(); 636 } 637 restart: 638 PROLOGUE 639 error = abort? EINTR : __nanosleep(&reltime, rmtp); 640 EPILOGUE 641 if (error == 0 && clock_id == CLOCK_HIGHRES) { 642 /* 643 * Don't return yet if we didn't really get a timeout. 644 * This can happen if we return because someone resets 645 * the system clock. 646 */ 647 if (flags & TIMER_ABSTIME) { 648 if ((hrtime_t)(uint32_t)rqtp->tv_sec * NANOSEC + 649 rqtp->tv_nsec > gethrtime()) { 650 abstime_to_reltime(clock_id, rqtp, &reltime); 651 goto restart; 652 } 653 } else { 654 rqlapse = (hrtime_t)(uint32_t)rqtp->tv_sec * NANOSEC + 655 rqtp->tv_nsec; 656 lapse = gethrtime() - start; 657 if (rqlapse > lapse) { 658 hrt2ts(rqlapse - lapse, &reltime); 659 goto restart; 660 } 661 } 662 } 663 if (error == 0 && clock_id == CLOCK_REALTIME && 664 (flags & TIMER_ABSTIME)) { 665 /* 666 * Don't return yet just because someone reset the 667 * system clock. Recompute the new relative time 668 * and reissue the nanosleep() call if necessary. 669 * 670 * Resetting the system clock causes all sorts of 671 * problems and the SUSV3 standards body should 672 * have made the behavior of clock_nanosleep() be 673 * implementation-defined in such a case rather than 674 * being specific about honoring the new system time. 675 * Standards bodies are filled with fools and idiots. 676 */ 677 abstime_to_reltime(clock_id, rqtp, &reltime); 678 if (reltime.tv_sec != 0 || reltime.tv_nsec != 0) 679 goto restart; 680 } 681 return (error); 682 } 683 684 unsigned int 685 sleep(unsigned int sec) 686 { 687 unsigned int rem = 0; 688 timespec_t ts; 689 timespec_t tsr; 690 691 ts.tv_sec = (time_t)sec; 692 ts.tv_nsec = 0; 693 if (nanosleep(&ts, &tsr) == -1 && errno == EINTR) { 694 rem = (unsigned int)tsr.tv_sec; 695 if (tsr.tv_nsec >= NANOSEC / 2) 696 rem++; 697 } 698 return (rem); 699 } 700 701 int 702 usleep(useconds_t usec) 703 { 704 timespec_t ts; 705 706 ts.tv_sec = usec / MICROSEC; 707 ts.tv_nsec = (long)(usec % MICROSEC) * 1000; 708 (void) nanosleep(&ts, NULL); 709 return (0); 710 } 711 712 int 713 close(int fildes) 714 { 715 extern void _aio_close(int); 716 extern int __close(int); 717 int rv; 718 719 /* 720 * If we call _aio_close() while in a critical region, 721 * we will draw an ASSERT() failure, so don't do it. 722 * No calls to close() from within libc need _aio_close(); 723 * only the application's calls to close() need this, 724 * and such calls are never from a libc critical region. 725 */ 726 if (curthread->ul_critical == 0) 727 _aio_close(fildes); 728 PERFORM(__close(fildes)) 729 } 730 731 int 732 door_call(int d, door_arg_t *params) 733 { 734 extern int __door_call(int, door_arg_t *); 735 int rv; 736 737 PERFORM(__door_call(d, params)) 738 } 739 740 int 741 fcntl(int fildes, int cmd, ...) 742 { 743 extern int __fcntl(int, int, ...); 744 intptr_t arg; 745 int rv; 746 va_list ap; 747 748 va_start(ap, cmd); 749 arg = va_arg(ap, intptr_t); 750 va_end(ap); 751 if (cmd != F_SETLKW) 752 return (__fcntl(fildes, cmd, arg)); 753 PERFORM(__fcntl(fildes, cmd, arg)) 754 } 755 756 int 757 fdatasync(int fildes) 758 { 759 extern int __fdsync(int, int); 760 int rv; 761 762 PERFORM(__fdsync(fildes, FDSYNC)) 763 } 764 765 int 766 fsync(int fildes) 767 { 768 extern int __fdsync(int, int); 769 int rv; 770 771 PERFORM(__fdsync(fildes, FSYNC)) 772 } 773 774 int 775 lockf(int fildes, int function, off_t size) 776 { 777 extern int __lockf(int, int, off_t); 778 int rv; 779 780 PERFORM(__lockf(fildes, function, size)) 781 } 782 783 #if !defined(_LP64) 784 int 785 lockf64(int fildes, int function, off64_t size) 786 { 787 extern int __lockf64(int, int, off64_t); 788 int rv; 789 790 PERFORM(__lockf64(fildes, function, size)) 791 } 792 #endif /* !_LP64 */ 793 794 ssize_t 795 msgrcv(int msqid, void *msgp, size_t msgsz, long msgtyp, int msgflg) 796 { 797 extern ssize_t __msgrcv(int, void *, size_t, long, int); 798 ssize_t rv; 799 800 PERFORM(__msgrcv(msqid, msgp, msgsz, msgtyp, msgflg)) 801 } 802 803 int 804 msgsnd(int msqid, const void *msgp, size_t msgsz, int msgflg) 805 { 806 extern int __msgsnd(int, const void *, size_t, int); 807 int rv; 808 809 PERFORM(__msgsnd(msqid, msgp, msgsz, msgflg)) 810 } 811 812 int 813 msync(void *addr, size_t len, int flags) 814 { 815 extern int __msync(caddr_t, size_t, int); 816 int rv; 817 818 PERFORM(__msync(addr, len, flags)) 819 } 820 821 int 822 openat(int fd, const char *path, int oflag, ...) 823 { 824 mode_t mode; 825 int rv; 826 va_list ap; 827 828 va_start(ap, oflag); 829 mode = va_arg(ap, mode_t); 830 va_end(ap); 831 PERFORM(__openat(fd, path, oflag, mode)) 832 } 833 834 int 835 open(const char *path, int oflag, ...) 836 { 837 mode_t mode; 838 int rv; 839 va_list ap; 840 841 va_start(ap, oflag); 842 mode = va_arg(ap, mode_t); 843 va_end(ap); 844 PERFORM(__open(path, oflag, mode)) 845 } 846 847 int 848 creat(const char *path, mode_t mode) 849 { 850 return (open(path, O_WRONLY | O_CREAT | O_TRUNC, mode)); 851 } 852 853 #if !defined(_LP64) 854 int 855 openat64(int fd, const char *path, int oflag, ...) 856 { 857 mode_t mode; 858 int rv; 859 va_list ap; 860 861 va_start(ap, oflag); 862 mode = va_arg(ap, mode_t); 863 va_end(ap); 864 PERFORM(__openat64(fd, path, oflag, mode)) 865 } 866 867 int 868 open64(const char *path, int oflag, ...) 869 { 870 mode_t mode; 871 int rv; 872 va_list ap; 873 874 va_start(ap, oflag); 875 mode = va_arg(ap, mode_t); 876 va_end(ap); 877 PERFORM(__open64(path, oflag, mode)) 878 } 879 880 int 881 creat64(const char *path, mode_t mode) 882 { 883 return (open64(path, O_WRONLY | O_CREAT | O_TRUNC, mode)); 884 } 885 #endif /* !_LP64 */ 886 887 int 888 pause(void) 889 { 890 extern int __pause(void); 891 int rv; 892 893 PERFORM(__pause()) 894 } 895 896 ssize_t 897 pread(int fildes, void *buf, size_t nbyte, off_t offset) 898 { 899 extern ssize_t __pread(int, void *, size_t, off_t); 900 ssize_t rv; 901 902 PERFORM(__pread(fildes, buf, nbyte, offset)) 903 } 904 905 #if !defined(_LP64) 906 ssize_t 907 pread64(int fildes, void *buf, size_t nbyte, off64_t offset) 908 { 909 extern ssize_t __pread64(int, void *, size_t, off64_t); 910 ssize_t rv; 911 912 PERFORM(__pread64(fildes, buf, nbyte, offset)) 913 } 914 915 ssize_t 916 preadv64(int fildes, const struct iovec *iov, int iovcnt, off64_t offset) 917 { 918 919 extern ssize_t __preadv64(int, const struct iovec *, int, off_t, off_t); 920 ssize_t rv; 921 922 PERFORM(__preadv64(fildes, iov, iovcnt, offset & 0xffffffffULL, 923 offset>>32)) 924 } 925 #endif /* !_LP64 */ 926 927 ssize_t 928 preadv(int fildes, const struct iovec *iov, int iovcnt, off_t offset) 929 { 930 931 extern ssize_t __preadv(int, const struct iovec *, int, off_t, off_t); 932 ssize_t rv; 933 934 PERFORM(__preadv(fildes, iov, iovcnt, offset, 0)) 935 } 936 ssize_t 937 pwrite(int fildes, const void *buf, size_t nbyte, off_t offset) 938 { 939 extern ssize_t __pwrite(int, const void *, size_t, off_t); 940 ssize_t rv; 941 942 PERFORM(__pwrite(fildes, buf, nbyte, offset)) 943 } 944 945 #if !defined(_LP64) 946 ssize_t 947 pwrite64(int fildes, const void *buf, size_t nbyte, off64_t offset) 948 { 949 extern ssize_t __pwrite64(int, const void *, size_t, off64_t); 950 ssize_t rv; 951 952 PERFORM(__pwrite64(fildes, buf, nbyte, offset)) 953 } 954 955 ssize_t 956 pwritev64(int fildes, const struct iovec *iov, int iovcnt, off64_t offset) 957 { 958 959 extern ssize_t __pwritev64(int, 960 const struct iovec *, int, off_t, off_t); 961 ssize_t rv; 962 963 PERFORM(__pwritev64(fildes, iov, iovcnt, offset & 964 0xffffffffULL, offset>>32)) 965 } 966 967 #endif /* !_LP64 */ 968 969 ssize_t 970 pwritev(int fildes, const struct iovec *iov, int iovcnt, off_t offset) 971 { 972 extern ssize_t __pwritev(int, const struct iovec *, int, off_t, off_t); 973 ssize_t rv; 974 975 PERFORM(__pwritev(fildes, iov, iovcnt, offset, 0)) 976 } 977 978 ssize_t 979 readv(int fildes, const struct iovec *iov, int iovcnt) 980 { 981 extern ssize_t __readv(int, const struct iovec *, int); 982 ssize_t rv; 983 984 PERFORM(__readv(fildes, iov, iovcnt)) 985 } 986 987 int 988 sigpause(int sig) 989 { 990 extern int __sigpause(int); 991 int rv; 992 993 PERFORM(__sigpause(sig)) 994 } 995 996 int 997 sigsuspend(const sigset_t *set) 998 { 999 extern int __sigsuspend(const sigset_t *); 1000 int rv; 1001 1002 PROLOGUE_MASK(set) 1003 rv = __sigsuspend(set); 1004 EPILOGUE_MASK 1005 return (rv); 1006 } 1007 1008 int 1009 _pollsys(struct pollfd *fds, nfds_t nfd, const timespec_t *timeout, 1010 const sigset_t *sigmask) 1011 { 1012 extern int __pollsys(struct pollfd *, nfds_t, const timespec_t *, 1013 const sigset_t *); 1014 int rv; 1015 1016 PROLOGUE_MASK(sigmask) 1017 rv = __pollsys(fds, nfd, timeout, sigmask); 1018 EPILOGUE_MASK 1019 return (rv); 1020 } 1021 1022 int 1023 sigtimedwait(const sigset_t *set, siginfo_t *infop, const timespec_t *timeout) 1024 { 1025 extern int __sigtimedwait(const sigset_t *, siginfo_t *, 1026 const timespec_t *); 1027 siginfo_t info; 1028 int sig; 1029 1030 PROLOGUE 1031 if (abort) { 1032 *self->ul_errnop = EINTR; 1033 sig = -1; 1034 } else { 1035 sig = __sigtimedwait(set, &info, timeout); 1036 if (sig == SIGCANCEL && 1037 (SI_FROMKERNEL(&info) || info.si_code == SI_LWP)) { 1038 do_sigcancel(); 1039 *self->ul_errnop = EINTR; 1040 sig = -1; 1041 } 1042 } 1043 EPILOGUE 1044 if (sig != -1 && infop) 1045 (void) memcpy(infop, &info, sizeof (*infop)); 1046 return (sig); 1047 } 1048 1049 int 1050 sigwait(sigset_t *set) 1051 { 1052 return (sigtimedwait(set, NULL, NULL)); 1053 } 1054 1055 int 1056 sigwaitinfo(const sigset_t *set, siginfo_t *info) 1057 { 1058 return (sigtimedwait(set, info, NULL)); 1059 } 1060 1061 int 1062 sigqueue(pid_t pid, int signo, const union sigval value) 1063 { 1064 extern int __sigqueue(pid_t pid, int signo, 1065 /* const union sigval */ void *value, int si_code, int block); 1066 return (__sigqueue(pid, signo, value.sival_ptr, SI_QUEUE, 0)); 1067 } 1068 1069 int 1070 _so_accept(int sock, struct sockaddr *addr, uint_t *addrlen, int version, 1071 int flags) 1072 { 1073 extern int __so_accept(int, struct sockaddr *, uint_t *, int, int); 1074 int rv; 1075 1076 PERFORM(__so_accept(sock, addr, addrlen, version, flags)) 1077 } 1078 1079 int 1080 _so_connect(int sock, struct sockaddr *addr, uint_t addrlen, int version) 1081 { 1082 extern int __so_connect(int, struct sockaddr *, uint_t, int); 1083 int rv; 1084 1085 PERFORM(__so_connect(sock, addr, addrlen, version)) 1086 } 1087 1088 int 1089 _so_recv(int sock, void *buf, size_t len, int flags) 1090 { 1091 extern int __so_recv(int, void *, size_t, int); 1092 int rv; 1093 1094 PERFORM(__so_recv(sock, buf, len, flags)) 1095 } 1096 1097 int 1098 _so_recvfrom(int sock, void *buf, size_t len, int flags, 1099 struct sockaddr *addr, int *addrlen) 1100 { 1101 extern int __so_recvfrom(int, void *, size_t, int, 1102 struct sockaddr *, int *); 1103 int rv; 1104 1105 PERFORM(__so_recvfrom(sock, buf, len, flags, addr, addrlen)) 1106 } 1107 1108 int 1109 _so_recvmsg(int sock, struct msghdr *msg, int flags) 1110 { 1111 extern int __so_recvmsg(int, struct msghdr *, int); 1112 int rv; 1113 1114 PERFORM(__so_recvmsg(sock, msg, flags)) 1115 } 1116 1117 int 1118 _so_send(int sock, const void *buf, size_t len, int flags) 1119 { 1120 extern int __so_send(int, const void *, size_t, int); 1121 int rv; 1122 1123 PERFORM(__so_send(sock, buf, len, flags)) 1124 } 1125 1126 int 1127 _so_sendmsg(int sock, const struct msghdr *msg, int flags) 1128 { 1129 extern int __so_sendmsg(int, const struct msghdr *, int); 1130 int rv; 1131 1132 PERFORM(__so_sendmsg(sock, msg, flags)) 1133 } 1134 1135 int 1136 _so_sendto(int sock, const void *buf, size_t len, int flags, 1137 const struct sockaddr *addr, int *addrlen) 1138 { 1139 extern int __so_sendto(int, const void *, size_t, int, 1140 const struct sockaddr *, int *); 1141 int rv; 1142 1143 PERFORM(__so_sendto(sock, buf, len, flags, addr, addrlen)) 1144 } 1145 1146 int 1147 tcdrain(int fildes) 1148 { 1149 extern int __tcdrain(int); 1150 int rv; 1151 1152 PERFORM(__tcdrain(fildes)) 1153 } 1154 1155 int 1156 waitid(idtype_t idtype, id_t id, siginfo_t *infop, int options) 1157 { 1158 extern int __waitid(idtype_t, id_t, siginfo_t *, int); 1159 int rv; 1160 1161 if (options & WNOHANG) 1162 return (__waitid(idtype, id, infop, options)); 1163 PERFORM(__waitid(idtype, id, infop, options)) 1164 } 1165 1166 ssize_t 1167 writev(int fildes, const struct iovec *iov, int iovcnt) 1168 { 1169 extern ssize_t __writev(int, const struct iovec *, int); 1170 ssize_t rv; 1171 1172 PERFORM(__writev(fildes, iov, iovcnt)) 1173 } 1174