1 /* $NetBSD: clnt_vc.c,v 1.4 2000/07/14 08:40:42 fvdl Exp $ */ 2 3 /* 4 * Sun RPC is a product of Sun Microsystems, Inc. and is provided for 5 * unrestricted use provided that this legend is included on all tape 6 * media and as a part of the software program in whole or part. Users 7 * may copy or modify Sun RPC without charge, but are not authorized 8 * to license or distribute it to anyone else except as part of a product or 9 * program developed by the user. 10 * 11 * SUN RPC IS PROVIDED AS IS WITH NO WARRANTIES OF ANY KIND INCLUDING THE 12 * WARRANTIES OF DESIGN, MERCHANTIBILITY AND FITNESS FOR A PARTICULAR 13 * PURPOSE, OR ARISING FROM A COURSE OF DEALING, USAGE OR TRADE PRACTICE. 14 * 15 * Sun RPC is provided with no support and without any obligation on the 16 * part of Sun Microsystems, Inc. to assist in its use, correction, 17 * modification or enhancement. 18 * 19 * SUN MICROSYSTEMS, INC. SHALL HAVE NO LIABILITY WITH RESPECT TO THE 20 * INFRINGEMENT OF COPYRIGHTS, TRADE SECRETS OR ANY PATENTS BY SUN RPC 21 * OR ANY PART THEREOF. 22 * 23 * In no event will Sun Microsystems, Inc. be liable for any lost revenue 24 * or profits or other special, indirect and consequential damages, even if 25 * Sun has been advised of the possibility of such damages. 26 * 27 * Sun Microsystems, Inc. 28 * 2550 Garcia Avenue 29 * Mountain View, California 94043 30 */ 31 32 #if defined(LIBC_SCCS) && !defined(lint) 33 static char *sccsid2 = "@(#)clnt_tcp.c 1.37 87/10/05 Copyr 1984 Sun Micro"; 34 static char *sccsid = "@(#)clnt_tcp.c 2.2 88/08/01 4.0 RPCSRC"; 35 static char sccsid3[] = "@(#)clnt_vc.c 1.19 89/03/16 Copyr 1988 Sun Micro"; 36 #endif 37 #include <sys/cdefs.h> 38 __FBSDID("$FreeBSD$"); 39 40 /* 41 * clnt_tcp.c, Implements a TCP/IP based, client side RPC. 42 * 43 * Copyright (C) 1984, Sun Microsystems, Inc. 44 * 45 * TCP based RPC supports 'batched calls'. 46 * A sequence of calls may be batched-up in a send buffer. The rpc call 47 * return immediately to the client even though the call was not necessarily 48 * sent. The batching occurs if the results' xdr routine is NULL (0) AND 49 * the rpc timeout value is zero (see clnt.h, rpc). 50 * 51 * Clients should NOT casually batch calls that in fact return results; that is, 52 * the server side should be aware that a call is batched and not produce any 53 * return message. Batched calls that produce many result messages can 54 * deadlock (netlock) the client and the server.... 55 * 56 * Now go hang yourself. 57 */ 58 59 #include <sys/param.h> 60 #include <sys/systm.h> 61 #include <sys/lock.h> 62 #include <sys/malloc.h> 63 #include <sys/mbuf.h> 64 #include <sys/mutex.h> 65 #include <sys/pcpu.h> 66 #include <sys/proc.h> 67 #include <sys/protosw.h> 68 #include <sys/socket.h> 69 #include <sys/socketvar.h> 70 #include <sys/syslog.h> 71 #include <sys/time.h> 72 #include <sys/uio.h> 73 74 #include <net/vnet.h> 75 76 #include <netinet/tcp.h> 77 78 #include <rpc/rpc.h> 79 #include <rpc/rpc_com.h> 80 81 #define MCALL_MSG_SIZE 24 82 83 struct cmessage { 84 struct cmsghdr cmsg; 85 struct cmsgcred cmcred; 86 }; 87 88 static enum clnt_stat clnt_vc_call(CLIENT *, struct rpc_callextra *, 89 rpcproc_t, struct mbuf *, struct mbuf **, struct timeval); 90 static void clnt_vc_geterr(CLIENT *, struct rpc_err *); 91 static bool_t clnt_vc_freeres(CLIENT *, xdrproc_t, void *); 92 static void clnt_vc_abort(CLIENT *); 93 static bool_t clnt_vc_control(CLIENT *, u_int, void *); 94 static void clnt_vc_close(CLIENT *); 95 static void clnt_vc_destroy(CLIENT *); 96 static bool_t time_not_ok(struct timeval *); 97 static int clnt_vc_soupcall(struct socket *so, void *arg, int waitflag); 98 99 static struct clnt_ops clnt_vc_ops = { 100 .cl_call = clnt_vc_call, 101 .cl_abort = clnt_vc_abort, 102 .cl_geterr = clnt_vc_geterr, 103 .cl_freeres = clnt_vc_freeres, 104 .cl_close = clnt_vc_close, 105 .cl_destroy = clnt_vc_destroy, 106 .cl_control = clnt_vc_control 107 }; 108 109 /* 110 * A pending RPC request which awaits a reply. Requests which have 111 * received their reply will have cr_xid set to zero and cr_mrep to 112 * the mbuf chain of the reply. 113 */ 114 struct ct_request { 115 TAILQ_ENTRY(ct_request) cr_link; 116 uint32_t cr_xid; /* XID of request */ 117 struct mbuf *cr_mrep; /* reply received by upcall */ 118 int cr_error; /* any error from upcall */ 119 char cr_verf[MAX_AUTH_BYTES]; /* reply verf */ 120 }; 121 122 TAILQ_HEAD(ct_request_list, ct_request); 123 124 struct ct_data { 125 struct mtx ct_lock; 126 int ct_threads; /* number of threads in clnt_vc_call */ 127 bool_t ct_closing; /* TRUE if we are closing */ 128 bool_t ct_closed; /* TRUE if we are closed */ 129 struct socket *ct_socket; /* connection socket */ 130 bool_t ct_closeit; /* close it on destroy */ 131 struct timeval ct_wait; /* wait interval in milliseconds */ 132 struct sockaddr_storage ct_addr; /* remote addr */ 133 struct rpc_err ct_error; 134 uint32_t ct_xid; 135 char ct_mcallc[MCALL_MSG_SIZE]; /* marshalled callmsg */ 136 size_t ct_mpos; /* pos after marshal */ 137 const char *ct_waitchan; 138 int ct_waitflag; 139 struct mbuf *ct_record; /* current reply record */ 140 size_t ct_record_resid; /* how much left of reply to read */ 141 bool_t ct_record_eor; /* true if reading last fragment */ 142 struct ct_request_list ct_pending; 143 int ct_upcallrefs; /* Ref cnt of upcalls in prog. */ 144 }; 145 146 static void clnt_vc_upcallsdone(struct ct_data *); 147 148 static const char clnt_vc_errstr[] = "%s : %s"; 149 static const char clnt_vc_str[] = "clnt_vc_create"; 150 static const char clnt_read_vc_str[] = "read_vc"; 151 static const char __no_mem_str[] = "out of memory"; 152 153 /* 154 * Create a client handle for a connection. 155 * Default options are set, which the user can change using clnt_control()'s. 156 * The rpc/vc package does buffering similar to stdio, so the client 157 * must pick send and receive buffer sizes, 0 => use the default. 158 * NB: fd is copied into a private area. 159 * NB: The rpch->cl_auth is set null authentication. Caller may wish to 160 * set this something more useful. 161 * 162 * fd should be an open socket 163 */ 164 CLIENT * 165 clnt_vc_create( 166 struct socket *so, /* open file descriptor */ 167 struct sockaddr *raddr, /* servers address */ 168 const rpcprog_t prog, /* program number */ 169 const rpcvers_t vers, /* version number */ 170 size_t sendsz, /* buffer recv size */ 171 size_t recvsz) /* buffer send size */ 172 { 173 CLIENT *cl; /* client handle */ 174 struct ct_data *ct = NULL; /* client handle */ 175 struct timeval now; 176 struct rpc_msg call_msg; 177 static uint32_t disrupt; 178 struct __rpc_sockinfo si; 179 XDR xdrs; 180 int error, interrupted, one = 1; 181 struct sockopt sopt; 182 183 if (disrupt == 0) 184 disrupt = (uint32_t)(long)raddr; 185 186 cl = (CLIENT *)mem_alloc(sizeof (*cl)); 187 ct = (struct ct_data *)mem_alloc(sizeof (*ct)); 188 189 mtx_init(&ct->ct_lock, "ct->ct_lock", NULL, MTX_DEF); 190 ct->ct_threads = 0; 191 ct->ct_closing = FALSE; 192 ct->ct_closed = FALSE; 193 ct->ct_upcallrefs = 0; 194 195 if ((so->so_state & (SS_ISCONNECTED|SS_ISCONFIRMING)) == 0) { 196 error = soconnect(so, raddr, curthread); 197 SOCK_LOCK(so); 198 interrupted = 0; 199 while ((so->so_state & SS_ISCONNECTING) 200 && so->so_error == 0) { 201 error = msleep(&so->so_timeo, SOCK_MTX(so), 202 PSOCK | PCATCH | PBDRY, "connec", 0); 203 if (error) { 204 if (error == EINTR || error == ERESTART) 205 interrupted = 1; 206 break; 207 } 208 } 209 if (error == 0) { 210 error = so->so_error; 211 so->so_error = 0; 212 } 213 SOCK_UNLOCK(so); 214 if (error) { 215 if (!interrupted) 216 so->so_state &= ~SS_ISCONNECTING; 217 rpc_createerr.cf_stat = RPC_SYSTEMERROR; 218 rpc_createerr.cf_error.re_errno = error; 219 goto err; 220 } 221 } 222 223 CURVNET_SET(so->so_vnet); 224 if (!__rpc_socket2sockinfo(so, &si)) { 225 CURVNET_RESTORE(); 226 goto err; 227 } 228 229 if (so->so_proto->pr_flags & PR_CONNREQUIRED) { 230 bzero(&sopt, sizeof(sopt)); 231 sopt.sopt_dir = SOPT_SET; 232 sopt.sopt_level = SOL_SOCKET; 233 sopt.sopt_name = SO_KEEPALIVE; 234 sopt.sopt_val = &one; 235 sopt.sopt_valsize = sizeof(one); 236 sosetopt(so, &sopt); 237 } 238 239 if (so->so_proto->pr_protocol == IPPROTO_TCP) { 240 bzero(&sopt, sizeof(sopt)); 241 sopt.sopt_dir = SOPT_SET; 242 sopt.sopt_level = IPPROTO_TCP; 243 sopt.sopt_name = TCP_NODELAY; 244 sopt.sopt_val = &one; 245 sopt.sopt_valsize = sizeof(one); 246 sosetopt(so, &sopt); 247 } 248 CURVNET_RESTORE(); 249 250 ct->ct_closeit = FALSE; 251 252 /* 253 * Set up private data struct 254 */ 255 ct->ct_socket = so; 256 ct->ct_wait.tv_sec = -1; 257 ct->ct_wait.tv_usec = -1; 258 memcpy(&ct->ct_addr, raddr, raddr->sa_len); 259 260 /* 261 * Initialize call message 262 */ 263 getmicrotime(&now); 264 ct->ct_xid = ((uint32_t)++disrupt) ^ __RPC_GETXID(&now); 265 call_msg.rm_xid = ct->ct_xid; 266 call_msg.rm_direction = CALL; 267 call_msg.rm_call.cb_rpcvers = RPC_MSG_VERSION; 268 call_msg.rm_call.cb_prog = (uint32_t)prog; 269 call_msg.rm_call.cb_vers = (uint32_t)vers; 270 271 /* 272 * pre-serialize the static part of the call msg and stash it away 273 */ 274 xdrmem_create(&xdrs, ct->ct_mcallc, MCALL_MSG_SIZE, 275 XDR_ENCODE); 276 if (! xdr_callhdr(&xdrs, &call_msg)) { 277 if (ct->ct_closeit) { 278 soclose(ct->ct_socket); 279 } 280 goto err; 281 } 282 ct->ct_mpos = XDR_GETPOS(&xdrs); 283 XDR_DESTROY(&xdrs); 284 ct->ct_waitchan = "rpcrecv"; 285 ct->ct_waitflag = 0; 286 287 /* 288 * Create a client handle which uses xdrrec for serialization 289 * and authnone for authentication. 290 */ 291 sendsz = __rpc_get_t_size(si.si_af, si.si_proto, (int)sendsz); 292 recvsz = __rpc_get_t_size(si.si_af, si.si_proto, (int)recvsz); 293 error = soreserve(ct->ct_socket, sendsz, recvsz); 294 if (error != 0) { 295 if (ct->ct_closeit) { 296 soclose(ct->ct_socket); 297 } 298 goto err; 299 } 300 cl->cl_refs = 1; 301 cl->cl_ops = &clnt_vc_ops; 302 cl->cl_private = ct; 303 cl->cl_auth = authnone_create(); 304 305 SOCKBUF_LOCK(&ct->ct_socket->so_rcv); 306 soupcall_set(ct->ct_socket, SO_RCV, clnt_vc_soupcall, ct); 307 SOCKBUF_UNLOCK(&ct->ct_socket->so_rcv); 308 309 ct->ct_record = NULL; 310 ct->ct_record_resid = 0; 311 TAILQ_INIT(&ct->ct_pending); 312 return (cl); 313 314 err: 315 if (cl) { 316 if (ct) { 317 mtx_destroy(&ct->ct_lock); 318 mem_free(ct, sizeof (struct ct_data)); 319 } 320 if (cl) 321 mem_free(cl, sizeof (CLIENT)); 322 } 323 return ((CLIENT *)NULL); 324 } 325 326 static enum clnt_stat 327 clnt_vc_call( 328 CLIENT *cl, /* client handle */ 329 struct rpc_callextra *ext, /* call metadata */ 330 rpcproc_t proc, /* procedure number */ 331 struct mbuf *args, /* pointer to args */ 332 struct mbuf **resultsp, /* pointer to results */ 333 struct timeval utimeout) 334 { 335 struct ct_data *ct = (struct ct_data *) cl->cl_private; 336 AUTH *auth; 337 struct rpc_err *errp; 338 enum clnt_stat stat; 339 XDR xdrs; 340 struct rpc_msg reply_msg; 341 bool_t ok; 342 int nrefreshes = 2; /* number of times to refresh cred */ 343 struct timeval timeout; 344 uint32_t xid; 345 struct mbuf *mreq = NULL, *results; 346 struct ct_request *cr; 347 int error; 348 349 cr = malloc(sizeof(struct ct_request), M_RPC, M_WAITOK); 350 351 mtx_lock(&ct->ct_lock); 352 353 if (ct->ct_closing || ct->ct_closed) { 354 mtx_unlock(&ct->ct_lock); 355 free(cr, M_RPC); 356 return (RPC_CANTSEND); 357 } 358 ct->ct_threads++; 359 360 if (ext) { 361 auth = ext->rc_auth; 362 errp = &ext->rc_err; 363 } else { 364 auth = cl->cl_auth; 365 errp = &ct->ct_error; 366 } 367 368 cr->cr_mrep = NULL; 369 cr->cr_error = 0; 370 371 if (ct->ct_wait.tv_usec == -1) { 372 timeout = utimeout; /* use supplied timeout */ 373 } else { 374 timeout = ct->ct_wait; /* use default timeout */ 375 } 376 377 call_again: 378 mtx_assert(&ct->ct_lock, MA_OWNED); 379 380 ct->ct_xid++; 381 xid = ct->ct_xid; 382 383 mtx_unlock(&ct->ct_lock); 384 385 /* 386 * Leave space to pre-pend the record mark. 387 */ 388 MGETHDR(mreq, M_WAIT, MT_DATA); 389 mreq->m_data += sizeof(uint32_t); 390 KASSERT(ct->ct_mpos + sizeof(uint32_t) <= MHLEN, 391 ("RPC header too big")); 392 bcopy(ct->ct_mcallc, mreq->m_data, ct->ct_mpos); 393 mreq->m_len = ct->ct_mpos; 394 395 /* 396 * The XID is the first thing in the request. 397 */ 398 *mtod(mreq, uint32_t *) = htonl(xid); 399 400 xdrmbuf_create(&xdrs, mreq, XDR_ENCODE); 401 402 errp->re_status = stat = RPC_SUCCESS; 403 404 if ((! XDR_PUTINT32(&xdrs, &proc)) || 405 (! AUTH_MARSHALL(auth, xid, &xdrs, 406 m_copym(args, 0, M_COPYALL, M_WAITOK)))) { 407 errp->re_status = stat = RPC_CANTENCODEARGS; 408 mtx_lock(&ct->ct_lock); 409 goto out; 410 } 411 mreq->m_pkthdr.len = m_length(mreq, NULL); 412 413 /* 414 * Prepend a record marker containing the packet length. 415 */ 416 M_PREPEND(mreq, sizeof(uint32_t), M_WAIT); 417 *mtod(mreq, uint32_t *) = 418 htonl(0x80000000 | (mreq->m_pkthdr.len - sizeof(uint32_t))); 419 420 cr->cr_xid = xid; 421 mtx_lock(&ct->ct_lock); 422 /* 423 * Check to see if the other end has already started to close down 424 * the connection. The upcall will have set ct_error.re_status 425 * to RPC_CANTRECV if this is the case. 426 * If the other end starts to close down the connection after this 427 * point, it will be detected later when cr_error is checked, 428 * since the request is in the ct_pending queue. 429 */ 430 if (ct->ct_error.re_status == RPC_CANTRECV) { 431 if (errp != &ct->ct_error) { 432 errp->re_errno = ct->ct_error.re_errno; 433 errp->re_status = RPC_CANTRECV; 434 } 435 stat = RPC_CANTRECV; 436 goto out; 437 } 438 TAILQ_INSERT_TAIL(&ct->ct_pending, cr, cr_link); 439 mtx_unlock(&ct->ct_lock); 440 441 /* 442 * sosend consumes mreq. 443 */ 444 error = sosend(ct->ct_socket, NULL, NULL, mreq, NULL, 0, curthread); 445 mreq = NULL; 446 if (error == EMSGSIZE) { 447 SOCKBUF_LOCK(&ct->ct_socket->so_snd); 448 sbwait(&ct->ct_socket->so_snd); 449 SOCKBUF_UNLOCK(&ct->ct_socket->so_snd); 450 AUTH_VALIDATE(auth, xid, NULL, NULL); 451 mtx_lock(&ct->ct_lock); 452 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link); 453 goto call_again; 454 } 455 456 reply_msg.acpted_rply.ar_verf.oa_flavor = AUTH_NULL; 457 reply_msg.acpted_rply.ar_verf.oa_base = cr->cr_verf; 458 reply_msg.acpted_rply.ar_verf.oa_length = 0; 459 reply_msg.acpted_rply.ar_results.where = NULL; 460 reply_msg.acpted_rply.ar_results.proc = (xdrproc_t)xdr_void; 461 462 mtx_lock(&ct->ct_lock); 463 if (error) { 464 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link); 465 errp->re_errno = error; 466 errp->re_status = stat = RPC_CANTSEND; 467 goto out; 468 } 469 470 /* 471 * Check to see if we got an upcall while waiting for the 472 * lock. In both these cases, the request has been removed 473 * from ct->ct_pending. 474 */ 475 if (cr->cr_error) { 476 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link); 477 errp->re_errno = cr->cr_error; 478 errp->re_status = stat = RPC_CANTRECV; 479 goto out; 480 } 481 if (cr->cr_mrep) { 482 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link); 483 goto got_reply; 484 } 485 486 /* 487 * Hack to provide rpc-based message passing 488 */ 489 if (timeout.tv_sec == 0 && timeout.tv_usec == 0) { 490 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link); 491 errp->re_status = stat = RPC_TIMEDOUT; 492 goto out; 493 } 494 495 error = msleep(cr, &ct->ct_lock, ct->ct_waitflag, ct->ct_waitchan, 496 tvtohz(&timeout)); 497 498 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link); 499 500 if (error) { 501 /* 502 * The sleep returned an error so our request is still 503 * on the list. Turn the error code into an 504 * appropriate client status. 505 */ 506 errp->re_errno = error; 507 switch (error) { 508 case EINTR: 509 case ERESTART: 510 stat = RPC_INTR; 511 break; 512 case EWOULDBLOCK: 513 stat = RPC_TIMEDOUT; 514 break; 515 default: 516 stat = RPC_CANTRECV; 517 } 518 errp->re_status = stat; 519 goto out; 520 } else { 521 /* 522 * We were woken up by the upcall. If the 523 * upcall had a receive error, report that, 524 * otherwise we have a reply. 525 */ 526 if (cr->cr_error) { 527 errp->re_errno = cr->cr_error; 528 errp->re_status = stat = RPC_CANTRECV; 529 goto out; 530 } 531 } 532 533 got_reply: 534 /* 535 * Now decode and validate the response. We need to drop the 536 * lock since xdr_replymsg may end up sleeping in malloc. 537 */ 538 mtx_unlock(&ct->ct_lock); 539 540 if (ext && ext->rc_feedback) 541 ext->rc_feedback(FEEDBACK_OK, proc, ext->rc_feedback_arg); 542 543 xdrmbuf_create(&xdrs, cr->cr_mrep, XDR_DECODE); 544 ok = xdr_replymsg(&xdrs, &reply_msg); 545 cr->cr_mrep = NULL; 546 547 if (ok) { 548 if ((reply_msg.rm_reply.rp_stat == MSG_ACCEPTED) && 549 (reply_msg.acpted_rply.ar_stat == SUCCESS)) 550 errp->re_status = stat = RPC_SUCCESS; 551 else 552 stat = _seterr_reply(&reply_msg, errp); 553 554 if (stat == RPC_SUCCESS) { 555 results = xdrmbuf_getall(&xdrs); 556 if (!AUTH_VALIDATE(auth, xid, 557 &reply_msg.acpted_rply.ar_verf, 558 &results)) { 559 errp->re_status = stat = RPC_AUTHERROR; 560 errp->re_why = AUTH_INVALIDRESP; 561 } else { 562 KASSERT(results, 563 ("auth validated but no result")); 564 *resultsp = results; 565 } 566 } /* end successful completion */ 567 /* 568 * If unsuccesful AND error is an authentication error 569 * then refresh credentials and try again, else break 570 */ 571 else if (stat == RPC_AUTHERROR) 572 /* maybe our credentials need to be refreshed ... */ 573 if (nrefreshes > 0 && 574 AUTH_REFRESH(auth, &reply_msg)) { 575 nrefreshes--; 576 XDR_DESTROY(&xdrs); 577 mtx_lock(&ct->ct_lock); 578 goto call_again; 579 } 580 /* end of unsuccessful completion */ 581 } /* end of valid reply message */ 582 else { 583 errp->re_status = stat = RPC_CANTDECODERES; 584 } 585 XDR_DESTROY(&xdrs); 586 mtx_lock(&ct->ct_lock); 587 out: 588 mtx_assert(&ct->ct_lock, MA_OWNED); 589 590 KASSERT(stat != RPC_SUCCESS || *resultsp, 591 ("RPC_SUCCESS without reply")); 592 593 if (mreq) 594 m_freem(mreq); 595 if (cr->cr_mrep) 596 m_freem(cr->cr_mrep); 597 598 ct->ct_threads--; 599 if (ct->ct_closing) 600 wakeup(ct); 601 602 mtx_unlock(&ct->ct_lock); 603 604 if (auth && stat != RPC_SUCCESS) 605 AUTH_VALIDATE(auth, xid, NULL, NULL); 606 607 free(cr, M_RPC); 608 609 return (stat); 610 } 611 612 static void 613 clnt_vc_geterr(CLIENT *cl, struct rpc_err *errp) 614 { 615 struct ct_data *ct = (struct ct_data *) cl->cl_private; 616 617 *errp = ct->ct_error; 618 } 619 620 static bool_t 621 clnt_vc_freeres(CLIENT *cl, xdrproc_t xdr_res, void *res_ptr) 622 { 623 XDR xdrs; 624 bool_t dummy; 625 626 xdrs.x_op = XDR_FREE; 627 dummy = (*xdr_res)(&xdrs, res_ptr); 628 629 return (dummy); 630 } 631 632 /*ARGSUSED*/ 633 static void 634 clnt_vc_abort(CLIENT *cl) 635 { 636 } 637 638 static bool_t 639 clnt_vc_control(CLIENT *cl, u_int request, void *info) 640 { 641 struct ct_data *ct = (struct ct_data *)cl->cl_private; 642 void *infop = info; 643 644 mtx_lock(&ct->ct_lock); 645 646 switch (request) { 647 case CLSET_FD_CLOSE: 648 ct->ct_closeit = TRUE; 649 mtx_unlock(&ct->ct_lock); 650 return (TRUE); 651 case CLSET_FD_NCLOSE: 652 ct->ct_closeit = FALSE; 653 mtx_unlock(&ct->ct_lock); 654 return (TRUE); 655 default: 656 break; 657 } 658 659 /* for other requests which use info */ 660 if (info == NULL) { 661 mtx_unlock(&ct->ct_lock); 662 return (FALSE); 663 } 664 switch (request) { 665 case CLSET_TIMEOUT: 666 if (time_not_ok((struct timeval *)info)) { 667 mtx_unlock(&ct->ct_lock); 668 return (FALSE); 669 } 670 ct->ct_wait = *(struct timeval *)infop; 671 break; 672 case CLGET_TIMEOUT: 673 *(struct timeval *)infop = ct->ct_wait; 674 break; 675 case CLGET_SERVER_ADDR: 676 (void) memcpy(info, &ct->ct_addr, (size_t)ct->ct_addr.ss_len); 677 break; 678 case CLGET_SVC_ADDR: 679 /* 680 * Slightly different semantics to userland - we use 681 * sockaddr instead of netbuf. 682 */ 683 memcpy(info, &ct->ct_addr, ct->ct_addr.ss_len); 684 break; 685 case CLSET_SVC_ADDR: /* set to new address */ 686 mtx_unlock(&ct->ct_lock); 687 return (FALSE); 688 case CLGET_XID: 689 *(uint32_t *)info = ct->ct_xid; 690 break; 691 case CLSET_XID: 692 /* This will set the xid of the NEXT call */ 693 /* decrement by 1 as clnt_vc_call() increments once */ 694 ct->ct_xid = *(uint32_t *)info - 1; 695 break; 696 case CLGET_VERS: 697 /* 698 * This RELIES on the information that, in the call body, 699 * the version number field is the fifth field from the 700 * begining of the RPC header. MUST be changed if the 701 * call_struct is changed 702 */ 703 *(uint32_t *)info = 704 ntohl(*(uint32_t *)(void *)(ct->ct_mcallc + 705 4 * BYTES_PER_XDR_UNIT)); 706 break; 707 708 case CLSET_VERS: 709 *(uint32_t *)(void *)(ct->ct_mcallc + 710 4 * BYTES_PER_XDR_UNIT) = 711 htonl(*(uint32_t *)info); 712 break; 713 714 case CLGET_PROG: 715 /* 716 * This RELIES on the information that, in the call body, 717 * the program number field is the fourth field from the 718 * begining of the RPC header. MUST be changed if the 719 * call_struct is changed 720 */ 721 *(uint32_t *)info = 722 ntohl(*(uint32_t *)(void *)(ct->ct_mcallc + 723 3 * BYTES_PER_XDR_UNIT)); 724 break; 725 726 case CLSET_PROG: 727 *(uint32_t *)(void *)(ct->ct_mcallc + 728 3 * BYTES_PER_XDR_UNIT) = 729 htonl(*(uint32_t *)info); 730 break; 731 732 case CLSET_WAITCHAN: 733 ct->ct_waitchan = (const char *)info; 734 break; 735 736 case CLGET_WAITCHAN: 737 *(const char **) info = ct->ct_waitchan; 738 break; 739 740 case CLSET_INTERRUPTIBLE: 741 if (*(int *) info) 742 ct->ct_waitflag = PCATCH | PBDRY; 743 else 744 ct->ct_waitflag = 0; 745 break; 746 747 case CLGET_INTERRUPTIBLE: 748 if (ct->ct_waitflag) 749 *(int *) info = TRUE; 750 else 751 *(int *) info = FALSE; 752 break; 753 754 default: 755 mtx_unlock(&ct->ct_lock); 756 return (FALSE); 757 } 758 759 mtx_unlock(&ct->ct_lock); 760 return (TRUE); 761 } 762 763 static void 764 clnt_vc_close(CLIENT *cl) 765 { 766 struct ct_data *ct = (struct ct_data *) cl->cl_private; 767 struct ct_request *cr; 768 769 mtx_lock(&ct->ct_lock); 770 771 if (ct->ct_closed) { 772 mtx_unlock(&ct->ct_lock); 773 return; 774 } 775 776 if (ct->ct_closing) { 777 while (ct->ct_closing) 778 msleep(ct, &ct->ct_lock, 0, "rpcclose", 0); 779 KASSERT(ct->ct_closed, ("client should be closed")); 780 mtx_unlock(&ct->ct_lock); 781 return; 782 } 783 784 if (ct->ct_socket) { 785 ct->ct_closing = TRUE; 786 mtx_unlock(&ct->ct_lock); 787 788 SOCKBUF_LOCK(&ct->ct_socket->so_rcv); 789 soupcall_clear(ct->ct_socket, SO_RCV); 790 clnt_vc_upcallsdone(ct); 791 SOCKBUF_UNLOCK(&ct->ct_socket->so_rcv); 792 793 /* 794 * Abort any pending requests and wait until everyone 795 * has finished with clnt_vc_call. 796 */ 797 mtx_lock(&ct->ct_lock); 798 TAILQ_FOREACH(cr, &ct->ct_pending, cr_link) { 799 cr->cr_xid = 0; 800 cr->cr_error = ESHUTDOWN; 801 wakeup(cr); 802 } 803 804 while (ct->ct_threads) 805 msleep(ct, &ct->ct_lock, 0, "rpcclose", 0); 806 } 807 808 ct->ct_closing = FALSE; 809 ct->ct_closed = TRUE; 810 mtx_unlock(&ct->ct_lock); 811 wakeup(ct); 812 } 813 814 static void 815 clnt_vc_destroy(CLIENT *cl) 816 { 817 struct ct_data *ct = (struct ct_data *) cl->cl_private; 818 struct socket *so = NULL; 819 820 clnt_vc_close(cl); 821 822 mtx_lock(&ct->ct_lock); 823 824 if (ct->ct_socket) { 825 if (ct->ct_closeit) { 826 so = ct->ct_socket; 827 } 828 } 829 830 mtx_unlock(&ct->ct_lock); 831 832 mtx_destroy(&ct->ct_lock); 833 if (so) { 834 soshutdown(so, SHUT_WR); 835 soclose(so); 836 } 837 mem_free(ct, sizeof(struct ct_data)); 838 mem_free(cl, sizeof(CLIENT)); 839 } 840 841 /* 842 * Make sure that the time is not garbage. -1 value is disallowed. 843 * Note this is different from time_not_ok in clnt_dg.c 844 */ 845 static bool_t 846 time_not_ok(struct timeval *t) 847 { 848 return (t->tv_sec <= -1 || t->tv_sec > 100000000 || 849 t->tv_usec <= -1 || t->tv_usec > 1000000); 850 } 851 852 int 853 clnt_vc_soupcall(struct socket *so, void *arg, int waitflag) 854 { 855 struct ct_data *ct = (struct ct_data *) arg; 856 struct uio uio; 857 struct mbuf *m; 858 struct ct_request *cr; 859 int error, rcvflag, foundreq; 860 uint32_t xid, header; 861 bool_t do_read; 862 863 ct->ct_upcallrefs++; 864 uio.uio_td = curthread; 865 do { 866 /* 867 * If ct_record_resid is zero, we are waiting for a 868 * record mark. 869 */ 870 if (ct->ct_record_resid == 0) { 871 872 /* 873 * Make sure there is either a whole record 874 * mark in the buffer or there is some other 875 * error condition 876 */ 877 do_read = FALSE; 878 if (so->so_rcv.sb_cc >= sizeof(uint32_t) 879 || (so->so_rcv.sb_state & SBS_CANTRCVMORE) 880 || so->so_error) 881 do_read = TRUE; 882 883 if (!do_read) 884 break; 885 886 SOCKBUF_UNLOCK(&so->so_rcv); 887 uio.uio_resid = sizeof(uint32_t); 888 m = NULL; 889 rcvflag = MSG_DONTWAIT | MSG_SOCALLBCK; 890 error = soreceive(so, NULL, &uio, &m, NULL, &rcvflag); 891 SOCKBUF_LOCK(&so->so_rcv); 892 893 if (error == EWOULDBLOCK) 894 break; 895 896 /* 897 * If there was an error, wake up all pending 898 * requests. 899 */ 900 if (error || uio.uio_resid > 0) { 901 wakeup_all: 902 mtx_lock(&ct->ct_lock); 903 if (!error) { 904 /* 905 * We must have got EOF trying 906 * to read from the stream. 907 */ 908 error = ECONNRESET; 909 } 910 ct->ct_error.re_status = RPC_CANTRECV; 911 ct->ct_error.re_errno = error; 912 TAILQ_FOREACH(cr, &ct->ct_pending, cr_link) { 913 cr->cr_error = error; 914 wakeup(cr); 915 } 916 mtx_unlock(&ct->ct_lock); 917 break; 918 } 919 bcopy(mtod(m, uint32_t *), &header, sizeof(uint32_t)); 920 header = ntohl(header); 921 ct->ct_record = NULL; 922 ct->ct_record_resid = header & 0x7fffffff; 923 ct->ct_record_eor = ((header & 0x80000000) != 0); 924 m_freem(m); 925 } else { 926 /* 927 * Wait until the socket has the whole record 928 * buffered. 929 */ 930 do_read = FALSE; 931 if (so->so_rcv.sb_cc >= ct->ct_record_resid 932 || (so->so_rcv.sb_state & SBS_CANTRCVMORE) 933 || so->so_error) 934 do_read = TRUE; 935 936 if (!do_read) 937 break; 938 939 /* 940 * We have the record mark. Read as much as 941 * the socket has buffered up to the end of 942 * this record. 943 */ 944 SOCKBUF_UNLOCK(&so->so_rcv); 945 uio.uio_resid = ct->ct_record_resid; 946 m = NULL; 947 rcvflag = MSG_DONTWAIT | MSG_SOCALLBCK; 948 error = soreceive(so, NULL, &uio, &m, NULL, &rcvflag); 949 SOCKBUF_LOCK(&so->so_rcv); 950 951 if (error == EWOULDBLOCK) 952 break; 953 954 if (error || uio.uio_resid == ct->ct_record_resid) 955 goto wakeup_all; 956 957 /* 958 * If we have part of the record already, 959 * chain this bit onto the end. 960 */ 961 if (ct->ct_record) 962 m_last(ct->ct_record)->m_next = m; 963 else 964 ct->ct_record = m; 965 966 ct->ct_record_resid = uio.uio_resid; 967 968 /* 969 * If we have the entire record, see if we can 970 * match it to a request. 971 */ 972 if (ct->ct_record_resid == 0 973 && ct->ct_record_eor) { 974 /* 975 * The XID is in the first uint32_t of 976 * the reply. 977 */ 978 if (ct->ct_record->m_len < sizeof(xid)) 979 ct->ct_record = 980 m_pullup(ct->ct_record, 981 sizeof(xid)); 982 if (!ct->ct_record) 983 break; 984 bcopy(mtod(ct->ct_record, uint32_t *), 985 &xid, sizeof(uint32_t)); 986 xid = ntohl(xid); 987 988 mtx_lock(&ct->ct_lock); 989 foundreq = 0; 990 TAILQ_FOREACH(cr, &ct->ct_pending, cr_link) { 991 if (cr->cr_xid == xid) { 992 /* 993 * This one 994 * matches. We leave 995 * the reply mbuf in 996 * cr->cr_mrep. Set 997 * the XID to zero so 998 * that we will ignore 999 * any duplicaed 1000 * replies. 1001 */ 1002 cr->cr_xid = 0; 1003 cr->cr_mrep = ct->ct_record; 1004 cr->cr_error = 0; 1005 foundreq = 1; 1006 wakeup(cr); 1007 break; 1008 } 1009 } 1010 mtx_unlock(&ct->ct_lock); 1011 1012 if (!foundreq) 1013 m_freem(ct->ct_record); 1014 ct->ct_record = NULL; 1015 } 1016 } 1017 } while (m); 1018 ct->ct_upcallrefs--; 1019 if (ct->ct_upcallrefs < 0) 1020 panic("rpcvc upcall refcnt"); 1021 if (ct->ct_upcallrefs == 0) 1022 wakeup(&ct->ct_upcallrefs); 1023 return (SU_OK); 1024 } 1025 1026 /* 1027 * Wait for all upcalls in progress to complete. 1028 */ 1029 static void 1030 clnt_vc_upcallsdone(struct ct_data *ct) 1031 { 1032 1033 SOCKBUF_LOCK_ASSERT(&ct->ct_socket->so_rcv); 1034 1035 while (ct->ct_upcallrefs > 0) 1036 (void) msleep(&ct->ct_upcallrefs, 1037 SOCKBUF_MTX(&ct->ct_socket->so_rcv), 0, "rpcvcup", 0); 1038 } 1039