/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright 2008 Sun Microsystems, Inc. All rights reserved. * Use is subject to license terms. */ /* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */ /* All Rights Reserved */ /* * Portions of this source code were derived from Berkeley * 4.3 BSD under license from the Regents of the University of * California. */ /* * svc.h, Server-side remote procedure call interface. */ #ifndef _RPC_SVC_H #define _RPC_SVC_H #pragma ident "%Z%%M% %I% %E% SMI" #include #include #include #include #include #ifdef _KERNEL #include #include #endif /* _KERNEL */ /* * This interface must manage two items concerning remote procedure calling: * * 1) An arbitrary number of transport connections upon which rpc requests * are received. They are created and registered by routines in svc_generic.c, * svc_vc.c and svc_dg.c; they in turn call xprt_register and * xprt_unregister. * * 2) An arbitrary number of locally registered services. Services are * described by the following four data: program number, version number, * "service dispatch" function, a transport handle, and a boolean that * indicates whether or not the exported program should be registered with a * local binder service; if true the program's number and version and the * address from the transport handle are registered with the binder. * These data are registered with rpcbind via svc_reg(). * * A service's dispatch function is called whenever an rpc request comes in * on a transport. The request's program and version numbers must match * those of the registered service. The dispatch function is passed two * parameters, struct svc_req * and SVCXPRT *, defined below. */ #ifdef __cplusplus extern "C" { #endif /* * Server-side transport handles. * The actual type definitions are below. */ #ifdef _KERNEL typedef struct __svcmasterxprt SVCMASTERXPRT; /* Master transport handle */ typedef struct __svcxprt SVCXPRT; /* Per-thread clone handle */ typedef struct __svcpool SVCPOOL; /* Kernel thread pool */ #else /* _KERNEL */ typedef struct __svcxprt SVCXPRT; /* Server transport handle */ #endif /* _KERNEL */ /* * Prototype of error handler callback */ #ifndef _KERNEL typedef void (*svc_errorhandler_t)(const SVCXPRT* svc, const bool_t isAConn); #endif /* * Service request. * * PSARC 2003/523 Contract Private Interface * svc_req * Changes must be reviewed by Solaris File Sharing * Changes must be communicated to contract-2003-523@sun.com */ struct svc_req { rpcprog_t rq_prog; /* service program number */ rpcvers_t rq_vers; /* service protocol version */ rpcproc_t rq_proc; /* the desired procedure */ struct opaque_auth rq_cred; /* raw creds from the wire */ caddr_t rq_clntcred; /* read only cooked cred */ SVCXPRT *rq_xprt; /* associated transport */ bslabel_t *rq_label; /* TSOL label of the request */ }; #ifdef _KERNEL struct dupreq { uint32_t dr_xid; rpcproc_t dr_proc; rpcvers_t dr_vers; rpcprog_t dr_prog; struct netbuf dr_addr; struct netbuf dr_resp; void (*dr_resfree)(); int dr_status; struct dupreq *dr_next; struct dupreq *dr_chain; }; /* * States of requests for duplicate request caching. */ #define DUP_NEW 0x00 /* new entry */ #define DUP_INPROGRESS 0x01 /* request already going */ #define DUP_DONE 0x02 /* request done */ #define DUP_DROP 0x03 /* request dropped */ #define DUP_ERROR 0x04 /* error in dup req cache */ /* * Prototype for a service dispatch routine. */ typedef void (SVC_DISPATCH)(struct svc_req *, SVCXPRT *); /* * The service provider callout. * Each entry identifies a dispatch routine to be called * for a given RPC program number and a version fitting * into the registered range. */ typedef struct { rpcprog_t sc_prog; /* RPC Program number */ rpcvers_t sc_versmin; /* Min version number */ rpcvers_t sc_versmax; /* Max version number */ SVC_DISPATCH *sc_dispatch; /* Dispatch routine */ } SVC_CALLOUT; /* * Table of service provider `callouts' for an RPC * transport handle. If sct_free is TRUE then transport * destructor is supposed to deallocate this table. */ typedef struct { size_t sct_size; /* Number of entries */ bool_t sct_free; /* Deallocate if true */ SVC_CALLOUT *sct_sc; /* Callout entries */ } SVC_CALLOUT_TABLE; struct svc_ops { bool_t (*xp_recv)(SVCXPRT *, mblk_t *, struct rpc_msg *); /* receive incoming requests */ bool_t (*xp_getargs)(SVCXPRT *, xdrproc_t, caddr_t); /* get arguments */ bool_t (*xp_reply)(SVCXPRT *, struct rpc_msg *); /* send reply */ bool_t (*xp_freeargs)(SVCXPRT *, xdrproc_t, caddr_t); /* free mem allocated for args */ void (*xp_destroy)(SVCMASTERXPRT *); /* destroy this struct */ int (*xp_dup)(struct svc_req *, caddr_t, int, struct dupreq **, bool_t *); /* check for dup */ void (*xp_dupdone)(struct dupreq *, caddr_t, void (*)(), int, int); /* mark dup entry as completed */ int32_t *(*xp_getres)(SVCXPRT *, int); /* get pointer to response buffer */ void (*xp_freeres)(SVCXPRT *); /* destroy pre-serialized response */ void (*xp_clone_destroy)(SVCXPRT *); /* destroy a clone xprt */ void (*xp_start)(SVCMASTERXPRT *); /* `ready-to-receive' */ }; #else /* _KERNEL */ /* * Service control requests */ #define SVCGET_VERSQUIET 1 #define SVCSET_VERSQUIET 2 #define SVCGET_XID 4 #define SVCSET_KEEPALIVE 5 #define SVCSET_CONNMAXREC 6 #define SVCGET_CONNMAXREC 7 #define SVCGET_RECVERRHANDLER 8 #define SVCSET_RECVERRHANDLER 9 enum xprt_stat { XPRT_DIED, XPRT_MOREREQS, XPRT_IDLE }; struct xp_ops { #ifdef __STDC__ bool_t (*xp_recv)(SVCXPRT *, struct rpc_msg *); /* receive incoming requests */ enum xprt_stat (*xp_stat)(SVCXPRT *); /* get transport status */ bool_t (*xp_getargs)(SVCXPRT *, xdrproc_t, caddr_t); /* get arguments */ bool_t (*xp_reply)(SVCXPRT *, struct rpc_msg *); /* send reply */ bool_t (*xp_freeargs)(SVCXPRT *, xdrproc_t, caddr_t); /* free mem allocated for args */ void (*xp_destroy)(SVCXPRT *); /* destroy this struct */ bool_t (*xp_control)(SVCXPRT *, const uint_t, void *); /* catch-all control function */ #else /* __STDC__ */ bool_t (*xp_recv)(); /* receive incoming requests */ enum xprt_stat (*xp_stat)(); /* get transport status */ bool_t (*xp_getargs)(); /* get arguments */ bool_t (*xp_reply)(); /* send reply */ bool_t (*xp_freeargs)(); /* free mem allocated for args */ void (*xp_destroy)(); /* destroy this struct */ bool_t (*xp_control)(); /* catch-all control function */ #endif /* __STDC__ */ }; #endif /* _KERNEL */ #ifdef _KERNEL /* * SVCPOOL * Kernel RPC server-side thread pool structure. */ typedef struct __svcxprt_qnode __SVCXPRT_QNODE; /* Defined in svc.c */ struct __svcpool { /* * Thread pool variables. * * The pool's thread lock p_thread_lock protects: * - p_threads, p_detached_threads, p_reserved_threads and p_closing * The pool's request lock protects: * - p_asleep, p_drowsy, p_reqs, p_walkers, p_req_cv. * The following fields are `initialized constants': * - p_id, p_stksize, p_timeout. * Access to p_next and p_prev is protected by the pool * list lock. */ SVCPOOL *p_next; /* Next pool in the list */ SVCPOOL *p_prev; /* Prev pool in the list */ int p_id; /* Pool id */ int p_threads; /* Non-detached threads */ int p_detached_threads; /* Detached threads */ int p_maxthreads; /* Max threads in the pool */ int p_redline; /* `Redline' for the pool */ int p_reserved_threads; /* Reserved threads */ kmutex_t p_thread_lock; /* Thread lock */ int p_asleep; /* Asleep threads */ int p_drowsy; /* Drowsy flag */ kcondvar_t p_req_cv; /* svc_poll() sleep var. */ clock_t p_timeout; /* svc_poll() timeout */ kmutex_t p_req_lock; /* Request lock */ int p_reqs; /* Pending requests */ int p_walkers; /* Walking threads */ int p_max_same_xprt; /* Max reqs from the xprt */ int p_stksize; /* Stack size for svc_run */ bool_t p_closing : 1; /* Pool is closing */ /* * Thread creator variables. * The `creator signaled' flag is turned on when a signal is send * to the creator thread (to create a new service thread). The * creator clears when the thread is created. The protocol is not * to signal the creator thread when the flag is on. However, * a new thread should signal the creator if there are more * requests in the queue. * * When the pool is closing (ie it has been already unregistered from * the pool list) the last thread on the last transport should turn * the p_creator_exit flag on. This tells the creator thread to * free the pool structure and exit. */ bool_t p_creator_signaled : 1; /* Create requested flag */ bool_t p_creator_exit : 1; /* If true creator exits */ kcondvar_t p_creator_cv; /* Creator cond. variable */ kmutex_t p_creator_lock; /* Creator lock */ /* * Doubly linked list containing `registered' master transport handles. * There is no special structure for a list node. Instead the * SVCMASTERXPRT structure has the xp_next and xp_prev fields. * * The p_lrwlock protects access to xprt->xp_next and xprt->xp_prev. * A service thread should also acquire a reader lock before accessing * any transports it is no longer linked to (to prevent them from * being destroyed). * * The list lock governs also the `pool is closing' flag. */ size_t p_lcount; /* Current count */ SVCMASTERXPRT *p_lhead; /* List head */ krwlock_t p_lrwlock; /* R/W lock */ /* * Circular linked list for the `xprt-ready' queue (FIFO). * Must be initialized with svc_xprt_qinit() before it is used. * * The writer's end is protected by the pool's request lock * (pool->p_req_lock). The reader's end is protected by q_end_lock. * * When the queue is full the p_qoverflow flag is raised. It stays * on until all the pending request are drained. */ size_t p_qsize; /* Number of queue nodes */ int p_qoverflow : 1; /* Overflow flag */ __SVCXPRT_QNODE *p_qbody; /* Queue body (array) */ __SVCXPRT_QNODE *p_qtop; /* Writer's end of FIFO */ __SVCXPRT_QNODE *p_qend; /* Reader's end of FIFO */ kmutex_t p_qend_lock; /* Reader's end lock */ /* * Userspace thread creator variables. * Thread creation is actually done in userland, via a thread * that is parked in the kernel. When that thread is signaled, * it returns back down to the daemon from whence it came and * does the lwp create. * * A parallel "creator" thread runs in the kernel. That is the * thread that will signal for the user thread to return to * userland and do its work. * * Since the thread doesn't always exist (there could be a race * if two threads are created in rapid succession), we set * p_signal_create_thread to FALSE when we're ready to accept work. * * p_user_exit is set to true when the service pool is about * to close. This is done so that the user creation thread * can be informed and cleanup any userland state. */ bool_t p_signal_create_thread : 1; /* Create requested flag */ bool_t p_user_exit : 1; /* If true creator exits */ bool_t p_user_waiting : 1; /* Thread waiting for work */ kcondvar_t p_user_cv; /* Creator cond. variable */ kmutex_t p_user_lock; /* Creator lock */ void (*p_offline)(); /* callout for unregister */ void (*p_shutdown)(); /* callout for shutdown */ }; /* * Server side transport handle (SVCMASTERXPRT). * xprt->xp_req_lock governs the following fields in xprt: * xp_req_head, xp_req_tail. * xprt->xp_thread_lock governs the following fields in xprt: * xp_threads, xp_detached_threads. * * xp_req_tail is only valid if xp_req_head is non-NULL * * The xp_threads count is the number of attached threads. These threads * are able to handle new requests, and it is expected that they will not * block for a very long time handling a given request. The * xp_detached_threads count is the number of threads that have detached * themselves from the transport. These threads can block indefinitely * while handling a request. Once they complete the request, they exit. * * A kernel service provider may register a callback function "closeproc" * for a transport. When the transport is closing the last exiting attached * thread - xp_threads goes to zero - it calls the callback function, passing * it a reference to the transport. This call is made with xp_thread_lock * held, so any cleanup bookkeeping it does should be done quickly. * * When the transport is closing the last exiting thread is supposed * to destroy/free the data structure. */ typedef struct __svcxprt_common { struct file *xpc_fp; struct svc_ops *xpc_ops; queue_t *xpc_wq; /* queue to write onto */ cred_t *xpc_cred; /* cached cred for server to use */ int32_t xpc_type; /* transport type */ int xpc_msg_size; /* TSDU or TIDU size */ struct netbuf xpc_rtaddr; /* remote transport address */ struct netbuf xpc_lcladdr; /* local transport address */ SVC_CALLOUT_TABLE *xpc_sct; } __SVCXPRT_COMMON; #define xp_fp xp_xpc.xpc_fp #define xp_ops xp_xpc.xpc_ops #define xp_wq xp_xpc.xpc_wq #define xp_cred xp_xpc.xpc_cred #define xp_type xp_xpc.xpc_type #define xp_msg_size xp_xpc.xpc_msg_size #define xp_rtaddr xp_xpc.xpc_rtaddr #define xp_lcladdr xp_xpc.xpc_lcladdr #define xp_sct xp_xpc.xpc_sct struct __svcmasterxprt { SVCMASTERXPRT *xp_next; /* Next transport in the list */ SVCMASTERXPRT *xp_prev; /* Prev transport in the list */ __SVCXPRT_COMMON xp_xpc; /* Fields common with the clone */ SVCPOOL *xp_pool; /* Pointer to the pool */ mblk_t *xp_req_head; /* Request queue head */ mblk_t *xp_req_tail; /* Request queue tail */ kmutex_t xp_req_lock; /* Request lock */ int xp_threads; /* Current num. of attached threads */ int xp_detached_threads; /* num. of detached threads */ kmutex_t xp_thread_lock; /* Thread count lock */ void (*xp_closeproc)(const SVCMASTERXPRT *); /* optional; see comments above */ char *xp_netid; /* network token */ struct netbuf xp_addrmask; /* address mask */ caddr_t xp_p2; /* private: for use by svc ops */ }; /* * Service thread `clone' transport handle (SVCXPRT) * * PSARC 2003/523 Contract Private Interface * SVCXPRT * Changes must be reviewed by Solaris File Sharing * Changes must be communicated to contract-2003-523@sun.com * * The xp_p2buf buffer is used as the storage for a transport type * specific structure. It is private for the svc ops for a given * transport type. */ #define SVC_P2LEN 64 struct __svcxprt { __SVCXPRT_COMMON xp_xpc; SVCMASTERXPRT *xp_master; /* back ptr to master */ /* The following fileds are on a per-thread basis */ callb_cpr_t *xp_cprp; /* unused padding for Contract */ bool_t xp_reserved : 1; /* is thread reserved? */ bool_t xp_detached : 1; /* is thread detached? */ int xp_same_xprt; /* Reqs from the same xprt */ /* The following fields are used on a per-request basis */ struct opaque_auth xp_verf; /* raw response verifier */ SVCAUTH xp_auth; /* auth flavor of current req */ void *xp_cookie; /* a cookie */ uint32_t xp_xid; /* id */ XDR xp_xdrin; /* input xdr stream */ XDR xp_xdrout; /* output xdr stream */ /* Private for svc ops */ char xp_p2buf[SVC_P2LEN]; /* udp_data or cots_data_t */ /* or clone_rdma_data_t */ }; #else /* _KERNEL */ struct __svcxprt { int xp_fd; #define xp_sock xp_fd ushort_t xp_port; /* * associated port number. * Obsolete, but still used to * specify whether rendezvouser * or normal connection */ struct xp_ops *xp_ops; int xp_addrlen; /* length of remote addr. Obsoleted */ char *xp_tp; /* transport provider device name */ char *xp_netid; /* network token */ struct netbuf xp_ltaddr; /* local transport address */ struct netbuf xp_rtaddr; /* remote transport address */ char xp_raddr[16]; /* remote address. Now obsoleted */ struct opaque_auth xp_verf; /* raw response verifier */ caddr_t xp_p1; /* private: for use by svc ops */ caddr_t xp_p2; /* private: for use by svc ops */ caddr_t xp_p3; /* private: for use by svc lib */ int xp_type; /* transport type */ /* * callback on client death * First parameter is the current structure, * Second parameter : * - FALSE for the service listener * - TRUE for a real connected socket */ svc_errorhandler_t xp_closeclnt; }; #endif /* _KERNEL */ /* * Approved way of getting address of caller, * address mask, and netid of transport. */ #define svc_getrpccaller(x) (&(x)->xp_rtaddr) #ifdef _KERNEL #define svc_getcaller(x) (&(x)->xp_rtaddr.buf) #define svc_getaddrmask(x) (&(x)->xp_master->xp_addrmask) #define svc_getnetid(x) ((x)->xp_master->xp_netid) #endif /* _KERNEL */ /* * Operations defined on an SVCXPRT handle */ #ifdef _KERNEL #define SVC_RECV(clone_xprt, mp, msg) \ (*(clone_xprt)->xp_ops->xp_recv)((clone_xprt), (mp), (msg)) /* * PSARC 2003/523 Contract Private Interface * SVC_GETARGS * Changes must be reviewed by Solaris File Sharing * Changes must be communicated to contract-2003-523@sun.com */ #define SVC_GETARGS(clone_xprt, xargs, argsp) \ (*(clone_xprt)->xp_ops->xp_getargs)((clone_xprt), (xargs), (argsp)) #define SVC_REPLY(clone_xprt, msg) \ (*(clone_xprt)->xp_ops->xp_reply) ((clone_xprt), (msg)) #define SVC_FREEARGS(clone_xprt, xargs, argsp) \ (*(clone_xprt)->xp_ops->xp_freeargs)((clone_xprt), (xargs), (argsp)) #define SVC_GETRES(clone_xprt, size) \ (*(clone_xprt)->xp_ops->xp_getres)((clone_xprt), (size)) #define SVC_FREERES(clone_xprt) \ (*(clone_xprt)->xp_ops->xp_freeres)(clone_xprt) #define SVC_DESTROY(xprt) \ (*(xprt)->xp_ops->xp_destroy)(xprt) /* * PSARC 2003/523 Contract Private Interfaces * SVC_DUP, SVC_DUPDONE, SVC_DUP_EXT, SVC_DUPDONE_EXT * Changes must be reviewed by Solaris File Sharing * Changes must be communicated to contract-2003-523@sun.com * * SVC_DUP and SVC_DUPDONE are defined here for backward compatibility. */ #define SVC_DUP_EXT(clone_xprt, req, res, size, drpp, dupcachedp) \ (*(clone_xprt)->xp_ops->xp_dup)(req, res, size, drpp, dupcachedp) #define SVC_DUPDONE_EXT(clone_xprt, dr, res, resfree, size, status) \ (*(clone_xprt)->xp_ops->xp_dupdone)(dr, res, resfree, size, status) #define SVC_DUP(clone_xprt, req, res, size, drpp) \ (*(clone_xprt)->xp_ops->xp_dup)(req, res, size, drpp, NULL) #define SVC_DUPDONE(clone_xprt, dr, res, size, status) \ (*(clone_xprt)->xp_ops->xp_dupdone)(dr, res, NULL, size, status) #define SVC_CLONE_DESTROY(clone_xprt) \ (*(clone_xprt)->xp_ops->xp_clone_destroy)(clone_xprt) #define SVC_START(xprt) \ (*(xprt)->xp_ops->xp_start)(xprt) #else /* _KERNEL */ #define SVC_RECV(xprt, msg) \ (*(xprt)->xp_ops->xp_recv)((xprt), (msg)) #define svc_recv(xprt, msg) \ (*(xprt)->xp_ops->xp_recv)((xprt), (msg)) #define SVC_STAT(xprt) \ (*(xprt)->xp_ops->xp_stat)(xprt) #define svc_stat(xprt) \ (*(xprt)->xp_ops->xp_stat)(xprt) #define SVC_GETARGS(xprt, xargs, argsp) \ (*(xprt)->xp_ops->xp_getargs)((xprt), (xargs), (argsp)) #define svc_getargs(xprt, xargs, argsp) \ (*(xprt)->xp_ops->xp_getargs)((xprt), (xargs), (argsp)) #define SVC_REPLY(xprt, msg) \ (*(xprt)->xp_ops->xp_reply) ((xprt), (msg)) #define svc_reply(xprt, msg) \ (*(xprt)->xp_ops->xp_reply) ((xprt), (msg)) #define SVC_FREEARGS(xprt, xargs, argsp) \ (*(xprt)->xp_ops->xp_freeargs)((xprt), (xargs), (argsp)) #define svc_freeargs(xprt, xargs, argsp) \ (*(xprt)->xp_ops->xp_freeargs)((xprt), (xargs), (argsp)) #define SVC_GETRES(xprt, size) \ (*(xprt)->xp_ops->xp_getres)((xprt), (size)) #define svc_getres(xprt, size) \ (*(xprt)->xp_ops->xp_getres)((xprt), (size)) #define SVC_FREERES(xprt) \ (*(xprt)->xp_ops->xp_freeres)(xprt) #define svc_freeres(xprt) \ (*(xprt)->xp_ops->xp_freeres)(xprt) #define SVC_DESTROY(xprt) \ (*(xprt)->xp_ops->xp_destroy)(xprt) #define svc_destroy(xprt) \ (*(xprt)->xp_ops->xp_destroy)(xprt) /* * PSARC 2003/523 Contract Private Interface * SVC_CONTROL * Changes must be reviewed by Solaris File Sharing * Changes must be communicated to contract-2003-523@sun.com */ #define SVC_CONTROL(xprt, rq, in) \ (*(xprt)->xp_ops->xp_control)((xprt), (rq), (in)) #endif /* _KERNEL */ /* * Pool id's reserved for NFS, NLM, and the NFSv4 callback program. */ #define NFS_SVCPOOL_ID 0x01 #define NLM_SVCPOOL_ID 0x02 #define NFS_CB_SVCPOOL_ID 0x03 #define RDC_SVCPOOL_ID 0x05 /* SNDR, PSARC 2001/699 */ struct svcpool_args { uint32_t id; /* Pool id */ uint32_t maxthreads; /* Max threads in the pool */ uint32_t redline; /* `Redline' for the pool */ uint32_t qsize; /* `xprt-ready' queue size */ uint32_t timeout; /* svc_poll() timeout */ uint32_t stksize; /* svc_run() stack size */ uint32_t max_same_xprt; /* Max reqs from the same xprt */ }; #ifdef _KERNEL /* * Transport registration and thread pool creation. */ extern int svc_xprt_register(SVCMASTERXPRT *, int); extern void svc_xprt_unregister(SVCMASTERXPRT *); extern int svc_pool_create(struct svcpool_args *); extern int svc_wait(int); extern int svc_do_run(int); #define SVCPSET_SHUTDOWN_PROC 1 #define SVCPSET_UNREGISTER_PROC 2 extern int svc_pool_control(int, int, void *); #else /* _KERNEL */ #ifdef __STDC__ extern bool_t rpc_reg(const rpcprog_t, const rpcvers_t, const rpcproc_t, char *(*)(char *), const xdrproc_t, const xdrproc_t, const char *); /* * Service registration * * svc_reg(xprt, prog, vers, dispatch, nconf) * const SVCXPRT *xprt; * const rpcprog_t prog; * const rpcvers_t vers; * const void (*dispatch)(); * const struct netconfig *nconf; */ extern bool_t svc_reg(const SVCXPRT *, const rpcprog_t, const rpcvers_t, void (*)(struct svc_req *, SVCXPRT *), const struct netconfig *); /* * Service authentication registration * * svc_auth_reg(cred_flavor, handler) * int cred_flavor; * enum auth_stat (*handler)(); */ extern int svc_auth_reg(int, enum auth_stat (*)()); /* * Service un-registration * * svc_unreg(prog, vers) * const rpcprog_t prog; * const rpcvers_t vers; */ extern void svc_unreg(const rpcprog_t, const rpcvers_t); /* * Transport registration/unregistration. * * xprt_register(xprt) * const SVCXPRT *xprt; * * xprt_unregister(xprt) * const SVCXPRT *xprt; */ extern void xprt_register(const SVCXPRT *); extern void xprt_unregister(const SVCXPRT *); #else /* __STDC__ */ extern bool_t rpc_reg(); extern bool_t svc_reg(); extern bool_t svc_auth_reg(); extern void svc_unreg(); extern void xprt_register(); extern void xprt_unregister(); #endif /* __STDC__ */ #endif /* _KERNEL */ /* * When the service routine is called, it must first check to see if it * knows about the procedure; if not, it should call svcerr_noproc * and return. If so, it should deserialize its arguments via * SVC_GETARGS (defined above). If the deserialization does not work, * svcerr_decode should be called followed by a return. Successful * decoding of the arguments should be followed the execution of the * procedure's code and a call to svc_sendreply. * * Also, if the service refuses to execute the procedure due to too- * weak authentication parameters, svcerr_weakauth should be called. * Note: do not confuse access-control failure with weak authentication! * * NB: In pure implementations of rpc, the caller always waits for a reply * msg. This message is sent when svc_sendreply is called. * Therefore pure service implementations should always call * svc_sendreply even if the function logically returns void; use * xdr.h - xdr_void for the xdr routine. HOWEVER, connectionful rpc allows * for the abuse of pure rpc via batched calling or pipelining. In the * case of a batched call, svc_sendreply should NOT be called since * this would send a return message, which is what batching tries to avoid. * It is the service/protocol writer's responsibility to know which calls are * batched and which are not. Warning: responding to batch calls may * deadlock the caller and server processes! */ #ifdef __STDC__ extern bool_t svc_sendreply(const SVCXPRT *, const xdrproc_t, const caddr_t); extern void svcerr_decode(const SVCXPRT *); extern void svcerr_weakauth(const SVCXPRT *); extern void svcerr_noproc(const SVCXPRT *); extern void svcerr_progvers(const SVCXPRT *, const rpcvers_t, const rpcvers_t); extern void svcerr_auth(const SVCXPRT *, const enum auth_stat); extern void svcerr_noprog(const SVCXPRT *); extern void svcerr_systemerr(const SVCXPRT *); extern void svcerr_badcred(const SVCXPRT *); #else /* __STDC__ */ extern bool_t svc_sendreply(); extern void svcerr_decode(); extern void svcerr_weakauth(); extern void svcerr_noproc(); extern void svcerr_progvers(); extern void svcerr_auth(); extern void svcerr_noprog(); extern void svcerr_systemerr(); extern void svcerr_badcred(); #endif /* __STDC__ */ #ifdef _KERNEL /* * Kernel RPC functions. */ extern void svc_init(void); extern void svc_cots_init(void); extern void svc_clts_init(void); extern void mt_kstat_init(void); extern void mt_kstat_fini(void); extern int svc_tli_kcreate(struct file *, uint_t, char *, struct netbuf *, SVCMASTERXPRT **, SVC_CALLOUT_TABLE *, void (*closeproc)(const SVCMASTERXPRT *), int, bool_t); extern int svc_clts_kcreate(struct file *, uint_t, struct T_info_ack *, SVCMASTERXPRT **); extern int svc_cots_kcreate(struct file *, uint_t, struct T_info_ack *, SVCMASTERXPRT **); extern void svc_queuereq(queue_t *, mblk_t *); extern void svc_queueclean(queue_t *); extern void svc_queueclose(queue_t *); extern int svc_reserve_thread(SVCXPRT *); extern void svc_unreserve_thread(SVCXPRT *); extern callb_cpr_t *svc_detach_thread(SVCXPRT *); /* * For RDMA based kRPC. * "rdma_xprt_record" is a reference to master transport handles * in kRPC thread pools. This is an easy way of tracking and shuting * down rdma based kRPC transports on demand. * "rdma_xprt_group" is a list of RDMA based mster transport handles * or records in a kRPC thread pool. */ typedef struct rdma_xprt_record rdma_xprt_record_t; struct rdma_xprt_record { int rtr_type; /* Type of rdma; IB/VI/RDDP */ SVCMASTERXPRT *rtr_xprt_ptr; /* Ptr to master xprt handle */ rdma_xprt_record_t *rtr_next; /* Ptr to next record */ }; typedef struct { int rtg_count; /* Number transport records */ int rtg_poolid; /* Pool Id for this group */ rdma_xprt_record_t *rtg_listhead; /* Head of the records list */ } rdma_xprt_group_t; extern int svc_rdma_kcreate(char *, SVC_CALLOUT_TABLE *, int, rdma_xprt_group_t *); extern void svc_rdma_kstop(SVCMASTERXPRT *); extern void svc_rdma_kdestroy(SVCMASTERXPRT *); extern void rdma_stop(rdma_xprt_group_t); /* * GSS cleanup method. */ extern void rpc_gss_cleanup(SVCXPRT *); #else /* _KERNEL */ /* * Lowest level dispatching -OR- who owns this process anyway. * Somebody has to wait for incoming requests and then call the correct * service routine. The routine svc_run does infinite waiting; i.e., * svc_run never returns. * Since another (co-existant) package may wish to selectively wait for * incoming calls or other events outside of the rpc architecture, the * routine svc_getreq_poll is provided. It must be passed pollfds, the * "in-place" results of a poll call (see poll, section 2). */ /* * Global keeper of rpc service descriptors in use * dynamic; must be inspected before each call to select or poll */ extern pollfd_t *svc_pollfd; extern int svc_max_pollfd; extern fd_set svc_fdset; #if !defined(_LP64) && FD_SETSIZE > 1024 extern fd_set _new_svc_fdset; #ifdef __PRAGMA_REDEFINE_EXTNAME #pragma redefine_extname svc_fdset _new_svc_fdset #else /* __PRAGMA_REDEFINE_EXTNAME */ #define svc_fdset _new_svc_fdset #endif /* __PRAGMA_REDEFINE_EXTNAME */ #endif /* LP64 && FD_SETSIZE > 1024 */ #define svc_fds svc_fdset.fds_bits[0] /* compatibility */ /* * A small program implemented by the svc_rpc implementation itself. * Also see clnt.h for protocol numbers. */ #ifdef __STDC__ extern void svc_getreq(int); extern void svc_getreq_common(const int); extern void svc_getreqset(fd_set *); /* takes fdset instead of int */ extern void svc_getreq_poll(struct pollfd *, const int); extern void svc_run(void); extern void svc_exit(void); #else /* __STDC__ */ extern void rpctest_service(); extern void svc_getreqset(); extern void svc_getreq(); extern void svc_getreq_common(); extern void svc_getreqset(); /* takes fdset instead of int */ extern void svc_getreq_poll(); extern void svc_run(); extern void svc_exit(); #endif /* __STDC__ */ /* * Functions used to manage user file descriptors */ typedef int svc_input_id_t; typedef void (*svc_callback_t)(svc_input_id_t id, int fd, unsigned int events, void* cookie); #ifdef __STDC__ extern svc_input_id_t svc_add_input(int fd, unsigned int events, svc_callback_t user_callback, void* cookie); extern int svc_remove_input(svc_input_id_t id); #else /* __STDC__ */ extern svc_input_id_t svc_add_input(); extern int svc_remove_input(); #endif /* * These are the existing service side transport implementations. * * Transport independent svc_create routine. */ #ifdef __STDC__ extern int svc_create(void (*)(struct svc_req *, SVCXPRT *), const rpcprog_t, const rpcvers_t, const char *); /* * void (*dispatch)(); -- dispatch routine * const rpcprog_t prognum; -- program number * const rpcvers_t versnum; -- version number * const char *nettype; -- network type */ /* * Generic server creation routine. It takes a netconfig structure * instead of a nettype. */ extern SVCXPRT *svc_tp_create(void (*)(struct svc_req *, SVCXPRT *), const rpcprog_t, const rpcvers_t, const struct netconfig *); /* * void (*dispatch)(); -- dispatch routine * const rpcprog_t prognum; -- program number * const rpcvers_t versnum; -- version number * const struct netconfig *nconf; -- netconfig structure */ /* * Generic TLI create routine */ extern SVCXPRT *svc_tli_create(const int, const struct netconfig *, const struct t_bind *, const uint_t, const uint_t); /* * const int fd; -- connection end point * const struct netconfig *nconf; -- netconfig structure * const struct t_bind *bindaddr; -- local bind address * const uint_t sendsz; -- max sendsize * const uint_t recvsz; -- max recvsize */ /* * Connectionless and connectionful create routines. */ extern SVCXPRT *svc_vc_create(const int, const uint_t, const uint_t); /* * const int fd; -- open connection end point * const uint_t sendsize; -- max send size * const uint_t recvsize; -- max recv size */ extern SVCXPRT *svc_dg_create(const int, const uint_t, const uint_t); /* * const int fd; -- open connection * const uint_t sendsize; -- max send size * const uint_t recvsize; -- max recv size */ /* * the routine takes any *open* TLI file * descriptor as its first input and is used for open connections. */ extern SVCXPRT *svc_fd_create(const int, const uint_t, const uint_t); /* * const int fd; -- open connection end point * const uint_t sendsize; -- max send size * const uint_t recvsize; -- max recv size */ /* * Memory based rpc (for speed check and testing) */ extern SVCXPRT *svc_raw_create(void); /* * Creation of service over doors transport. */ extern SVCXPRT *svc_door_create(void (*)(struct svc_req *, SVCXPRT *), const rpcprog_t, const rpcvers_t, const uint_t); /* * void (*dispatch)(); -- dispatch routine * const rpcprog_t prognum; -- program number * const rpcvers_t versnum; -- version number * const uint_t sendsize; -- send buffer size */ /* * Service control interface */ extern bool_t svc_control(SVCXPRT *, const uint_t, void *); /* * SVCXPRT *svc; -- service to manipulate * const uint_t req; -- request * void *info; -- argument to request */ /* * svc_dg_enable_cache() enables the cache on dg transports. */ extern int svc_dg_enablecache(SVCXPRT *, const uint_t); #else /* __STDC__ */ extern int svc_create(); extern SVCXPRT *svc_tp_create(); extern SVCXPRT *svc_tli_create(); extern SVCXPRT *svc_vc_create(); extern SVCXPRT *svc_dg_create(); extern SVCXPRT *svc_fd_create(); extern SVCXPRT *svc_raw_create(); extern SVCXPRT *svc_door_create(); extern int svc_dg_enablecache(); #endif /* __STDC__ */ extern boolean_t is_multilevel(rpcprog_t); #ifdef PORTMAP /* For backward compatibility */ #include #endif /* PORTMAP */ /* * For user level MT hot server functions */ /* * Different MT modes */ #define RPC_SVC_MT_NONE 0 /* default, single-threaded */ #define RPC_SVC_MT_AUTO 1 /* automatic MT mode */ #define RPC_SVC_MT_USER 2 /* user MT mode */ #ifdef __STDC__ extern void svc_done(SVCXPRT *); #else extern void svc_done(); #endif /* __STDC__ */ /* * Obtaining local credentials. */ typedef struct __svc_local_cred_t { uid_t euid; /* effective uid */ gid_t egid; /* effective gid */ uid_t ruid; /* real uid */ gid_t rgid; /* real gid */ pid_t pid; /* caller's pid, or -1 if not available */ } svc_local_cred_t; #ifdef __STDC__ struct ucred_s; extern void svc_fd_negotiate_ucred(int); extern int svc_getcallerucred(const SVCXPRT *, struct ucred_s **); extern bool_t svc_get_local_cred(SVCXPRT *, svc_local_cred_t *); #else extern void svc_fd_negotiate_ucred(); extern int svc_getcallerucred(); extern bool_t svc_get_local_cred(); #endif /* __STDC__ */ /* * Private interfaces and structures for user level duplicate request caching. * The interfaces and data structures are not committed and subject to * change in future releases. Currently only intended for use by automountd. */ struct dupreq { uint32_t dr_xid; rpcproc_t dr_proc; rpcvers_t dr_vers; rpcprog_t dr_prog; struct netbuf dr_addr; struct netbuf dr_resp; int dr_status; time_t dr_time; uint_t dr_hash; struct dupreq *dr_next; struct dupreq *dr_prev; struct dupreq *dr_chain; struct dupreq *dr_prevchain; }; /* * The fixedtime state is defined if we want to expand the routines to * handle and encompass fixed size caches. */ #define DUPCACHE_FIXEDTIME 0 /* * States of requests for duplicate request caching. * These are the same as defined for the kernel. */ #define DUP_NEW 0x00 /* new entry */ #define DUP_INPROGRESS 0x01 /* request already going */ #define DUP_DONE 0x02 /* request done */ #define DUP_DROP 0x03 /* request dropped */ #define DUP_ERROR 0x04 /* error in dup req cache */ #ifdef __STDC__ extern bool_t __svc_dupcache_init(void *, int, char **); extern int __svc_dup(struct svc_req *, caddr_t *, uint_t *, char *); extern int __svc_dupdone(struct svc_req *, caddr_t, uint_t, int, char *); extern bool_t __svc_vc_dupcache_init(SVCXPRT *, void *, int); extern int __svc_vc_dup(struct svc_req *, caddr_t *, uint_t *); extern int __svc_vc_dupdone(struct svc_req *, caddr_t, uint_t, int); #else extern bool_t __svc_dupcache_init(); extern int __svc_dup(); extern int __svc_dupdone(); extern bool_t __svc_vc_dupcache_init(); extern int __svc_vc_dup(); extern int __svc_vc_dupdone(); #endif /* __STDC__ */ #endif /* _KERNEL */ #ifdef __cplusplus } #endif #endif /* !_RPC_SVC_H */