/* * Copyright (c) 2008 Isilon Inc http://www.isilon.com/ * Authors: Doug Rabson * Developed with Red Inc: Alfred Perlstein * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ /* * Copyright (c) 2012 by Delphix. All rights reserved. * Copyright 2013 Nexenta Systems, Inc. All rights reserved. */ /* * NFS Lock Manager service functions (nlm_do_...) * Called from nlm_rpc_svc.c wrappers. * * Source code derived from FreeBSD nlm_prot_impl.c */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "nlm_impl.h" #define NLM_IN_GRACE(g) (ddi_get_lbolt() < (g)->grace_threshold) struct nlm_block_cb_data { struct nlm_host *hostp; struct nlm_vhold *nvp; struct flock64 *flp; }; /* * Invoke an asyncronous RPC callbeck * (used when NLM server needs to reply to MSG NLM procedure). */ #define NLM_INVOKE_CALLBACK(descr, rpcp, resp, callb) \ do { \ enum clnt_stat _stat; \ \ _stat = (*(callb))(resp, NULL, (rpcp)->nr_handle); \ if (_stat != RPC_SUCCESS && _stat != RPC_TIMEDOUT) { \ struct rpc_err _err; \ \ CLNT_GETERR((rpcp)->nr_handle, &_err); \ NLM_ERR("NLM: %s callback failed: " \ "stat %d, err %d\n", descr, _stat, \ _err.re_errno); \ } \ \ _NOTE(CONSTCOND) } while (0) static void nlm_block( nlm4_lockargs *lockargs, struct nlm_host *host, struct nlm_vhold *nvp, nlm_rpc_t *rpcp, struct flock64 *fl, nlm_testargs_cb grant_cb); static vnode_t *nlm_fh_to_vp(struct netobj *); static struct nlm_vhold *nlm_fh_to_vhold(struct nlm_host *, struct netobj *); static void nlm_init_shrlock(struct shrlock *, nlm4_share *, struct nlm_host *); static callb_cpr_t *nlm_block_callback(flk_cb_when_t, void *); static int nlm_vop_frlock(vnode_t *, int, flock64_t *, int, offset_t, struct flk_callback *, cred_t *, caller_context_t *); /* * Convert a lock from network to local form, and * check for valid range (no overflow). */ static int nlm_init_flock(struct flock64 *fl, struct nlm4_lock *nl, struct nlm_host *host, rpcvers_t vers, short type) { uint64_t off, len; bzero(fl, sizeof (*fl)); off = nl->l_offset; len = nl->l_len; if (vers < NLM4_VERS) { if (off > MAX_UOFF32 || len > MAX_UOFF32) return (EINVAL); if (off + len > MAX_UOFF32 + 1) return (EINVAL); } else { /* * Check range for 64-bit client (no overflow). * Again allow len == ~0 to mean lock to EOF. */ if (len == MAX_U_OFFSET_T) len = 0; if (len != 0 && off + (len - 1) < off) return (EINVAL); } fl->l_type = type; fl->l_whence = SEEK_SET; fl->l_start = off; fl->l_len = len; fl->l_sysid = host->nh_sysid; fl->l_pid = nl->svid; /* l_pad */ return (0); } /* * Gets vnode from client's filehandle * NOTE: Holds vnode, it _must_ be explicitly * released by VN_RELE(). */ static vnode_t * nlm_fh_to_vp(struct netobj *fh) { fhandle_t *fhp; /* * Get a vnode pointer for the given NFS file handle. * Note that it could be an NFSv2 for NFSv3 handle, * which means the size might vary. (don't copy) */ if (fh->n_len < sizeof (*fhp)) return (NULL); /* We know this is aligned (kmem_alloc) */ /* LINTED E_BAD_PTR_CAST_ALIGN */ fhp = (fhandle_t *)fh->n_bytes; return (lm_fhtovp(fhp)); } /* * Get vhold from client's filehandle, but in contrast to * The function tries to check some access rights as well. * * NOTE: vhold object _must_ be explicitly released by * nlm_vhold_release(). */ static struct nlm_vhold * nlm_fh_to_vhold(struct nlm_host *hostp, struct netobj *fh) { vnode_t *vp; struct nlm_vhold *nvp; vp = nlm_fh_to_vp(fh); if (vp == NULL) return (NULL); nvp = nlm_vhold_get(hostp, vp); /* * Both nlm_fh_to_vp() and nlm_vhold_get() * do VN_HOLD(), so we need to drop one * reference on vnode. */ VN_RELE(vp); return (nvp); } /* ******************************************************************* */ /* * NLM implementation details, called from the RPC svc code. */ /* * Call-back from NFS statd, used to notify that one of our * hosts had a status change. The host can be either an * NFS client, NFS server or both. * According to NSM protocol description, the state is a * number that is increases monotonically each time the * state of host changes. An even number indicates that * the host is down, while an odd number indicates that * the host is up. * * Here we ignore this even/odd difference of status number * reported by the NSM, we launch notification handlers * every time the state is changed. The reason we why do so * is that client and server can talk to each other using * connectionless transport and it's easy to lose packet * containing NSM notification with status number update. * * In nlm_host_monitor(), we put the sysid in the private data * that statd carries in this callback, so we can easliy find * the host this call applies to. */ /* ARGSUSED */ void nlm_do_notify1(nlm_sm_status *argp, void *res, struct svc_req *sr) { struct nlm_globals *g; struct nlm_host *host; uint16_t sysid; g = zone_getspecific(nlm_zone_key, curzone); bcopy(&argp->priv, &sysid, sizeof (sysid)); DTRACE_PROBE2(nsm__notify, uint16_t, sysid, int, argp->state); host = nlm_host_find_by_sysid(g, (sysid_t)sysid); if (host == NULL) return; nlm_host_notify_server(host, argp->state); nlm_host_notify_client(host, argp->state); nlm_host_release(g, host); } /* * Another available call-back for NFS statd. * Not currently used. */ /* ARGSUSED */ void nlm_do_notify2(nlm_sm_status *argp, void *res, struct svc_req *sr) { ASSERT(0); } /* * NLM_TEST, NLM_TEST_MSG, * NLM4_TEST, NLM4_TEST_MSG, * Client inquiry about locks, non-blocking. */ void nlm_do_test(nlm4_testargs *argp, nlm4_testres *resp, struct svc_req *sr, nlm_testres_cb cb) { struct nlm_globals *g; struct nlm_host *host; struct nlm4_holder *lh; struct nlm_owner_handle *oh; nlm_rpc_t *rpcp = NULL; vnode_t *vp = NULL; struct netbuf *addr; char *netid; char *name; int error; struct flock64 fl; nlm_copy_netobj(&resp->cookie, &argp->cookie); name = argp->alock.caller_name; netid = svc_getnetid(sr->rq_xprt); addr = svc_getrpccaller(sr->rq_xprt); g = zone_getspecific(nlm_zone_key, curzone); host = nlm_host_findcreate(g, name, netid, addr); if (host == NULL) { resp->stat.stat = nlm4_denied_nolocks; return; } if (cb != NULL) { error = nlm_host_get_rpc(host, sr->rq_vers, &rpcp); if (error != 0) { resp->stat.stat = nlm4_denied_nolocks; goto out; } } vp = nlm_fh_to_vp(&argp->alock.fh); if (vp == NULL) { resp->stat.stat = nlm4_stale_fh; goto out; } if (NLM_IN_GRACE(g)) { resp->stat.stat = nlm4_denied_grace_period; goto out; } /* Convert to local form. */ error = nlm_init_flock(&fl, &argp->alock, host, sr->rq_vers, (argp->exclusive) ? F_WRLCK : F_RDLCK); if (error) { resp->stat.stat = nlm4_failed; goto out; } /* BSD: VOP_ADVLOCK(nv->nv_vp, NULL, F_GETLK, &fl, F_REMOTE); */ error = nlm_vop_frlock(vp, F_GETLK, &fl, F_REMOTELOCK | FREAD | FWRITE, (u_offset_t)0, NULL, CRED(), NULL); if (error) { resp->stat.stat = nlm4_failed; goto out; } if (fl.l_type == F_UNLCK) { resp->stat.stat = nlm4_granted; goto out; } resp->stat.stat = nlm4_denied; /* * This lock "test" fails due to a conflicting lock. * * If this is a v1 client, make sure the conflicting * lock range we report can be expressed with 32-bit * offsets. The lock range requested was expressed * as 32-bit offset and length, so at least part of * the conflicting lock should lie below MAX_UOFF32. * If the conflicting lock extends past that, we'll * trim the range to end at MAX_UOFF32 so this lock * can be represented in a 32-bit response. Check * the start also (paranoid, but a low cost check). */ if (sr->rq_vers < NLM4_VERS) { uint64 maxlen; if (fl.l_start > MAX_UOFF32) fl.l_start = MAX_UOFF32; maxlen = MAX_UOFF32 + 1 - fl.l_start; if (fl.l_len > maxlen) fl.l_len = maxlen; } /* * Build the nlm4_holder result structure. * * Note that lh->oh is freed via xdr_free, * xdr_nlm4_holder, xdr_netobj, xdr_bytes. */ oh = kmem_zalloc(sizeof (*oh), KM_SLEEP); oh->oh_sysid = (sysid_t)fl.l_sysid; lh = &resp->stat.nlm4_testrply_u.holder; lh->exclusive = (fl.l_type == F_WRLCK); lh->svid = fl.l_pid; lh->oh.n_len = sizeof (*oh); lh->oh.n_bytes = (void *)oh; lh->l_offset = fl.l_start; lh->l_len = fl.l_len; out: /* * If we have a callback function, use that to * deliver the response via another RPC call. */ if (cb != NULL && rpcp != NULL) NLM_INVOKE_CALLBACK("test", rpcp, resp, cb); if (vp != NULL) VN_RELE(vp); if (rpcp != NULL) nlm_host_rele_rpc(host, rpcp); nlm_host_release(g, host); } /* * NLM_LOCK, NLM_LOCK_MSG, NLM_NM_LOCK * NLM4_LOCK, NLM4_LOCK_MSG, NLM4_NM_LOCK * * Client request to set a lock, possibly blocking. * * If the lock needs to block, we return status blocked to * this RPC call, and then later call back the client with * a "granted" callback. Tricky aspects of this include: * sending a reply before this function returns, and then * borrowing this thread from the RPC service pool for the * wait on the lock and doing the later granted callback. * * We also have to keep a list of locks (pending + granted) * both to handle retransmitted requests, and to keep the * vnodes for those locks active. */ void nlm_do_lock(nlm4_lockargs *argp, nlm4_res *resp, struct svc_req *sr, nlm_reply_cb reply_cb, nlm_res_cb res_cb, nlm_testargs_cb grant_cb) { struct nlm_globals *g; struct flock64 fl; struct nlm_host *host = NULL; struct netbuf *addr; struct nlm_vhold *nvp = NULL; nlm_rpc_t *rpcp = NULL; char *netid; char *name; int error, flags; bool_t do_blocking = FALSE; bool_t do_mon_req = FALSE; enum nlm4_stats status; nlm_copy_netobj(&resp->cookie, &argp->cookie); name = argp->alock.caller_name; netid = svc_getnetid(sr->rq_xprt); addr = svc_getrpccaller(sr->rq_xprt); g = zone_getspecific(nlm_zone_key, curzone); host = nlm_host_findcreate(g, name, netid, addr); if (host == NULL) { DTRACE_PROBE4(no__host, struct nlm_globals *, g, char *, name, char *, netid, struct netbuf *, addr); status = nlm4_denied_nolocks; goto doreply; } DTRACE_PROBE3(start, struct nlm_globals *, g, struct nlm_host *, host, nlm4_lockargs *, argp); /* * If we may need to do _msg_ call needing an RPC * callback, get the RPC client handle now, * so we know if we can bind to the NLM service on * this client. * * Note: host object carries transport type. * One client using multiple transports gets * separate sysids for each of its transports. */ if (res_cb != NULL || (grant_cb != NULL && argp->block == TRUE)) { error = nlm_host_get_rpc(host, sr->rq_vers, &rpcp); if (error != 0) { status = nlm4_denied_nolocks; goto doreply; } } /* * During the "grace period", only allow reclaim. */ if (argp->reclaim == 0 && NLM_IN_GRACE(g)) { status = nlm4_denied_grace_period; goto doreply; } /* * Check whether we missed host shutdown event */ if (nlm_host_get_state(host) != argp->state) nlm_host_notify_server(host, argp->state); /* * Get a hold on the vnode for a lock operation. * Only lock() and share() need vhold objects. */ nvp = nlm_fh_to_vhold(host, &argp->alock.fh); if (nvp == NULL) { status = nlm4_stale_fh; goto doreply; } /* Convert to local form. */ error = nlm_init_flock(&fl, &argp->alock, host, sr->rq_vers, (argp->exclusive) ? F_WRLCK : F_RDLCK); if (error) { status = nlm4_failed; goto doreply; } /* * Try to lock non-blocking first. If we succeed * getting the lock, we can reply with the granted * status directly and avoid the complications of * making the "granted" RPC callback later. * * This also let's us find out now about some * possible errors like EROFS, etc. */ flags = F_REMOTELOCK | FREAD | FWRITE; error = nlm_vop_frlock(nvp->nv_vp, F_SETLK, &fl, flags, (u_offset_t)0, NULL, CRED(), NULL); DTRACE_PROBE3(setlk__res, struct flock64 *, &fl, int, flags, int, error); switch (error) { case 0: /* Got it without waiting! */ status = nlm4_granted; do_mon_req = TRUE; break; /* EINPROGRESS too? */ case EAGAIN: /* We did not get the lock. Should we block? */ if (argp->block == FALSE || grant_cb == NULL) { status = nlm4_denied; break; } /* * Should block. Try to reserve this thread * so we can use it to wait for the lock and * later send the granted message. If this * reservation fails, say "no resources". */ if (!svc_reserve_thread(sr->rq_xprt)) { status = nlm4_denied_nolocks; break; } /* * OK, can detach this thread, so this call * will block below (after we reply). */ status = nlm4_blocked; do_blocking = TRUE; do_mon_req = TRUE; break; case ENOLCK: /* Failed for lack of resources. */ status = nlm4_denied_nolocks; break; case EROFS: /* read-only file system */ status = nlm4_rofs; break; case EFBIG: /* file too big */ status = nlm4_fbig; break; case EDEADLK: /* dead lock condition */ status = nlm4_deadlck; break; default: status = nlm4_denied; break; } doreply: resp->stat.stat = status; /* * We get one of two function pointers; one for a * normal RPC reply, and another for doing an RPC * "callback" _res reply for a _msg function. * Use either of those to send the reply now. * * If sending this reply fails, just leave the * lock in the list for retransmitted requests. * Cleanup is via unlock or host rele (statmon). */ if (reply_cb != NULL) { /* i.e. nlm_lock_1_reply */ if (!(*reply_cb)(sr->rq_xprt, resp)) svcerr_systemerr(sr->rq_xprt); } if (res_cb != NULL && rpcp != NULL) NLM_INVOKE_CALLBACK("lock", rpcp, resp, res_cb); /* * The reply has been sent to the client. * Start monitoring this client (maybe). * * Note that the non-monitored (NM) calls pass grant_cb=NULL * indicating that the client doesn't support RPC callbacks. * No monitoring for these (lame) clients. */ if (do_mon_req && grant_cb != NULL) nlm_host_monitor(g, host, argp->state); if (do_blocking) { /* * We need to block on this lock, and when that * completes, do the granted RPC call. Note that * we "reserved" this thread above, so we can now * "detach" it from the RPC SVC pool, allowing it * to block indefinitely if needed. */ ASSERT(rpcp != NULL); (void) svc_detach_thread(sr->rq_xprt); nlm_block(argp, host, nvp, rpcp, &fl, grant_cb); } DTRACE_PROBE3(lock__end, struct nlm_globals *, g, struct nlm_host *, host, nlm4_res *, resp); if (rpcp != NULL) nlm_host_rele_rpc(host, rpcp); nlm_vhold_release(host, nvp); nlm_host_release(g, host); } /* * Helper for nlm_do_lock(), partly for observability, * (we'll see a call blocked in this function) and * because nlm_do_lock() was getting quite long. */ static void nlm_block(nlm4_lockargs *lockargs, struct nlm_host *host, struct nlm_vhold *nvp, nlm_rpc_t *rpcp, struct flock64 *flp, nlm_testargs_cb grant_cb) { nlm4_testargs args; int error; flk_callback_t flk_cb; struct nlm_block_cb_data cb_data; /* * Keep a list of blocked locks on nh_pending, and use it * to cancel these threads in nlm_destroy_client_pending. * * Check to see if this lock is already in the list * and if not, add an entry for it. Allocate first, * then if we don't insert, free the new one. * Caller already has vp held. */ error = nlm_slreq_register(host, nvp, flp); if (error != 0) { /* * Sleeping lock request with given fl is already * registered by someone else. This means that * some other thread is handling the request, let * him to do its work. */ ASSERT(error == EEXIST); return; } cb_data.hostp = host; cb_data.nvp = nvp; cb_data.flp = flp; flk_init_callback(&flk_cb, nlm_block_callback, &cb_data); /* BSD: VOP_ADVLOCK(vp, NULL, F_SETLK, fl, F_REMOTE); */ error = nlm_vop_frlock(nvp->nv_vp, F_SETLKW, flp, F_REMOTELOCK | FREAD | FWRITE, (u_offset_t)0, &flk_cb, CRED(), NULL); if (error != 0) { /* * We failed getting the lock, but have no way to * tell the client about that. Let 'em time out. */ (void) nlm_slreq_unregister(host, nvp, flp); return; } /* * Do the "granted" call-back to the client. */ args.cookie = lockargs->cookie; args.exclusive = lockargs->exclusive; args.alock = lockargs->alock; NLM_INVOKE_CALLBACK("grant", rpcp, &args, grant_cb); } /* * The function that is used as flk callback when NLM server * sets new sleeping lock. The function unregisters NLM * sleeping lock request (nlm_slreq) associated with the * sleeping lock _before_ lock becomes active. It prevents * potential race condition between nlm_block() and * nlm_do_cancel(). */ static callb_cpr_t * nlm_block_callback(flk_cb_when_t when, void *data) { struct nlm_block_cb_data *cb_data; cb_data = (struct nlm_block_cb_data *)data; if (when == FLK_AFTER_SLEEP) { (void) nlm_slreq_unregister(cb_data->hostp, cb_data->nvp, cb_data->flp); } return (0); } /* * NLM_CANCEL, NLM_CANCEL_MSG, * NLM4_CANCEL, NLM4_CANCEL_MSG, * Client gives up waiting for a blocking lock. */ void nlm_do_cancel(nlm4_cancargs *argp, nlm4_res *resp, struct svc_req *sr, nlm_res_cb cb) { struct nlm_globals *g; struct nlm_host *host; struct netbuf *addr; struct nlm_vhold *nvp = NULL; nlm_rpc_t *rpcp = NULL; char *netid; char *name; int error; struct flock64 fl; nlm_copy_netobj(&resp->cookie, &argp->cookie); netid = svc_getnetid(sr->rq_xprt); addr = svc_getrpccaller(sr->rq_xprt); name = argp->alock.caller_name; g = zone_getspecific(nlm_zone_key, curzone); host = nlm_host_findcreate(g, name, netid, addr); if (host == NULL) { resp->stat.stat = nlm4_denied_nolocks; return; } if (cb != NULL) { error = nlm_host_get_rpc(host, sr->rq_vers, &rpcp); if (error != 0) { resp->stat.stat = nlm4_denied_nolocks; return; } } DTRACE_PROBE3(start, struct nlm_globals *, g, struct nlm_host *, host, nlm4_cancargs *, argp); if (NLM_IN_GRACE(g)) { resp->stat.stat = nlm4_denied_grace_period; goto out; } nvp = nlm_fh_to_vhold(host, &argp->alock.fh); if (nvp == NULL) { resp->stat.stat = nlm4_stale_fh; goto out; } /* Convert to local form. */ error = nlm_init_flock(&fl, &argp->alock, host, sr->rq_vers, (argp->exclusive) ? F_WRLCK : F_RDLCK); if (error) { resp->stat.stat = nlm4_failed; goto out; } error = nlm_slreq_unregister(host, nvp, &fl); if (error != 0) { /* * There's no sleeping lock request corresponding * to the lock. Then requested sleeping lock * doesn't exist. */ resp->stat.stat = nlm4_denied; goto out; } fl.l_type = F_UNLCK; error = nlm_vop_frlock(nvp->nv_vp, F_SETLK, &fl, F_REMOTELOCK | FREAD | FWRITE, (u_offset_t)0, NULL, CRED(), NULL); resp->stat.stat = (error == 0) ? nlm4_granted : nlm4_denied; out: /* * If we have a callback function, use that to * deliver the response via another RPC call. */ if (cb != NULL && rpcp != NULL) NLM_INVOKE_CALLBACK("cancel", rpcp, resp, cb); DTRACE_PROBE3(cancel__end, struct nlm_globals *, g, struct nlm_host *, host, nlm4_res *, resp); if (rpcp != NULL) nlm_host_rele_rpc(host, rpcp); nlm_vhold_release(host, nvp); nlm_host_release(g, host); } /* * NLM_UNLOCK, NLM_UNLOCK_MSG, * NLM4_UNLOCK, NLM4_UNLOCK_MSG, * Client removes one of their locks. */ void nlm_do_unlock(nlm4_unlockargs *argp, nlm4_res *resp, struct svc_req *sr, nlm_res_cb cb) { struct nlm_globals *g; struct nlm_host *host; struct netbuf *addr; nlm_rpc_t *rpcp = NULL; vnode_t *vp = NULL; char *netid; char *name; int error; struct flock64 fl; nlm_copy_netobj(&resp->cookie, &argp->cookie); netid = svc_getnetid(sr->rq_xprt); addr = svc_getrpccaller(sr->rq_xprt); name = argp->alock.caller_name; /* * NLM_UNLOCK operation doesn't have an error code * denoting that operation failed, so we always * return nlm4_granted except when the server is * in a grace period. */ resp->stat.stat = nlm4_granted; g = zone_getspecific(nlm_zone_key, curzone); host = nlm_host_findcreate(g, name, netid, addr); if (host == NULL) return; if (cb != NULL) { error = nlm_host_get_rpc(host, sr->rq_vers, &rpcp); if (error != 0) goto out; } DTRACE_PROBE3(start, struct nlm_globals *, g, struct nlm_host *, host, nlm4_unlockargs *, argp); if (NLM_IN_GRACE(g)) { resp->stat.stat = nlm4_denied_grace_period; goto out; } vp = nlm_fh_to_vp(&argp->alock.fh); if (vp == NULL) goto out; /* Convert to local form. */ error = nlm_init_flock(&fl, &argp->alock, host, sr->rq_vers, F_UNLCK); if (error) goto out; /* BSD: VOP_ADVLOCK(nv->nv_vp, NULL, F_UNLCK, &fl, F_REMOTE); */ error = nlm_vop_frlock(vp, F_SETLK, &fl, F_REMOTELOCK | FREAD | FWRITE, (u_offset_t)0, NULL, CRED(), NULL); DTRACE_PROBE1(unlock__res, int, error); out: /* * If we have a callback function, use that to * deliver the response via another RPC call. */ if (cb != NULL && rpcp != NULL) NLM_INVOKE_CALLBACK("unlock", rpcp, resp, cb); DTRACE_PROBE3(unlock__end, struct nlm_globals *, g, struct nlm_host *, host, nlm4_res *, resp); if (vp != NULL) VN_RELE(vp); if (rpcp != NULL) nlm_host_rele_rpc(host, rpcp); nlm_host_release(g, host); } /* * NLM_GRANTED, NLM_GRANTED_MSG, * NLM4_GRANTED, NLM4_GRANTED_MSG, * * This service routine is special. It's the only one that's * really part of our NLM _client_ support, used by _servers_ * to "call back" when a blocking lock from this NLM client * is granted by the server. In this case, we _know_ there is * already an nlm_host allocated and held by the client code. * We want to find that nlm_host here. * * Over in nlm_call_lock(), the client encoded the sysid for this * server in the "owner handle" netbuf sent with our lock request. * We can now use that to find the nlm_host object we used there. * (NB: The owner handle is opaque to the server.) */ void nlm_do_granted(nlm4_testargs *argp, nlm4_res *resp, struct svc_req *sr, nlm_res_cb cb) { struct nlm_globals *g; struct nlm_owner_handle *oh; struct nlm_host *host; nlm_rpc_t *rpcp = NULL; int error; nlm_copy_netobj(&resp->cookie, &argp->cookie); resp->stat.stat = nlm4_denied; g = zone_getspecific(nlm_zone_key, curzone); oh = (void *) argp->alock.oh.n_bytes; if (oh == NULL) return; host = nlm_host_find_by_sysid(g, oh->oh_sysid); if (host == NULL) return; if (cb != NULL) { error = nlm_host_get_rpc(host, sr->rq_vers, &rpcp); if (error != 0) goto out; } if (NLM_IN_GRACE(g)) { resp->stat.stat = nlm4_denied_grace_period; goto out; } error = nlm_slock_grant(g, host, &argp->alock); if (error == 0) resp->stat.stat = nlm4_granted; out: /* * If we have a callback function, use that to * deliver the response via another RPC call. */ if (cb != NULL && rpcp != NULL) NLM_INVOKE_CALLBACK("do_granted", rpcp, resp, cb); if (rpcp != NULL) nlm_host_rele_rpc(host, rpcp); nlm_host_release(g, host); } /* * NLM_FREE_ALL, NLM4_FREE_ALL * * Destroy all lock state for the calling client. */ void nlm_do_free_all(nlm4_notify *argp, void *res, struct svc_req *sr) { struct nlm_globals *g; struct nlm_host_list host_list; struct nlm_host *hostp; TAILQ_INIT(&host_list); g = zone_getspecific(nlm_zone_key, curzone); /* Serialize calls to clean locks. */ mutex_enter(&g->clean_lock); /* * Find all hosts that have the given node name and put them on a * local list. */ mutex_enter(&g->lock); for (hostp = avl_first(&g->nlm_hosts_tree); hostp != NULL; hostp = AVL_NEXT(&g->nlm_hosts_tree, hostp)) { if (strcasecmp(hostp->nh_name, argp->name) == 0) { /* * If needed take the host out of the idle list since * we are taking a reference. */ if (hostp->nh_flags & NLM_NH_INIDLE) { TAILQ_REMOVE(&g->nlm_idle_hosts, hostp, nh_link); hostp->nh_flags &= ~NLM_NH_INIDLE; } hostp->nh_refs++; TAILQ_INSERT_TAIL(&host_list, hostp, nh_link); } } mutex_exit(&g->lock); /* Free locks for all hosts on the local list. */ while (!TAILQ_EMPTY(&host_list)) { hostp = TAILQ_FIRST(&host_list); TAILQ_REMOVE(&host_list, hostp, nh_link); /* * Note that this does not do client-side cleanup. * We want to do that ONLY if statd tells us the * server has restarted. */ nlm_host_notify_server(hostp, argp->state); nlm_host_release(g, hostp); } mutex_exit(&g->clean_lock); (void) res; (void) sr; } static void nlm_init_shrlock(struct shrlock *shr, nlm4_share *nshare, struct nlm_host *host) { switch (nshare->access) { default: case fsa_NONE: shr->s_access = 0; break; case fsa_R: shr->s_access = F_RDACC; break; case fsa_W: shr->s_access = F_WRACC; break; case fsa_RW: shr->s_access = F_RWACC; break; } switch (nshare->mode) { default: case fsm_DN: shr->s_deny = F_NODNY; break; case fsm_DR: shr->s_deny = F_RDDNY; break; case fsm_DW: shr->s_deny = F_WRDNY; break; case fsm_DRW: shr->s_deny = F_RWDNY; break; } shr->s_sysid = host->nh_sysid; shr->s_pid = 0; shr->s_own_len = nshare->oh.n_len; shr->s_owner = nshare->oh.n_bytes; } /* * NLM_SHARE, NLM4_SHARE * * Request a DOS-style share reservation */ void nlm_do_share(nlm4_shareargs *argp, nlm4_shareres *resp, struct svc_req *sr) { struct nlm_globals *g; struct nlm_host *host; struct netbuf *addr; struct nlm_vhold *nvp = NULL; char *netid; char *name; int error; struct shrlock shr; nlm_copy_netobj(&resp->cookie, &argp->cookie); name = argp->share.caller_name; netid = svc_getnetid(sr->rq_xprt); addr = svc_getrpccaller(sr->rq_xprt); g = zone_getspecific(nlm_zone_key, curzone); host = nlm_host_findcreate(g, name, netid, addr); if (host == NULL) { resp->stat = nlm4_denied_nolocks; return; } DTRACE_PROBE3(share__start, struct nlm_globals *, g, struct nlm_host *, host, nlm4_shareargs *, argp); if (argp->reclaim == 0 && NLM_IN_GRACE(g)) { resp->stat = nlm4_denied_grace_period; goto out; } /* * Get holded vnode when on lock operation. * Only lock() and share() need vhold objects. */ nvp = nlm_fh_to_vhold(host, &argp->share.fh); if (nvp == NULL) { resp->stat = nlm4_stale_fh; goto out; } /* Convert to local form. */ nlm_init_shrlock(&shr, &argp->share, host); error = VOP_SHRLOCK(nvp->nv_vp, F_SHARE, &shr, FREAD | FWRITE, CRED(), NULL); if (error == 0) { resp->stat = nlm4_granted; nlm_host_monitor(g, host, 0); } else { resp->stat = nlm4_denied; } out: DTRACE_PROBE3(share__end, struct nlm_globals *, g, struct nlm_host *, host, nlm4_shareres *, resp); nlm_vhold_release(host, nvp); nlm_host_release(g, host); } /* * NLM_UNSHARE, NLM4_UNSHARE * * Release a DOS-style share reservation */ void nlm_do_unshare(nlm4_shareargs *argp, nlm4_shareres *resp, struct svc_req *sr) { struct nlm_globals *g; struct nlm_host *host; struct netbuf *addr; vnode_t *vp = NULL; char *netid; int error; struct shrlock shr; nlm_copy_netobj(&resp->cookie, &argp->cookie); netid = svc_getnetid(sr->rq_xprt); addr = svc_getrpccaller(sr->rq_xprt); g = zone_getspecific(nlm_zone_key, curzone); host = nlm_host_find(g, netid, addr); if (host == NULL) { resp->stat = nlm4_denied_nolocks; return; } DTRACE_PROBE3(unshare__start, struct nlm_globals *, g, struct nlm_host *, host, nlm4_shareargs *, argp); if (NLM_IN_GRACE(g)) { resp->stat = nlm4_denied_grace_period; goto out; } vp = nlm_fh_to_vp(&argp->share.fh); if (vp == NULL) { resp->stat = nlm4_stale_fh; goto out; } /* Convert to local form. */ nlm_init_shrlock(&shr, &argp->share, host); error = VOP_SHRLOCK(vp, F_UNSHARE, &shr, FREAD | FWRITE, CRED(), NULL); (void) error; resp->stat = nlm4_granted; out: DTRACE_PROBE3(unshare__end, struct nlm_globals *, g, struct nlm_host *, host, nlm4_shareres *, resp); if (vp != NULL) VN_RELE(vp); nlm_host_release(g, host); } /* * NLM wrapper to VOP_FRLOCK that checks the validity of the lock before * invoking the vnode operation. */ static int nlm_vop_frlock(vnode_t *vp, int cmd, flock64_t *bfp, int flag, offset_t offset, struct flk_callback *flk_cbp, cred_t *cr, caller_context_t *ct) { if (bfp->l_len != 0 && bfp->l_start + (bfp->l_len - 1) < bfp->l_start) { return (EOVERFLOW); } return (VOP_FRLOCK(vp, cmd, bfp, flag, offset, flk_cbp, cr, ct)); }