/* * CDDL HEADER START * * The contents of this file are subject to the terms of the * Common Development and Distribution License (the "License"). * You may not use this file except in compliance with the License. * * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE * or http://www.opensolaris.org/os/licensing. * See the License for the specific language governing permissions * and limitations under the License. * * When distributing Covered Code, include this CDDL HEADER in each * file and include the License file at usr/src/OPENSOLARIS.LICENSE. * If applicable, add the following below this CDDL HEADER, with the * fields enclosed by brackets "[]" replaced with your own identifying * information: Portions Copyright [yyyy] [name of copyright owner] * * CDDL HEADER END */ /* * Copyright 2009 Sun Microsystems, Inc. All rights reserved. * Use is subject to license terms. */ /* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */ /* All Rights Reserved */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #ifdef DEBUG /* * These are "special" state IDs and file handles that * match any delegation state ID or file handled. This * is for testing purposes only. */ stateid4 nfs4_deleg_any = { 0x7FFFFFF0 }; char nfs4_deleg_fh[] = "\0377\0376\0375\0374"; nfs_fh4 nfs4_deleg_anyfh = { sizeof (nfs4_deleg_fh)-1, nfs4_deleg_fh }; nfsstat4 cb4_getattr_fail = NFS4_OK; nfsstat4 cb4_recall_fail = NFS4_OK; int nfs4_callback_debug; int nfs4_recall_debug; int nfs4_drat_debug; #endif #define CB_NOTE(x) NFS4_DEBUG(nfs4_callback_debug, (CE_NOTE, x)) #define CB_WARN(x) NFS4_DEBUG(nfs4_callback_debug, (CE_WARN, x)) #define CB_WARN1(x, y) NFS4_DEBUG(nfs4_callback_debug, (CE_WARN, x, y)) enum nfs4_delegreturn_policy nfs4_delegreturn_policy = INACTIVE; static zone_key_t nfs4_callback_zone_key; /* * NFS4_MAPSIZE is the number of bytes we are willing to consume * for the block allocation map when the server grants a NFS_LIMIT_BLOCK * style delegation. */ #define NFS4_MAPSIZE 8192 #define NFS4_MAPWORDS NFS4_MAPSIZE/sizeof (uint_t) #define NbPW (NBBY*sizeof (uint_t)) static int nfs4_num_prognums = 1024; static SVC_CALLOUT_TABLE nfs4_cb_sct; struct nfs4_dnode { list_node_t linkage; rnode4_t *rnodep; int flags; /* Flags for nfs4delegreturn_impl() */ }; static const struct nfs4_callback_stats nfs4_callback_stats_tmpl = { { "delegations", KSTAT_DATA_UINT64 }, { "cb_getattr", KSTAT_DATA_UINT64 }, { "cb_recall", KSTAT_DATA_UINT64 }, { "cb_null", KSTAT_DATA_UINT64 }, { "cb_dispatch", KSTAT_DATA_UINT64 }, { "delegaccept_r", KSTAT_DATA_UINT64 }, { "delegaccept_rw", KSTAT_DATA_UINT64 }, { "delegreturn", KSTAT_DATA_UINT64 }, { "callbacks", KSTAT_DATA_UINT64 }, { "claim_cur", KSTAT_DATA_UINT64 }, { "claim_cur_ok", KSTAT_DATA_UINT64 }, { "recall_trunc", KSTAT_DATA_UINT64 }, { "recall_failed", KSTAT_DATA_UINT64 }, { "return_limit_write", KSTAT_DATA_UINT64 }, { "return_limit_addmap", KSTAT_DATA_UINT64 }, { "deleg_recover", KSTAT_DATA_UINT64 }, { "cb_illegal", KSTAT_DATA_UINT64 } }; struct nfs4_cb_port { list_node_t linkage; /* linkage into per-zone port list */ char netid[KNC_STRSIZE]; char uaddr[KNC_STRSIZE]; char protofmly[KNC_STRSIZE]; char proto[KNC_STRSIZE]; }; static int cb_getattr_bytes; struct cb_recall_pass { rnode4_t *rp; int flags; /* Flags for nfs4delegreturn_impl() */ bool_t truncate; }; static nfs4_open_stream_t *get_next_deleg_stream(rnode4_t *, int); static void nfs4delegreturn_thread(struct cb_recall_pass *); static int deleg_reopen(vnode_t *, bool_t *, struct nfs4_callback_globals *, int); static void nfs4_dlistadd(rnode4_t *, struct nfs4_callback_globals *, int); static void nfs4_dlistclean_impl(struct nfs4_callback_globals *, int); static int nfs4delegreturn_impl(rnode4_t *, int, struct nfs4_callback_globals *); static void nfs4delegreturn_cleanup_impl(rnode4_t *, nfs4_server_t *, struct nfs4_callback_globals *); static void nfs4_recall_sync_wait(nfs4_server_t *); static void cb_getattr(nfs_cb_argop4 *argop, nfs_cb_resop4 *resop, struct svc_req *req, struct compound_state *cs, struct nfs4_callback_globals *ncg) { CB_GETATTR4args *args = &argop->nfs_cb_argop4_u.opcbgetattr; CB_GETATTR4res *resp = &resop->nfs_cb_resop4_u.opcbgetattr; rnode4_t *rp; vnode_t *vp; bool_t found = FALSE; struct nfs4_server *sp; struct fattr4 *fap; rpc_inline_t *fdata; long mapcnt; fattr4_change change; fattr4_size size; uint_t rflag; ncg->nfs4_callback_stats.cb_getattr.value.ui64++; #ifdef DEBUG /* * error injection hook: set cb_getattr_fail global to * NFS4 pcol error to be returned */ if (cb4_getattr_fail != NFS4_OK) { *cs->statusp = resp->status = cb4_getattr_fail; return; } #endif resp->obj_attributes.attrmask = 0; mutex_enter(&ncg->nfs4_cb_lock); sp = ncg->nfs4prog2server[req->rq_prog - NFS4_CALLBACK]; mutex_exit(&ncg->nfs4_cb_lock); if (nfs4_server_vlock(sp, 0) == FALSE) { CB_WARN("cb_getattr: cannot find server\n"); *cs->statusp = resp->status = NFS4ERR_BADHANDLE; return; } /* * In cb_compound, callback_ident was validated against rq_prog, * but we couldn't verify that it was set to the value we provided * at setclientid time (because we didn't have server struct yet). * Now we have the server struct, but don't have callback_ident * handy. So, validate server struct program number against req * RPC's prog number. At this point, we know the RPC prog num * is valid (else we wouldn't be here); however, we don't know * that it was the prog number we supplied to this server at * setclientid time. If the prog numbers aren't equivalent, then * log the problem and fail the request because either cbserv * and/or cbclient are confused. This will probably never happen. */ if (sp->s_program != req->rq_prog) { #ifdef DEBUG zcmn_err(getzoneid(), CE_WARN, "cb_getattr: wrong server program number srv=%d req=%d\n", sp->s_program, req->rq_prog); #else zcmn_err(getzoneid(), CE_WARN, "cb_getattr: wrong server program number\n"); #endif mutex_exit(&sp->s_lock); nfs4_server_rele(sp); *cs->statusp = resp->status = NFS4ERR_BADHANDLE; return; } /* * Search the delegation list for a matching file handle; * mutex on sp prevents the list from changing. */ rp = list_head(&sp->s_deleg_list); for (; rp != NULL; rp = list_next(&sp->s_deleg_list, rp)) { nfs4_fhandle_t fhandle; sfh4_copyval(rp->r_fh, &fhandle); if ((fhandle.fh_len == args->fh.nfs_fh4_len && bcmp(fhandle.fh_buf, args->fh.nfs_fh4_val, fhandle.fh_len) == 0)) { found = TRUE; break; } #ifdef DEBUG if (nfs4_deleg_anyfh.nfs_fh4_len == args->fh.nfs_fh4_len && bcmp(nfs4_deleg_anyfh.nfs_fh4_val, args->fh.nfs_fh4_val, args->fh.nfs_fh4_len) == 0) { found = TRUE; break; } #endif } /* * VN_HOLD the vnode before releasing s_lock to guarantee * we have a valid vnode reference. */ if (found == TRUE) { vp = RTOV4(rp); VN_HOLD(vp); } mutex_exit(&sp->s_lock); nfs4_server_rele(sp); if (found == FALSE) { CB_WARN("cb_getattr: bad fhandle\n"); *cs->statusp = resp->status = NFS4ERR_BADHANDLE; return; } /* * Figure out which attributes the server wants. We only * offer FATTR4_CHANGE & FATTR4_SIZE; ignore the rest. */ fdata = kmem_alloc(cb_getattr_bytes, KM_SLEEP); /* * Don't actually need to create XDR to encode these * simple data structures. * xdrmem_create(&xdr, fdata, cb_getattr_bytes, XDR_ENCODE); */ fap = &resp->obj_attributes; fap->attrmask = 0; /* attrlist4_len starts at 0 and increases as attrs are processed */ fap->attrlist4 = (char *)fdata; fap->attrlist4_len = 0; /* don't supply attrs if request was zero */ if (args->attr_request != 0) { if (args->attr_request & FATTR4_CHANGE_MASK) { /* * If the file is mmapped, then increment the change * attribute and return it. This will guarantee that * the server will perceive that the file has changed * if there is any chance that the client application * has changed it. Otherwise, just return the change * attribute as it has been updated by nfs4write_deleg. */ mutex_enter(&rp->r_statelock); mapcnt = rp->r_mapcnt; rflag = rp->r_flags; mutex_exit(&rp->r_statelock); mutex_enter(&rp->r_statev4_lock); /* * If object mapped, then always return new change. * Otherwise, return change if object has dirty * pages. If object doesn't have any dirty pages, * then all changes have been pushed to server, so * reset change to grant change. */ if (mapcnt) rp->r_deleg_change++; else if (! (rflag & R4DIRTY)) rp->r_deleg_change = rp->r_deleg_change_grant; change = rp->r_deleg_change; mutex_exit(&rp->r_statev4_lock); /* * Use inline XDR code directly, we know that we * going to a memory buffer and it has enough * space so it cannot fail. */ IXDR_PUT_U_HYPER(fdata, change); fap->attrlist4_len += 2 * BYTES_PER_XDR_UNIT; fap->attrmask |= FATTR4_CHANGE_MASK; } if (args->attr_request & FATTR4_SIZE_MASK) { /* * Use an atomic add of 0 to fetch a consistent view * of r_size; this avoids having to take rw_lock * which could cause a deadlock. */ size = atomic_add_64_nv((uint64_t *)&rp->r_size, 0); /* * Use inline XDR code directly, we know that we * going to a memory buffer and it has enough * space so it cannot fail. */ IXDR_PUT_U_HYPER(fdata, size); fap->attrlist4_len += 2 * BYTES_PER_XDR_UNIT; fap->attrmask |= FATTR4_SIZE_MASK; } } VN_RELE(vp); *cs->statusp = resp->status = NFS4_OK; } static void cb_getattr_free(nfs_cb_resop4 *resop) { if (resop->nfs_cb_resop4_u.opcbgetattr.obj_attributes.attrlist4) kmem_free(resop->nfs_cb_resop4_u.opcbgetattr. obj_attributes.attrlist4, cb_getattr_bytes); } static void cb_recall(nfs_cb_argop4 *argop, nfs_cb_resop4 *resop, struct svc_req *req, struct compound_state *cs, struct nfs4_callback_globals *ncg) { CB_RECALL4args * args = &argop->nfs_cb_argop4_u.opcbrecall; CB_RECALL4res *resp = &resop->nfs_cb_resop4_u.opcbrecall; rnode4_t *rp; vnode_t *vp; struct nfs4_server *sp; bool_t found = FALSE; bool_t retried = FALSE; ncg->nfs4_callback_stats.cb_recall.value.ui64++; ASSERT(req->rq_prog >= NFS4_CALLBACK); ASSERT(req->rq_prog < NFS4_CALLBACK+nfs4_num_prognums); #ifdef DEBUG /* * error injection hook: set cb_recall_fail global to * NFS4 pcol error to be returned */ if (cb4_recall_fail != NFS4_OK) { *cs->statusp = resp->status = cb4_recall_fail; return; } #endif mutex_enter(&ncg->nfs4_cb_lock); sp = ncg->nfs4prog2server[req->rq_prog - NFS4_CALLBACK]; mutex_exit(&ncg->nfs4_cb_lock); if (nfs4_server_vlock(sp, 0) == FALSE) { CB_WARN("cb_recall: cannot find server\n"); *cs->statusp = resp->status = NFS4ERR_BADHANDLE; return; } retry: /* * Search the delegation list for a matching file handle * AND stateid; mutex on sp prevents the list from changing. */ rp = list_head(&sp->s_deleg_list); for (; rp != NULL; rp = list_next(&sp->s_deleg_list, rp)) { mutex_enter(&rp->r_statev4_lock); /* check both state id and file handle! */ if ((bcmp(&rp->r_deleg_stateid, &args->stateid, sizeof (stateid4)) == 0)) { nfs4_fhandle_t fhandle; sfh4_copyval(rp->r_fh, &fhandle); if ((fhandle.fh_len == args->fh.nfs_fh4_len && bcmp(fhandle.fh_buf, args->fh.nfs_fh4_val, fhandle.fh_len) == 0)) { found = TRUE; break; } else { #ifdef DEBUG CB_WARN("cb_recall: stateid OK, bad fh"); #endif } } #ifdef DEBUG if (bcmp(&args->stateid, &nfs4_deleg_any, sizeof (stateid4)) == 0) { found = TRUE; break; } #endif mutex_exit(&rp->r_statev4_lock); } /* * VN_HOLD the vnode before releasing s_lock to guarantee * we have a valid vnode reference. The async thread will * release the hold when it's done. */ if (found == TRUE) { mutex_exit(&rp->r_statev4_lock); vp = RTOV4(rp); VN_HOLD(vp); } mutex_exit(&sp->s_lock); if (found == FALSE && retried == FALSE) { nfs4_recall_sync_wait(sp); mutex_enter(&sp->s_lock); retried = TRUE; goto retry; } nfs4_server_rele(sp); if (found == FALSE) { CB_WARN("cb_recall: bad stateid\n"); *cs->statusp = resp->status = NFS4ERR_BAD_STATEID; return; } /* Fire up a thread to do the delegreturn */ nfs4delegreturn_async(rp, NFS4_DR_RECALL|NFS4_DR_REOPEN, args->truncate); *cs->statusp = resp->status = 0; } /* ARGSUSED */ static void cb_recall_free(nfs_cb_resop4 *resop) { /* nothing to do here, cb_recall doesn't kmem_alloc */ } /* * This function handles the CB_NULL proc call from an NFSv4 Server. * * We take note that the server has sent a CB_NULL for later processing * in the recovery logic. It is noted so we may pause slightly after the * setclientid and before reopening files. The pause is to allow the * NFSv4 Server time to receive the CB_NULL reply and adjust any of * its internal structures such that it has the opportunity to grant * delegations to reopened files. * */ /* ARGSUSED */ static void cb_null(CB_COMPOUND4args *args, CB_COMPOUND4res *resp, struct svc_req *req, struct nfs4_callback_globals *ncg) { struct nfs4_server *sp; ncg->nfs4_callback_stats.cb_null.value.ui64++; ASSERT(req->rq_prog >= NFS4_CALLBACK); ASSERT(req->rq_prog < NFS4_CALLBACK+nfs4_num_prognums); mutex_enter(&ncg->nfs4_cb_lock); sp = ncg->nfs4prog2server[req->rq_prog - NFS4_CALLBACK]; mutex_exit(&ncg->nfs4_cb_lock); if (nfs4_server_vlock(sp, 0) != FALSE) { sp->s_flags |= N4S_CB_PINGED; cv_broadcast(&sp->wait_cb_null); mutex_exit(&sp->s_lock); nfs4_server_rele(sp); } } /* * cb_illegal args: void * res : status (NFS4ERR_OP_CB_ILLEGAL) */ /* ARGSUSED */ static void cb_illegal(nfs_cb_argop4 *argop, nfs_cb_resop4 *resop, struct svc_req *req, struct compound_state *cs, struct nfs4_callback_globals *ncg) { CB_ILLEGAL4res *resp = &resop->nfs_cb_resop4_u.opcbillegal; ncg->nfs4_callback_stats.cb_illegal.value.ui64++; resop->resop = OP_CB_ILLEGAL; *cs->statusp = resp->status = NFS4ERR_OP_ILLEGAL; } static void cb_compound(CB_COMPOUND4args *args, CB_COMPOUND4res *resp, struct svc_req *req, struct nfs4_callback_globals *ncg) { uint_t i; struct compound_state cs; nfs_cb_argop4 *argop; nfs_cb_resop4 *resop, *new_res; uint_t op; bzero(&cs, sizeof (cs)); cs.statusp = &resp->status; cs.cont = TRUE; /* * Form a reply tag by copying over the reqeuest tag. */ resp->tag.utf8string_len = args->tag.utf8string_len; resp->tag.utf8string_val = kmem_alloc(resp->tag.utf8string_len, KM_SLEEP); bcopy(args->tag.utf8string_val, resp->tag.utf8string_val, args->tag.utf8string_len); /* * XXX for now, minorversion should be zero */ if (args->minorversion != CB4_MINORVERSION) { resp->array_len = 0; resp->array = NULL; resp->status = NFS4ERR_MINOR_VERS_MISMATCH; return; } #ifdef DEBUG /* * Verify callback_ident. It doesn't really matter if it's wrong * because we don't really use callback_ident -- we use prog number * of the RPC request instead. In this case, just print a DEBUG * console message to reveal brokenness of cbclient (at bkoff/cthon). */ if (args->callback_ident != req->rq_prog) zcmn_err(getzoneid(), CE_WARN, "cb_compound: cb_client using wrong " "callback_ident(%d), should be %d", args->callback_ident, req->rq_prog); #endif resp->array_len = args->array_len; resp->array = kmem_zalloc(args->array_len * sizeof (nfs_cb_resop4), KM_SLEEP); for (i = 0; i < args->array_len && cs.cont; i++) { argop = &args->array[i]; resop = &resp->array[i]; resop->resop = argop->argop; op = (uint_t)resop->resop; switch (op) { case OP_CB_GETATTR: cb_getattr(argop, resop, req, &cs, ncg); break; case OP_CB_RECALL: cb_recall(argop, resop, req, &cs, ncg); break; case OP_CB_ILLEGAL: /* fall through */ default: /* * Handle OP_CB_ILLEGAL and any undefined opcode. * Currently, the XDR code will return BADXDR * if cb op doesn't decode to legal value, so * it really only handles OP_CB_ILLEGAL. */ op = OP_CB_ILLEGAL; cb_illegal(argop, resop, req, &cs, ncg); } if (*cs.statusp != NFS4_OK) cs.cont = FALSE; /* * If not at last op, and if we are to stop, then * compact the results array. */ if ((i + 1) < args->array_len && !cs.cont) { new_res = kmem_alloc( (i+1) * sizeof (nfs_cb_resop4), KM_SLEEP); bcopy(resp->array, new_res, (i+1) * sizeof (nfs_cb_resop4)); kmem_free(resp->array, args->array_len * sizeof (nfs_cb_resop4)); resp->array_len = i + 1; resp->array = new_res; } } } static void cb_compound_free(CB_COMPOUND4res *resp) { uint_t i, op; nfs_cb_resop4 *resop; if (resp->tag.utf8string_val) { UTF8STRING_FREE(resp->tag) } for (i = 0; i < resp->array_len; i++) { resop = &resp->array[i]; op = (uint_t)resop->resop; switch (op) { case OP_CB_GETATTR: cb_getattr_free(resop); break; case OP_CB_RECALL: cb_recall_free(resop); break; default: break; } } if (resp->array != NULL) { kmem_free(resp->array, resp->array_len * sizeof (nfs_cb_resop4)); } } static void cb_dispatch(struct svc_req *req, SVCXPRT *xprt) { CB_COMPOUND4args args; CB_COMPOUND4res res; struct nfs4_callback_globals *ncg; bool_t (*xdr_args)(), (*xdr_res)(); void (*proc)(CB_COMPOUND4args *, CB_COMPOUND4res *, struct svc_req *, struct nfs4_callback_globals *); void (*freeproc)(CB_COMPOUND4res *); ncg = zone_getspecific(nfs4_callback_zone_key, nfs_zone()); ASSERT(ncg != NULL); ncg->nfs4_callback_stats.cb_dispatch.value.ui64++; switch (req->rq_proc) { case CB_NULL: xdr_args = xdr_void; xdr_res = xdr_void; proc = cb_null; freeproc = NULL; break; case CB_COMPOUND: xdr_args = xdr_CB_COMPOUND4args_clnt; xdr_res = xdr_CB_COMPOUND4res; proc = cb_compound; freeproc = cb_compound_free; break; default: CB_WARN("cb_dispatch: no proc\n"); svcerr_noproc(xprt); return; } args.tag.utf8string_val = NULL; args.array = NULL; if (!SVC_GETARGS(xprt, xdr_args, (caddr_t)&args)) { CB_WARN("cb_dispatch: cannot getargs\n"); svcerr_decode(xprt); return; } (*proc)(&args, &res, req, ncg); if (svc_sendreply(xprt, xdr_res, (caddr_t)&res) == FALSE) { CB_WARN("cb_dispatch: bad sendreply\n"); svcerr_systemerr(xprt); } if (freeproc) (*freeproc)(&res); if (!SVC_FREEARGS(xprt, xdr_args, (caddr_t)&args)) { CB_WARN("cb_dispatch: bad freeargs\n"); } } static rpcprog_t nfs4_getnextprogram(struct nfs4_callback_globals *ncg) { int i, j; j = ncg->nfs4_program_hint; for (i = 0; i < nfs4_num_prognums; i++, j++) { if (j >= nfs4_num_prognums) j = 0; if (ncg->nfs4prog2server[j] == NULL) { ncg->nfs4_program_hint = j+1; return (j+NFS4_CALLBACK); } } return (0); } void nfs4callback_destroy(nfs4_server_t *np) { struct nfs4_callback_globals *ncg; int i; if (np->s_program == 0) return; ncg = np->zone_globals; i = np->s_program - NFS4_CALLBACK; mutex_enter(&ncg->nfs4_cb_lock); ASSERT(ncg->nfs4prog2server[i] == np); ncg->nfs4prog2server[i] = NULL; if (i < ncg->nfs4_program_hint) ncg->nfs4_program_hint = i; mutex_exit(&ncg->nfs4_cb_lock); } /* * nfs4_setport - This function saves a netid and univeral address for * the callback program. These values will be used during setclientid. */ static void nfs4_setport(char *netid, char *uaddr, char *protofmly, char *proto, struct nfs4_callback_globals *ncg) { struct nfs4_cb_port *p; bool_t found = FALSE; ASSERT(MUTEX_HELD(&ncg->nfs4_cb_lock)); p = list_head(&ncg->nfs4_cb_ports); for (; p != NULL; p = list_next(&ncg->nfs4_cb_ports, p)) { if (strcmp(p->netid, netid) == 0) { found = TRUE; break; } } if (found == TRUE) (void) strcpy(p->uaddr, uaddr); else { p = kmem_alloc(sizeof (*p), KM_SLEEP); (void) strcpy(p->uaddr, uaddr); (void) strcpy(p->netid, netid); (void) strcpy(p->protofmly, protofmly); (void) strcpy(p->proto, proto); list_insert_head(&ncg->nfs4_cb_ports, p); } } /* * nfs4_cb_args - This function is used to construct the callback * portion of the arguments needed for setclientid. */ void nfs4_cb_args(nfs4_server_t *np, struct knetconfig *knc, SETCLIENTID4args *args) { struct nfs4_cb_port *p; bool_t found = FALSE; rpcprog_t pgm; struct nfs4_callback_globals *ncg = np->zone_globals; /* * This server structure may already have a program number * assigned to it. This happens when the client has to * re-issue SETCLIENTID. Just re-use the information. */ if (np->s_program >= NFS4_CALLBACK && np->s_program < NFS4_CALLBACK + nfs4_num_prognums) nfs4callback_destroy(np); mutex_enter(&ncg->nfs4_cb_lock); p = list_head(&ncg->nfs4_cb_ports); for (; p != NULL; p = list_next(&ncg->nfs4_cb_ports, p)) { if (strcmp(p->protofmly, knc->knc_protofmly) == 0 && strcmp(p->proto, knc->knc_proto) == 0) { found = TRUE; break; } } if (found == FALSE) { NFS4_DEBUG(nfs4_callback_debug, (CE_WARN, "nfs4_cb_args: could not find netid for %s/%s\n", knc->knc_protofmly, knc->knc_proto)); args->callback.cb_program = 0; args->callback.cb_location.r_netid = NULL; args->callback.cb_location.r_addr = NULL; args->callback_ident = 0; mutex_exit(&ncg->nfs4_cb_lock); return; } if ((pgm = nfs4_getnextprogram(ncg)) == 0) { CB_WARN("nfs4_cb_args: out of program numbers\n"); args->callback.cb_program = 0; args->callback.cb_location.r_netid = NULL; args->callback.cb_location.r_addr = NULL; args->callback_ident = 0; mutex_exit(&ncg->nfs4_cb_lock); return; } ncg->nfs4prog2server[pgm-NFS4_CALLBACK] = np; args->callback.cb_program = pgm; args->callback.cb_location.r_netid = p->netid; args->callback.cb_location.r_addr = p->uaddr; args->callback_ident = pgm; np->s_program = pgm; mutex_exit(&ncg->nfs4_cb_lock); } static int nfs4_dquery(struct nfs4_svc_args *arg, model_t model) { file_t *fp; vnode_t *vp; rnode4_t *rp; int error; STRUCT_HANDLE(nfs4_svc_args, uap); STRUCT_SET_HANDLE(uap, model, arg); if ((fp = getf(STRUCT_FGET(uap, fd))) == NULL) return (EBADF); vp = fp->f_vnode; if (vp == NULL || vp->v_type != VREG || !vn_matchops(vp, nfs4_vnodeops)) { releasef(STRUCT_FGET(uap, fd)); return (EBADF); } rp = VTOR4(vp); /* * I can't convince myself that we need locking here. The * rnode cannot disappear and the value returned is instantly * stale anway, so why bother? */ error = suword32(STRUCT_FGETP(uap, netid), rp->r_deleg_type); releasef(STRUCT_FGET(uap, fd)); return (error); } /* * NFS4 client system call. This service does the * necessary initialization for the callback program. * This is fashioned after the server side interaction * between nfsd and the kernel. On the client, the * mount command forks and the child process does the * necessary interaction with the kernel. * * uap->fd is the fd of an open transport provider */ int nfs4_svc(struct nfs4_svc_args *arg, model_t model) { file_t *fp; int error; int readsize; char buf[KNC_STRSIZE], uaddr[KNC_STRSIZE]; char protofmly[KNC_STRSIZE], proto[KNC_STRSIZE]; size_t len; STRUCT_HANDLE(nfs4_svc_args, uap); struct netbuf addrmask; int cmd; SVCMASTERXPRT *cb_xprt; struct nfs4_callback_globals *ncg; #ifdef lint model = model; /* STRUCT macros don't always refer to it */ #endif STRUCT_SET_HANDLE(uap, model, arg); if (STRUCT_FGET(uap, cmd) == NFS4_DQUERY) return (nfs4_dquery(arg, model)); if (secpolicy_nfs(CRED()) != 0) return (EPERM); if ((fp = getf(STRUCT_FGET(uap, fd))) == NULL) return (EBADF); /* * Set read buffer size to rsize * and add room for RPC headers. */ readsize = nfs3tsize() + (RPC_MAXDATASIZE - NFS_MAXDATA); if (readsize < RPC_MAXDATASIZE) readsize = RPC_MAXDATASIZE; error = copyinstr((const char *)STRUCT_FGETP(uap, netid), buf, KNC_STRSIZE, &len); if (error) { releasef(STRUCT_FGET(uap, fd)); return (error); } cmd = STRUCT_FGET(uap, cmd); if (cmd & NFS4_KRPC_START) { addrmask.len = STRUCT_FGET(uap, addrmask.len); addrmask.maxlen = STRUCT_FGET(uap, addrmask.maxlen); addrmask.buf = kmem_alloc(addrmask.maxlen, KM_SLEEP); error = copyin(STRUCT_FGETP(uap, addrmask.buf), addrmask.buf, addrmask.len); if (error) { releasef(STRUCT_FGET(uap, fd)); kmem_free(addrmask.buf, addrmask.maxlen); return (error); } } else addrmask.buf = NULL; error = copyinstr((const char *)STRUCT_FGETP(uap, addr), uaddr, sizeof (uaddr), &len); if (error) { releasef(STRUCT_FGET(uap, fd)); if (addrmask.buf) kmem_free(addrmask.buf, addrmask.maxlen); return (error); } error = copyinstr((const char *)STRUCT_FGETP(uap, protofmly), protofmly, sizeof (protofmly), &len); if (error) { releasef(STRUCT_FGET(uap, fd)); if (addrmask.buf) kmem_free(addrmask.buf, addrmask.maxlen); return (error); } error = copyinstr((const char *)STRUCT_FGETP(uap, proto), proto, sizeof (proto), &len); if (error) { releasef(STRUCT_FGET(uap, fd)); if (addrmask.buf) kmem_free(addrmask.buf, addrmask.maxlen); return (error); } ncg = zone_getspecific(nfs4_callback_zone_key, nfs_zone()); ASSERT(ncg != NULL); mutex_enter(&ncg->nfs4_cb_lock); if (cmd & NFS4_SETPORT) nfs4_setport(buf, uaddr, protofmly, proto, ncg); if (cmd & NFS4_KRPC_START) { error = svc_tli_kcreate(fp, readsize, buf, &addrmask, &cb_xprt, &nfs4_cb_sct, NULL, NFS_CB_SVCPOOL_ID, FALSE); if (error) { CB_WARN1("nfs4_svc: svc_tli_kcreate failed %d\n", error); kmem_free(addrmask.buf, addrmask.maxlen); } } mutex_exit(&ncg->nfs4_cb_lock); releasef(STRUCT_FGET(uap, fd)); return (error); } struct nfs4_callback_globals * nfs4_get_callback_globals(void) { return (zone_getspecific(nfs4_callback_zone_key, nfs_zone())); } static void * nfs4_callback_init_zone(zoneid_t zoneid) { kstat_t *nfs4_callback_kstat; struct nfs4_callback_globals *ncg; ncg = kmem_zalloc(sizeof (*ncg), KM_SLEEP); ncg->nfs4prog2server = kmem_zalloc(nfs4_num_prognums * sizeof (struct nfs4_server *), KM_SLEEP); /* initialize the dlist */ mutex_init(&ncg->nfs4_dlist_lock, NULL, MUTEX_DEFAULT, NULL); list_create(&ncg->nfs4_dlist, sizeof (struct nfs4_dnode), offsetof(struct nfs4_dnode, linkage)); /* initialize cb_port list */ mutex_init(&ncg->nfs4_cb_lock, NULL, MUTEX_DEFAULT, NULL); list_create(&ncg->nfs4_cb_ports, sizeof (struct nfs4_cb_port), offsetof(struct nfs4_cb_port, linkage)); /* get our own copy of the kstats */ bcopy(&nfs4_callback_stats_tmpl, &ncg->nfs4_callback_stats, sizeof (nfs4_callback_stats_tmpl)); /* register "nfs:0:nfs4_callback_stats" for this zone */ if ((nfs4_callback_kstat = kstat_create_zone("nfs", 0, "nfs4_callback_stats", "misc", KSTAT_TYPE_NAMED, sizeof (ncg->nfs4_callback_stats) / sizeof (kstat_named_t), KSTAT_FLAG_VIRTUAL | KSTAT_FLAG_WRITABLE, zoneid)) != NULL) { nfs4_callback_kstat->ks_data = &ncg->nfs4_callback_stats; kstat_install(nfs4_callback_kstat); } return (ncg); } static void nfs4_discard_delegations(struct nfs4_callback_globals *ncg) { nfs4_server_t *sp; int i, num_removed; /* * It's OK here to just run through the registered "programs", as * servers without programs won't have any delegations to handle. */ for (i = 0; i < nfs4_num_prognums; i++) { rnode4_t *rp; mutex_enter(&ncg->nfs4_cb_lock); sp = ncg->nfs4prog2server[i]; mutex_exit(&ncg->nfs4_cb_lock); if (nfs4_server_vlock(sp, 1) == FALSE) continue; num_removed = 0; while ((rp = list_head(&sp->s_deleg_list)) != NULL) { mutex_enter(&rp->r_statev4_lock); if (rp->r_deleg_type == OPEN_DELEGATE_NONE) { /* * We need to take matters into our own hands, * as nfs4delegreturn_cleanup_impl() won't * remove this from the list. */ list_remove(&sp->s_deleg_list, rp); mutex_exit(&rp->r_statev4_lock); nfs4_dec_state_ref_count_nolock(sp, VTOMI4(RTOV4(rp))); num_removed++; continue; } mutex_exit(&rp->r_statev4_lock); VN_HOLD(RTOV4(rp)); mutex_exit(&sp->s_lock); /* * The following will remove the node from the list. */ nfs4delegreturn_cleanup_impl(rp, sp, ncg); VN_RELE(RTOV4(rp)); mutex_enter(&sp->s_lock); } mutex_exit(&sp->s_lock); /* each removed list node reles a reference */ while (num_removed-- > 0) nfs4_server_rele(sp); /* remove our reference for nfs4_server_vlock */ nfs4_server_rele(sp); } } /* ARGSUSED */ static void nfs4_callback_shutdown_zone(zoneid_t zoneid, void *data) { struct nfs4_callback_globals *ncg = data; /* * Clean pending delegation return list. */ nfs4_dlistclean_impl(ncg, NFS4_DR_DISCARD); /* * Discard all delegations. */ nfs4_discard_delegations(ncg); } static void nfs4_callback_fini_zone(zoneid_t zoneid, void *data) { struct nfs4_callback_globals *ncg = data; struct nfs4_cb_port *p; nfs4_server_t *sp, *next; nfs4_server_t freelist; int i; kstat_delete_byname_zone("nfs", 0, "nfs4_callback_stats", zoneid); /* * Discard all delegations that may have crept in since we did the * _shutdown. */ nfs4_discard_delegations(ncg); /* * We're completely done with this zone and all associated * nfs4_server_t's. Any remaining nfs4_server_ts should only have one * more reference outstanding -- the reference we didn't release in * nfs4_renew_lease_thread(). * * Here we need to run through the global nfs4_server_lst as we need to * deal with nfs4_server_ts without programs, as they also have threads * created for them, and so have outstanding references that we need to * release. */ freelist.forw = &freelist; freelist.back = &freelist; mutex_enter(&nfs4_server_lst_lock); sp = nfs4_server_lst.forw; while (sp != &nfs4_server_lst) { next = sp->forw; if (sp->zoneid == zoneid) { remque(sp); insque(sp, &freelist); } sp = next; } mutex_exit(&nfs4_server_lst_lock); sp = freelist.forw; while (sp != &freelist) { next = sp->forw; nfs4_server_rele(sp); /* free the list's reference */ sp = next; } #ifdef DEBUG for (i = 0; i < nfs4_num_prognums; i++) { ASSERT(ncg->nfs4prog2server[i] == NULL); } #endif kmem_free(ncg->nfs4prog2server, nfs4_num_prognums * sizeof (struct nfs4_server *)); mutex_enter(&ncg->nfs4_cb_lock); while ((p = list_head(&ncg->nfs4_cb_ports)) != NULL) { list_remove(&ncg->nfs4_cb_ports, p); kmem_free(p, sizeof (*p)); } list_destroy(&ncg->nfs4_cb_ports); mutex_destroy(&ncg->nfs4_cb_lock); list_destroy(&ncg->nfs4_dlist); mutex_destroy(&ncg->nfs4_dlist_lock); kmem_free(ncg, sizeof (*ncg)); } void nfs4_callback_init(void) { int i; SVC_CALLOUT *nfs4_cb_sc; /* initialize the callback table */ nfs4_cb_sc = kmem_alloc(nfs4_num_prognums * sizeof (SVC_CALLOUT), KM_SLEEP); for (i = 0; i < nfs4_num_prognums; i++) { nfs4_cb_sc[i].sc_prog = NFS4_CALLBACK+i; nfs4_cb_sc[i].sc_versmin = NFS_CB; nfs4_cb_sc[i].sc_versmax = NFS_CB; nfs4_cb_sc[i].sc_dispatch = cb_dispatch; } nfs4_cb_sct.sct_size = nfs4_num_prognums; nfs4_cb_sct.sct_free = FALSE; nfs4_cb_sct.sct_sc = nfs4_cb_sc; /* * Compute max bytes required for dyamically allocated parts * of cb_getattr reply. Only size and change are supported now. * If CB_GETATTR is changed to reply with additional attrs, * additional sizes must be added below. * * fattr4_change + fattr4_size == uint64_t + uint64_t */ cb_getattr_bytes = 2 * BYTES_PER_XDR_UNIT + 2 * BYTES_PER_XDR_UNIT; zone_key_create(&nfs4_callback_zone_key, nfs4_callback_init_zone, nfs4_callback_shutdown_zone, nfs4_callback_fini_zone); } void nfs4_callback_fini(void) { } /* * NB: This function can be called from the *wrong* zone (ie, the zone that * 'rp' belongs to and the caller's zone may not be the same). This can happen * if the zone is going away and we get called from nfs4_async_inactive(). In * this case the globals will be NULL and we won't update the counters, which * doesn't matter as the zone is going away anyhow. */ static void nfs4delegreturn_cleanup_impl(rnode4_t *rp, nfs4_server_t *np, struct nfs4_callback_globals *ncg) { mntinfo4_t *mi = VTOMI4(RTOV4(rp)); boolean_t need_rele = B_FALSE; /* * Caller must be holding mi_recovlock in read mode * to call here. This is provided by start_op. * Delegation management requires to grab s_lock * first and then r_statev4_lock. */ if (np == NULL) { np = find_nfs4_server_all(mi, 1); if (np == NULL) return; need_rele = B_TRUE; } else { mutex_enter(&np->s_lock); } mutex_enter(&rp->r_statev4_lock); if (rp->r_deleg_type == OPEN_DELEGATE_NONE) { mutex_exit(&rp->r_statev4_lock); mutex_exit(&np->s_lock); if (need_rele) nfs4_server_rele(np); return; } /* * Free the cred originally held when * the delegation was granted. Caller must * hold this cred if it wants to use it after * this call. */ crfree(rp->r_deleg_cred); rp->r_deleg_cred = NULL; rp->r_deleg_type = OPEN_DELEGATE_NONE; rp->r_deleg_needs_recovery = OPEN_DELEGATE_NONE; rp->r_deleg_needs_recall = FALSE; rp->r_deleg_return_pending = FALSE; /* * Remove the rnode from the server's list and * update the ref counts. */ list_remove(&np->s_deleg_list, rp); mutex_exit(&rp->r_statev4_lock); nfs4_dec_state_ref_count_nolock(np, mi); mutex_exit(&np->s_lock); /* removed list node removes a reference */ nfs4_server_rele(np); if (need_rele) nfs4_server_rele(np); if (ncg != NULL) ncg->nfs4_callback_stats.delegations.value.ui64--; } void nfs4delegreturn_cleanup(rnode4_t *rp, nfs4_server_t *np) { struct nfs4_callback_globals *ncg; if (np != NULL) { ncg = np->zone_globals; } else if (nfs_zone() == VTOMI4(RTOV4(rp))->mi_zone) { ncg = zone_getspecific(nfs4_callback_zone_key, nfs_zone()); ASSERT(ncg != NULL); } else { /* * Request coming from the wrong zone. */ ASSERT(getzoneid() == GLOBAL_ZONEID); ncg = NULL; } nfs4delegreturn_cleanup_impl(rp, np, ncg); } static void nfs4delegreturn_save_lost_rqst(int error, nfs4_lost_rqst_t *lost_rqstp, cred_t *cr, vnode_t *vp) { if (error != ETIMEDOUT && error != EINTR && !NFS4_FRC_UNMT_ERR(error, vp->v_vfsp)) { lost_rqstp->lr_op = 0; return; } NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE, "nfs4close_save_lost_rqst: error %d", error)); lost_rqstp->lr_op = OP_DELEGRETURN; /* * The vp is held and rele'd via the recovery code. * See nfs4_save_lost_rqst. */ lost_rqstp->lr_vp = vp; lost_rqstp->lr_dvp = NULL; lost_rqstp->lr_oop = NULL; lost_rqstp->lr_osp = NULL; lost_rqstp->lr_lop = NULL; lost_rqstp->lr_cr = cr; lost_rqstp->lr_flk = NULL; lost_rqstp->lr_putfirst = FALSE; } static void nfs4delegreturn_otw(rnode4_t *rp, cred_t *cr, nfs4_error_t *ep) { COMPOUND4args_clnt args; COMPOUND4res_clnt res; nfs_argop4 argops[3]; nfs4_ga_res_t *garp = NULL; hrtime_t t; int numops; int doqueue = 1; args.ctag = TAG_DELEGRETURN; numops = 3; /* PUTFH, GETATTR, DELEGRETURN */ args.array = argops; args.array_len = numops; argops[0].argop = OP_CPUTFH; argops[0].nfs_argop4_u.opcputfh.sfh = rp->r_fh; argops[1].argop = OP_GETATTR; argops[1].nfs_argop4_u.opgetattr.attr_request = NFS4_VATTR_MASK; argops[1].nfs_argop4_u.opgetattr.mi = VTOMI4(RTOV4(rp)); argops[2].argop = OP_DELEGRETURN; argops[2].nfs_argop4_u.opdelegreturn.deleg_stateid = rp->r_deleg_stateid; t = gethrtime(); rfs4call(VTOMI4(RTOV4(rp)), &args, &res, cr, &doqueue, 0, ep); if (ep->error) return; if (res.status == NFS4_OK) { garp = &res.array[1].nfs_resop4_u.opgetattr.ga_res; nfs4_attr_cache(RTOV4(rp), garp, t, cr, TRUE, NULL); } (void) xdr_free(xdr_COMPOUND4res_clnt, (caddr_t)&res); } int nfs4_do_delegreturn(rnode4_t *rp, int flags, cred_t *cr, struct nfs4_callback_globals *ncg) { vnode_t *vp = RTOV4(rp); mntinfo4_t *mi = VTOMI4(vp); nfs4_lost_rqst_t lost_rqst; nfs4_recov_state_t recov_state; bool_t needrecov = FALSE, recovonly, done = FALSE; nfs4_error_t e = { 0, NFS4_OK, RPC_SUCCESS }; ncg->nfs4_callback_stats.delegreturn.value.ui64++; while (!done) { e.error = nfs4_start_fop(mi, vp, NULL, OH_DELEGRETURN, &recov_state, &recovonly); if (e.error) { if (flags & NFS4_DR_FORCE) { (void) nfs_rw_enter_sig(&mi->mi_recovlock, RW_READER, 0); nfs4delegreturn_cleanup_impl(rp, NULL, ncg); nfs_rw_exit(&mi->mi_recovlock); } break; } /* * Check to see if the delegation has already been * returned by the recovery thread. The state of * the delegation cannot change at this point due * to start_fop and the r_deleg_recall_lock. */ if (rp->r_deleg_type == OPEN_DELEGATE_NONE) { e.error = 0; nfs4_end_op(mi, vp, NULL, &recov_state, needrecov); break; } if (recovonly) { /* * Delegation will be returned via the * recovery framework. Build a lost request * structure, start recovery and get out. */ nfs4_error_init(&e, EINTR); nfs4delegreturn_save_lost_rqst(e.error, &lost_rqst, cr, vp); (void) nfs4_start_recovery(&e, mi, vp, NULL, &rp->r_deleg_stateid, lost_rqst.lr_op == OP_DELEGRETURN ? &lost_rqst : NULL, OP_DELEGRETURN, NULL, NULL, NULL); nfs4_end_op(mi, vp, NULL, &recov_state, needrecov); break; } nfs4delegreturn_otw(rp, cr, &e); /* * Ignore some errors on delegreturn; no point in marking * the file dead on a state destroying operation. */ if (e.error == 0 && (nfs4_recov_marks_dead(e.stat) || e.stat == NFS4ERR_BADHANDLE || e.stat == NFS4ERR_STALE || (e.stat == NFS4ERR_STALE_STATEID && !(rp->r_flags & R4HASHED)))) needrecov = FALSE; else needrecov = nfs4_needs_recovery(&e, TRUE, vp->v_vfsp); if (needrecov) { nfs4delegreturn_save_lost_rqst(e.error, &lost_rqst, cr, vp); (void) nfs4_start_recovery(&e, mi, vp, NULL, &rp->r_deleg_stateid, lost_rqst.lr_op == OP_DELEGRETURN ? &lost_rqst : NULL, OP_DELEGRETURN, NULL, NULL, NULL); } else { nfs4delegreturn_cleanup_impl(rp, NULL, ncg); done = TRUE; } nfs4_end_op(mi, vp, NULL, &recov_state, needrecov); } return (e.error); } /* * nfs4_resend_delegreturn - used to drive the delegreturn * operation via the recovery thread. */ void nfs4_resend_delegreturn(nfs4_lost_rqst_t *lorp, nfs4_error_t *ep, nfs4_server_t *np) { rnode4_t *rp = VTOR4(lorp->lr_vp); /* If the file failed recovery, just quit. */ mutex_enter(&rp->r_statelock); if (rp->r_flags & R4RECOVERR) { ep->error = EIO; } mutex_exit(&rp->r_statelock); if (!ep->error) nfs4delegreturn_otw(rp, lorp->lr_cr, ep); /* * If recovery is now needed, then return the error * and status and let the recovery thread handle it, * including re-driving another delegreturn. Otherwise, * just give up and clean up the delegation. */ if (nfs4_needs_recovery(ep, TRUE, lorp->lr_vp->v_vfsp)) return; if (rp->r_deleg_type != OPEN_DELEGATE_NONE) nfs4delegreturn_cleanup(rp, np); nfs4_error_zinit(ep); } /* * nfs4delegreturn - general function to return a delegation. * * NFS4_DR_FORCE - return the delegation even if start_op fails * NFS4_DR_PUSH - push modified data back to the server via VOP_PUTPAGE * NFS4_DR_DISCARD - discard the delegation w/o delegreturn * NFS4_DR_DID_OP - calling function already did nfs4_start_op * NFS4_DR_RECALL - delegreturned initiated via CB_RECALL * NFS4_DR_REOPEN - do file reopens, if applicable */ static int nfs4delegreturn_impl(rnode4_t *rp, int flags, struct nfs4_callback_globals *ncg) { int error = 0; cred_t *cr = NULL; vnode_t *vp; bool_t needrecov = FALSE; bool_t rw_entered = FALSE; bool_t do_reopen; vp = RTOV4(rp); /* * If NFS4_DR_DISCARD is set by itself, take a short-cut and * discard without doing an otw DELEGRETURN. This may only be used * by the recovery thread because it bypasses the synchronization * with r_deleg_recall_lock and mi->mi_recovlock. */ if (flags == NFS4_DR_DISCARD) { nfs4delegreturn_cleanup_impl(rp, NULL, ncg); return (0); } if (flags & NFS4_DR_DID_OP) { /* * Caller had already done start_op, which means the * r_deleg_recall_lock is already held in READ mode * so we cannot take it in write mode. Return the * delegation asynchronously. * * Remove the NFS4_DR_DID_OP flag so we don't * get stuck looping through here. */ VN_HOLD(vp); nfs4delegreturn_async(rp, (flags & ~NFS4_DR_DID_OP), FALSE); return (0); } /* * Verify we still have a delegation and crhold the credential. */ mutex_enter(&rp->r_statev4_lock); if (rp->r_deleg_type == OPEN_DELEGATE_NONE) { mutex_exit(&rp->r_statev4_lock); goto out; } cr = rp->r_deleg_cred; ASSERT(cr != NULL); crhold(cr); mutex_exit(&rp->r_statev4_lock); /* * Push the modified data back to the server synchronously * before doing DELEGRETURN. */ if (flags & NFS4_DR_PUSH) (void) VOP_PUTPAGE(vp, 0, 0, 0, cr, NULL); /* * Take r_deleg_recall_lock in WRITE mode, this will prevent * nfs4_is_otw_open_necessary from trying to use the delegation * while the DELEGRETURN is in progress. */ (void) nfs_rw_enter_sig(&rp->r_deleg_recall_lock, RW_WRITER, FALSE); rw_entered = TRUE; if (rp->r_deleg_type == OPEN_DELEGATE_NONE) goto out; if (flags & NFS4_DR_REOPEN) { /* * If R4RECOVERRP is already set, then skip re-opening * the delegation open streams and go straight to doing * delegreturn. (XXX if the file has failed recovery, then the * delegreturn attempt is likely to be futile.) */ mutex_enter(&rp->r_statelock); do_reopen = !(rp->r_flags & R4RECOVERRP); mutex_exit(&rp->r_statelock); if (do_reopen) { error = deleg_reopen(vp, &needrecov, ncg, flags); if (error != 0) { if ((flags & (NFS4_DR_FORCE | NFS4_DR_RECALL)) == 0) goto out; } else if (needrecov) { if ((flags & NFS4_DR_FORCE) == 0) goto out; } } } if (flags & NFS4_DR_DISCARD) { mntinfo4_t *mi = VTOMI4(RTOV4(rp)); mutex_enter(&rp->r_statelock); /* * deleg_return_pending is cleared inside of delegation_accept * when a delegation is accepted. if this flag has been * cleared, then a new delegation has overwritten the one we * were about to throw away. */ if (!rp->r_deleg_return_pending) { mutex_exit(&rp->r_statelock); goto out; } mutex_exit(&rp->r_statelock); (void) nfs_rw_enter_sig(&mi->mi_recovlock, RW_READER, FALSE); nfs4delegreturn_cleanup_impl(rp, NULL, ncg); nfs_rw_exit(&mi->mi_recovlock); } else { error = nfs4_do_delegreturn(rp, flags, cr, ncg); } out: if (cr) crfree(cr); if (rw_entered) nfs_rw_exit(&rp->r_deleg_recall_lock); return (error); } int nfs4delegreturn(rnode4_t *rp, int flags) { struct nfs4_callback_globals *ncg; ncg = zone_getspecific(nfs4_callback_zone_key, nfs_zone()); ASSERT(ncg != NULL); return (nfs4delegreturn_impl(rp, flags, ncg)); } void nfs4delegreturn_async(rnode4_t *rp, int flags, bool_t trunc) { struct cb_recall_pass *pp; pp = kmem_alloc(sizeof (struct cb_recall_pass), KM_SLEEP); pp->rp = rp; pp->flags = flags; pp->truncate = trunc; /* * Fire up a thread to do the actual delegreturn * Caller must guarantee that the rnode doesn't * vanish (by calling VN_HOLD). */ (void) zthread_create(NULL, 0, nfs4delegreturn_thread, pp, 0, minclsyspri); } static void delegreturn_all_thread(rpcprog_t *pp) { nfs4_server_t *np; bool_t found = FALSE; rpcprog_t prog; rnode4_t *rp; vnode_t *vp; zoneid_t zoneid = getzoneid(); struct nfs4_callback_globals *ncg; NFS4_DEBUG(nfs4_drat_debug, (CE_NOTE, "delereturn_all_thread: prog %d\n", *pp)); prog = *pp; kmem_free(pp, sizeof (*pp)); pp = NULL; mutex_enter(&nfs4_server_lst_lock); for (np = nfs4_server_lst.forw; np != &nfs4_server_lst; np = np->forw) { if (np->zoneid == zoneid && np->s_program == prog) { mutex_enter(&np->s_lock); found = TRUE; break; } } mutex_exit(&nfs4_server_lst_lock); /* * It's possible that the nfs4_server which was using this * program number has vanished since this thread is async. * If so, just return. Your work here is finished, my friend. */ if (!found) goto out; ncg = np->zone_globals; while ((rp = list_head(&np->s_deleg_list)) != NULL) { vp = RTOV4(rp); VN_HOLD(vp); mutex_exit(&np->s_lock); (void) nfs4delegreturn_impl(rp, NFS4_DR_PUSH|NFS4_DR_REOPEN, ncg); VN_RELE(vp); /* retake the s_lock for next trip through the loop */ mutex_enter(&np->s_lock); } mutex_exit(&np->s_lock); out: NFS4_DEBUG(nfs4_drat_debug, (CE_NOTE, "delereturn_all_thread: complete\n")); zthread_exit(); } void nfs4_delegreturn_all(nfs4_server_t *sp) { rpcprog_t pro, *pp; mutex_enter(&sp->s_lock); /* Check to see if the delegation list is empty */ if (list_head(&sp->s_deleg_list) == NULL) { mutex_exit(&sp->s_lock); return; } /* * Grab the program number; the async thread will use this * to find the nfs4_server. */ pro = sp->s_program; mutex_exit(&sp->s_lock); pp = kmem_alloc(sizeof (rpcprog_t), KM_SLEEP); *pp = pro; (void) zthread_create(NULL, 0, delegreturn_all_thread, pp, 0, minclsyspri); } /* * Discard any delegations * * Iterate over the servers s_deleg_list and * for matching mount-point rnodes discard * the delegation. */ void nfs4_deleg_discard(mntinfo4_t *mi, nfs4_server_t *sp) { rnode4_t *rp, *next; mntinfo4_t *r_mi; struct nfs4_callback_globals *ncg; ASSERT(mutex_owned(&sp->s_lock)); ncg = sp->zone_globals; for (rp = list_head(&sp->s_deleg_list); rp != NULL; rp = next) { r_mi = VTOMI4(RTOV4(rp)); next = list_next(&sp->s_deleg_list, rp); if (r_mi != mi) { /* * Skip if this rnode is in not on the * same mount-point */ continue; } ASSERT(rp->r_deleg_type == OPEN_DELEGATE_READ); #ifdef DEBUG if (nfs4_client_recov_debug) { zprintf(getzoneid(), "nfs4_deleg_discard: matched rnode %p " "-- discarding delegation\n", (void *)rp); } #endif mutex_enter(&rp->r_statev4_lock); /* * Free the cred originally held when the delegation * was granted. Also need to decrement the refcnt * on this server for each delegation we discard */ if (rp->r_deleg_cred) crfree(rp->r_deleg_cred); rp->r_deleg_cred = NULL; rp->r_deleg_type = OPEN_DELEGATE_NONE; rp->r_deleg_needs_recovery = OPEN_DELEGATE_NONE; rp->r_deleg_needs_recall = FALSE; ASSERT(sp->s_refcnt > 1); sp->s_refcnt--; list_remove(&sp->s_deleg_list, rp); mutex_exit(&rp->r_statev4_lock); nfs4_dec_state_ref_count_nolock(sp, mi); ncg->nfs4_callback_stats.delegations.value.ui64--; } } /* * Reopen any open streams that were covered by the given file's * delegation. * Returns zero or an errno value. If there was no error, *recovp * indicates whether recovery was initiated. */ static int deleg_reopen(vnode_t *vp, bool_t *recovp, struct nfs4_callback_globals *ncg, int flags) { nfs4_open_stream_t *osp; nfs4_recov_state_t recov_state; bool_t needrecov = FALSE; mntinfo4_t *mi; rnode4_t *rp; nfs4_error_t e = { 0, NFS4_OK, RPC_SUCCESS }; int claimnull; mi = VTOMI4(vp); rp = VTOR4(vp); recov_state.rs_flags = 0; recov_state.rs_num_retry_despite_err = 0; retry: if ((e.error = nfs4_start_op(mi, vp, NULL, &recov_state)) != 0) { return (e.error); } /* * if we mean to discard the delegation, it must be BAD, so don't * use it when doing the reopen or it will fail too. */ claimnull = (flags & NFS4_DR_DISCARD); /* * Loop through the open streams for this rnode to find * all of the ones created using the delegation state ID. * Each of these needs to be re-opened. */ while ((osp = get_next_deleg_stream(rp, claimnull)) != NULL) { if (claimnull) { nfs4_reopen(vp, osp, &e, CLAIM_NULL, FALSE, FALSE); } else { ncg->nfs4_callback_stats.claim_cur.value.ui64++; nfs4_reopen(vp, osp, &e, CLAIM_DELEGATE_CUR, FALSE, FALSE); if (e.error == 0 && e.stat == NFS4_OK) ncg->nfs4_callback_stats. claim_cur_ok.value.ui64++; } if (e.error == EAGAIN) { open_stream_rele(osp, rp); nfs4_end_op(mi, vp, NULL, &recov_state, TRUE); goto retry; } /* * if error is EINTR, ETIMEDOUT, or NFS4_FRC_UNMT_ERR, then * recovery has already been started inside of nfs4_reopen. */ if (e.error == EINTR || e.error == ETIMEDOUT || NFS4_FRC_UNMT_ERR(e.error, vp->v_vfsp)) { open_stream_rele(osp, rp); break; } needrecov = nfs4_needs_recovery(&e, TRUE, vp->v_vfsp); if (e.error != 0 && !needrecov) { /* * Recovery is not possible, but don't give up yet; * we'd still like to do delegreturn after * reopening as many streams as possible. * Continue processing the open streams. */ ncg->nfs4_callback_stats.recall_failed.value.ui64++; } else if (needrecov) { /* * Start recovery and bail out. The recovery * thread will take it from here. */ (void) nfs4_start_recovery(&e, mi, vp, NULL, NULL, NULL, OP_OPEN, NULL, NULL, NULL); open_stream_rele(osp, rp); *recovp = TRUE; break; } open_stream_rele(osp, rp); } nfs4_end_op(mi, vp, NULL, &recov_state, needrecov); return (e.error); } /* * get_next_deleg_stream - returns the next open stream which * represents a delegation for this rnode. In order to assure * forward progress, the caller must guarantee that each open * stream returned is changed so that a future call won't return * it again. * * There are several ways for the open stream to change. If the open * stream is !os_delegation, then we aren't interested in it. Also, if * either os_failed_reopen or !os_valid, then don't return the osp. * * If claimnull is false (doing reopen CLAIM_DELEGATE_CUR) then return * the osp if it is an os_delegation open stream. Also, if the rnode still * has r_deleg_return_pending, then return the os_delegation osp. Lastly, * if the rnode's r_deleg_stateid is different from the osp's open_stateid, * then return the osp. * * We have already taken the 'r_deleg_recall_lock' as WRITER, which * prevents new OPENs from going OTW (as start_fop takes this * lock in READ mode); thus, no new open streams can be created * (which inherently means no new delegation open streams are * being created). */ static nfs4_open_stream_t * get_next_deleg_stream(rnode4_t *rp, int claimnull) { nfs4_open_stream_t *osp; ASSERT(nfs_rw_lock_held(&rp->r_deleg_recall_lock, RW_WRITER)); /* * Search through the list of open streams looking for * one that was created while holding the delegation. */ mutex_enter(&rp->r_os_lock); for (osp = list_head(&rp->r_open_streams); osp != NULL; osp = list_next(&rp->r_open_streams, osp)) { mutex_enter(&osp->os_sync_lock); if (!osp->os_delegation || osp->os_failed_reopen || !osp->os_valid) { mutex_exit(&osp->os_sync_lock); continue; } if (!claimnull || rp->r_deleg_return_pending || !stateid4_cmp(&osp->open_stateid, &rp->r_deleg_stateid)) { osp->os_ref_count++; mutex_exit(&osp->os_sync_lock); mutex_exit(&rp->r_os_lock); return (osp); } mutex_exit(&osp->os_sync_lock); } mutex_exit(&rp->r_os_lock); return (NULL); } static void nfs4delegreturn_thread(struct cb_recall_pass *args) { rnode4_t *rp; vnode_t *vp; cred_t *cr; int dtype, error, flags; bool_t rdirty, rip; kmutex_t cpr_lock; callb_cpr_t cpr_info; struct nfs4_callback_globals *ncg; ncg = zone_getspecific(nfs4_callback_zone_key, nfs_zone()); ASSERT(ncg != NULL); mutex_init(&cpr_lock, NULL, MUTEX_DEFAULT, NULL); CALLB_CPR_INIT(&cpr_info, &cpr_lock, callb_generic_cpr, "nfsv4delegRtn"); rp = args->rp; vp = RTOV4(rp); mutex_enter(&rp->r_statev4_lock); if (rp->r_deleg_type == OPEN_DELEGATE_NONE) { mutex_exit(&rp->r_statev4_lock); goto out; } mutex_exit(&rp->r_statev4_lock); /* * Take the read-write lock in read mode to prevent other * threads from modifying the data during the recall. This * doesn't affect mmappers. */ (void) nfs_rw_enter_sig(&rp->r_rwlock, RW_READER, FALSE); /* Proceed with delegreturn */ mutex_enter(&rp->r_statev4_lock); if (rp->r_deleg_type == OPEN_DELEGATE_NONE) { mutex_exit(&rp->r_statev4_lock); nfs_rw_exit(&rp->r_rwlock); goto out; } dtype = rp->r_deleg_type; cr = rp->r_deleg_cred; ASSERT(cr != NULL); crhold(cr); mutex_exit(&rp->r_statev4_lock); flags = args->flags; /* * If the file is being truncated at the server, then throw * away all of the pages, it doesn't matter what flavor of * delegation we have. */ if (args->truncate) { ncg->nfs4_callback_stats.recall_trunc.value.ui64++; nfs4_invalidate_pages(vp, 0, cr); } else if (dtype == OPEN_DELEGATE_WRITE) { mutex_enter(&rp->r_statelock); rdirty = rp->r_flags & R4DIRTY; mutex_exit(&rp->r_statelock); if (rdirty) { error = VOP_PUTPAGE(vp, 0, 0, 0, cr, NULL); if (error) CB_WARN1("nfs4delegreturn_thread:" " VOP_PUTPAGE: %d\n", error); } /* turn off NFS4_DR_PUSH because we just did that above. */ flags &= ~NFS4_DR_PUSH; } mutex_enter(&rp->r_statelock); rip = rp->r_flags & R4RECOVERRP; mutex_exit(&rp->r_statelock); /* If a failed recovery is indicated, discard the pages */ if (rip) { error = VOP_PUTPAGE(vp, 0, 0, B_INVAL, cr, NULL); if (error) CB_WARN1("nfs4delegreturn_thread: VOP_PUTPAGE: %d\n", error); } /* * Pass the flags to nfs4delegreturn_impl, but be sure not to pass * NFS4_DR_DID_OP, which just calls nfs4delegreturn_async again. */ flags &= ~NFS4_DR_DID_OP; (void) nfs4delegreturn_impl(rp, flags, ncg); nfs_rw_exit(&rp->r_rwlock); crfree(cr); out: kmem_free(args, sizeof (struct cb_recall_pass)); VN_RELE(vp); mutex_enter(&cpr_lock); CALLB_CPR_EXIT(&cpr_info); mutex_destroy(&cpr_lock); zthread_exit(); } /* * This function has one assumption that the caller of this function is * either doing recovery (therefore cannot call nfs4_start_op) or has * already called nfs4_start_op(). */ void nfs4_delegation_accept(rnode4_t *rp, open_claim_type4 claim, OPEN4res *res, nfs4_ga_res_t *garp, cred_t *cr) { open_read_delegation4 *orp; open_write_delegation4 *owp; nfs4_server_t *np; bool_t already = FALSE; bool_t recall = FALSE; bool_t valid_garp = TRUE; bool_t delegation_granted = FALSE; bool_t dr_needed = FALSE; bool_t recov; int dr_flags = 0; long mapcnt; uint_t rflag; mntinfo4_t *mi; struct nfs4_callback_globals *ncg; open_delegation_type4 odt; ncg = zone_getspecific(nfs4_callback_zone_key, nfs_zone()); ASSERT(ncg != NULL); mi = VTOMI4(RTOV4(rp)); /* * Accept a delegation granted to the client via an OPEN. * Set the delegation fields in the rnode and insert the * rnode onto the list anchored in the nfs4_server_t. The * proper locking order requires the nfs4_server_t first, * even though it may not be needed in all cases. * * NB: find_nfs4_server returns with s_lock held. */ if ((np = find_nfs4_server(mi)) == NULL) return; /* grab the statelock too, for examining r_mapcnt */ mutex_enter(&rp->r_statelock); mutex_enter(&rp->r_statev4_lock); if (rp->r_deleg_type == OPEN_DELEGATE_READ || rp->r_deleg_type == OPEN_DELEGATE_WRITE) already = TRUE; odt = res->delegation.delegation_type; if (odt == OPEN_DELEGATE_READ) { rp->r_deleg_type = res->delegation.delegation_type; orp = &res->delegation.open_delegation4_u.read; rp->r_deleg_stateid = orp->stateid; rp->r_deleg_perms = orp->permissions; if (claim == CLAIM_PREVIOUS) if ((recall = orp->recall) != 0) dr_needed = TRUE; delegation_granted = TRUE; ncg->nfs4_callback_stats.delegations.value.ui64++; ncg->nfs4_callback_stats.delegaccept_r.value.ui64++; } else if (odt == OPEN_DELEGATE_WRITE) { rp->r_deleg_type = res->delegation.delegation_type; owp = &res->delegation.open_delegation4_u.write; rp->r_deleg_stateid = owp->stateid; rp->r_deleg_perms = owp->permissions; rp->r_deleg_limit = owp->space_limit; if (claim == CLAIM_PREVIOUS) if ((recall = owp->recall) != 0) dr_needed = TRUE; delegation_granted = TRUE; if (garp == NULL || !garp->n4g_change_valid) { valid_garp = FALSE; rp->r_deleg_change = 0; rp->r_deleg_change_grant = 0; } else { rp->r_deleg_change = garp->n4g_change; rp->r_deleg_change_grant = garp->n4g_change; } mapcnt = rp->r_mapcnt; rflag = rp->r_flags; /* * Update the delegation change attribute if * there are mappers for the file is dirty. This * might be the case during recovery after server * reboot. */ if (mapcnt > 0 || rflag & R4DIRTY) rp->r_deleg_change++; NFS4_DEBUG(nfs4_callback_debug, (CE_NOTE, "nfs4_delegation_accept: r_deleg_change: 0x%x\n", (int)(rp->r_deleg_change >> 32))); NFS4_DEBUG(nfs4_callback_debug, (CE_NOTE, "nfs4_delegation_accept: r_delg_change_grant: 0x%x\n", (int)(rp->r_deleg_change_grant >> 32))); ncg->nfs4_callback_stats.delegations.value.ui64++; ncg->nfs4_callback_stats.delegaccept_rw.value.ui64++; } else if (already) { /* * No delegation granted. If the rnode currently has * has one, then consider it tainted and return it. */ dr_needed = TRUE; } if (delegation_granted) { /* Add the rnode to the list. */ if (!already) { crhold(cr); rp->r_deleg_cred = cr; ASSERT(mutex_owned(&np->s_lock)); list_insert_head(&np->s_deleg_list, rp); /* added list node gets a reference */ np->s_refcnt++; nfs4_inc_state_ref_count_nolock(np, mi); } rp->r_deleg_needs_recovery = OPEN_DELEGATE_NONE; } /* * We've now safely accepted the delegation, if any. Drop the * locks and figure out what post-processing is needed. We'd * like to retain r_statev4_lock, but nfs4_server_rele takes * s_lock which would be a lock ordering violation. */ mutex_exit(&rp->r_statev4_lock); mutex_exit(&rp->r_statelock); mutex_exit(&np->s_lock); nfs4_server_rele(np); /* * Check to see if we are in recovery. Remember that * this function is protected by start_op, so a recovery * cannot begin until we are out of here. */ mutex_enter(&mi->mi_lock); recov = mi->mi_recovflags & MI4_RECOV_ACTIV; mutex_exit(&mi->mi_lock); mutex_enter(&rp->r_statev4_lock); if (nfs4_delegreturn_policy == IMMEDIATE || !valid_garp) dr_needed = TRUE; if (dr_needed && rp->r_deleg_return_pending == FALSE) { if (recov) { /* * We cannot call delegreturn from inside * of recovery or VOP_PUTPAGE will hang * due to nfs4_start_fop call in * nfs4write. Use dlistadd to add the * rnode to the list of rnodes needing * cleaning. We do not need to do reopen * here because recov_openfiles will do it. * In the non-recall case, just discard the * delegation as it is no longer valid. */ if (recall) dr_flags = NFS4_DR_PUSH; else dr_flags = NFS4_DR_PUSH|NFS4_DR_DISCARD; nfs4_dlistadd(rp, ncg, dr_flags); dr_flags = 0; } else { /* * Push the modified data back to the server, * reopen any delegation open streams, and return * the delegation. Drop the statev4_lock first! */ dr_flags = NFS4_DR_PUSH|NFS4_DR_DID_OP|NFS4_DR_REOPEN; } } mutex_exit(&rp->r_statev4_lock); if (dr_flags) (void) nfs4delegreturn_impl(rp, dr_flags, ncg); } /* * nfs4delegabandon - Abandon the delegation on an rnode4. This code * is called when the client receives EXPIRED, BAD_STATEID, OLD_STATEID * or BADSEQID and the recovery code is unable to recover. Push any * dirty data back to the server and return the delegation (if any). */ void nfs4delegabandon(rnode4_t *rp) { vnode_t *vp; struct cb_recall_pass *pp; open_delegation_type4 dt; mutex_enter(&rp->r_statev4_lock); dt = rp->r_deleg_type; mutex_exit(&rp->r_statev4_lock); if (dt == OPEN_DELEGATE_NONE) return; vp = RTOV4(rp); VN_HOLD(vp); pp = kmem_alloc(sizeof (struct cb_recall_pass), KM_SLEEP); pp->rp = rp; /* * Recovery on the file has failed and we want to return * the delegation. We don't want to reopen files and * nfs4delegreturn_thread() figures out what to do about * the data. The only thing to do is attempt to return * the delegation. */ pp->flags = 0; pp->truncate = FALSE; /* * Fire up a thread to do the delegreturn; this is * necessary because we could be inside a GETPAGE or * PUTPAGE and we cannot do another one. */ (void) zthread_create(NULL, 0, nfs4delegreturn_thread, pp, 0, minclsyspri); } static int wait_for_recall1(vnode_t *vp, nfs4_op_hint_t op, nfs4_recov_state_t *rsp, int flg) { rnode4_t *rp; int error = 0; #ifdef lint op = op; #endif if (vp && vp->v_type == VREG) { rp = VTOR4(vp); /* * Take r_deleg_recall_lock in read mode to synchronize * with delegreturn. */ error = nfs_rw_enter_sig(&rp->r_deleg_recall_lock, RW_READER, INTR4(vp)); if (error == 0) rsp->rs_flags |= flg; } return (error); } void nfs4_end_op_recall(vnode_t *vp1, vnode_t *vp2, nfs4_recov_state_t *rsp) { NFS4_DEBUG(nfs4_recall_debug, (CE_NOTE, "nfs4_end_op_recall: 0x%p, 0x%p\n", (void *)vp1, (void *)vp2)); if (vp2 && rsp->rs_flags & NFS4_RS_RECALL_HELD2) nfs_rw_exit(&VTOR4(vp2)->r_deleg_recall_lock); if (vp1 && rsp->rs_flags & NFS4_RS_RECALL_HELD1) nfs_rw_exit(&VTOR4(vp1)->r_deleg_recall_lock); } int wait_for_recall(vnode_t *vp1, vnode_t *vp2, nfs4_op_hint_t op, nfs4_recov_state_t *rsp) { int error; NFS4_DEBUG(nfs4_recall_debug, (CE_NOTE, "wait_for_recall: 0x%p, 0x%p\n", (void *)vp1, (void *) vp2)); rsp->rs_flags &= ~(NFS4_RS_RECALL_HELD1|NFS4_RS_RECALL_HELD2); if ((error = wait_for_recall1(vp1, op, rsp, NFS4_RS_RECALL_HELD1)) != 0) return (error); if ((error = wait_for_recall1(vp2, op, rsp, NFS4_RS_RECALL_HELD2)) != 0) { if (rsp->rs_flags & NFS4_RS_RECALL_HELD1) { nfs_rw_exit(&VTOR4(vp1)->r_deleg_recall_lock); rsp->rs_flags &= ~NFS4_RS_RECALL_HELD1; } return (error); } return (0); } /* * nfs4_dlistadd - Add this rnode to a list of rnodes to be * DELEGRETURN'd at the end of recovery. */ static void nfs4_dlistadd(rnode4_t *rp, struct nfs4_callback_globals *ncg, int flags) { struct nfs4_dnode *dp; ASSERT(mutex_owned(&rp->r_statev4_lock)); /* * Mark the delegation as having a return pending. * This will prevent the use of the delegation stateID * by read, write, setattr and open. */ rp->r_deleg_return_pending = TRUE; dp = kmem_alloc(sizeof (*dp), KM_SLEEP); VN_HOLD(RTOV4(rp)); dp->rnodep = rp; dp->flags = flags; mutex_enter(&ncg->nfs4_dlist_lock); list_insert_head(&ncg->nfs4_dlist, dp); #ifdef DEBUG ncg->nfs4_dlistadd_c++; #endif mutex_exit(&ncg->nfs4_dlist_lock); } /* * nfs4_dlistclean_impl - Do DELEGRETURN for each rnode on the list. * of files awaiting cleaning. If the override_flags are non-zero * then use them rather than the flags that were set when the rnode * was added to the dlist. */ static void nfs4_dlistclean_impl(struct nfs4_callback_globals *ncg, int override_flags) { rnode4_t *rp; struct nfs4_dnode *dp; int flags; ASSERT(override_flags == 0 || override_flags == NFS4_DR_DISCARD); mutex_enter(&ncg->nfs4_dlist_lock); while ((dp = list_head(&ncg->nfs4_dlist)) != NULL) { #ifdef DEBUG ncg->nfs4_dlistclean_c++; #endif list_remove(&ncg->nfs4_dlist, dp); mutex_exit(&ncg->nfs4_dlist_lock); rp = dp->rnodep; flags = (override_flags != 0) ? override_flags : dp->flags; kmem_free(dp, sizeof (*dp)); (void) nfs4delegreturn_impl(rp, flags, ncg); VN_RELE(RTOV4(rp)); mutex_enter(&ncg->nfs4_dlist_lock); } mutex_exit(&ncg->nfs4_dlist_lock); } void nfs4_dlistclean(void) { struct nfs4_callback_globals *ncg; ncg = zone_getspecific(nfs4_callback_zone_key, nfs_zone()); ASSERT(ncg != NULL); nfs4_dlistclean_impl(ncg, 0); } nfs4_rcsync_t * nfs4_recall_sync_start(mntinfo4_t *mi) { nfs4_server_t *sp = mi->mi_srv; nfs4_rcsync_t *rcp = kmem_zalloc(sizeof (*rcp), KM_SLEEP); if (nfs4_server_vlock(sp, 0) == FALSE) { rcp->rs_flags = RS_SERVER_GONE; return rcp; } rcp->rs_mi = mi; mutex_enter(&sp->s_rcsync_lock); rcp->rs_seq = sp->s_rcsync_seq++; list_insert_tail(&sp->s_rcsync_list, rcp); mutex_exit(&sp->s_rcsync_lock); mutex_exit(&sp->s_lock); nfs4_server_rele(sp); return rcp; } void nfs4_recall_sync_end(mntinfo4_t *mi, nfs4_rcsync_t *rcp) { nfs4_server_t *sp = mi->mi_srv; if (nfs4_server_vlock(sp, 1) == FALSE) { ASSERT((rcp->rs_flags & RS_SERVER_GONE) != 0); kmem_free(rcp, sizeof (*rcp)); return; } mutex_enter(&sp->s_rcsync_lock); if ((rcp->rs_flags & RS_SERVER_GONE) == 0) { list_remove(&sp->s_rcsync_list, rcp); cv_broadcast(&sp->s_rcsync_cv); } mutex_exit(&sp->s_rcsync_lock); mutex_exit(&sp->s_lock); nfs4_server_rele(sp); kmem_free(rcp, sizeof (*rcp)); } static void nfs4_recall_sync_wait(nfs4_server_t *sp) { uint64_t seq; mutex_enter(&sp->s_rcsync_lock); seq = sp->s_rcsync_seq; while (!list_is_empty(&sp->s_rcsync_list)) { nfs4_rcsync_t *rcp = list_head(&sp->s_rcsync_list); if (rcp->rs_seq >= seq) break; cv_wait(&sp->s_rcsync_cv, &sp->s_rcsync_lock); } mutex_exit(&sp->s_rcsync_lock); }