1 /* 2 * fs/nfs/nfs4proc.c 3 * 4 * Client-side procedure declarations for NFSv4. 5 * 6 * Copyright (c) 2002 The Regents of the University of Michigan. 7 * All rights reserved. 8 * 9 * Kendrick Smith <kmsmith@umich.edu> 10 * Andy Adamson <andros@umich.edu> 11 * 12 * Redistribution and use in source and binary forms, with or without 13 * modification, are permitted provided that the following conditions 14 * are met: 15 * 16 * 1. Redistributions of source code must retain the above copyright 17 * notice, this list of conditions and the following disclaimer. 18 * 2. Redistributions in binary form must reproduce the above copyright 19 * notice, this list of conditions and the following disclaimer in the 20 * documentation and/or other materials provided with the distribution. 21 * 3. Neither the name of the University nor the names of its 22 * contributors may be used to endorse or promote products derived 23 * from this software without specific prior written permission. 24 * 25 * THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED 26 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF 27 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE 28 * DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 29 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 30 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 31 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR 32 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF 33 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING 34 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS 35 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 36 */ 37 38 #include <linux/mm.h> 39 #include <linux/delay.h> 40 #include <linux/errno.h> 41 #include <linux/string.h> 42 #include <linux/slab.h> 43 #include <linux/sunrpc/clnt.h> 44 #include <linux/sunrpc/gss_api.h> 45 #include <linux/nfs.h> 46 #include <linux/nfs4.h> 47 #include <linux/nfs_fs.h> 48 #include <linux/nfs_page.h> 49 #include <linux/nfs_mount.h> 50 #include <linux/namei.h> 51 #include <linux/mount.h> 52 #include <linux/module.h> 53 #include <linux/sunrpc/bc_xprt.h> 54 #include <linux/xattr.h> 55 #include <linux/utsname.h> 56 57 #include "nfs4_fs.h" 58 #include "delegation.h" 59 #include "internal.h" 60 #include "iostat.h" 61 #include "callback.h" 62 #include "pnfs.h" 63 64 #define NFSDBG_FACILITY NFSDBG_PROC 65 66 #define NFS4_POLL_RETRY_MIN (HZ/10) 67 #define NFS4_POLL_RETRY_MAX (15*HZ) 68 69 #define NFS4_MAX_LOOP_ON_RECOVER (10) 70 71 struct nfs4_opendata; 72 static int _nfs4_proc_open(struct nfs4_opendata *data); 73 static int _nfs4_recover_proc_open(struct nfs4_opendata *data); 74 static int nfs4_do_fsinfo(struct nfs_server *, struct nfs_fh *, struct nfs_fsinfo *); 75 static int nfs4_async_handle_error(struct rpc_task *, const struct nfs_server *, struct nfs4_state *); 76 static int _nfs4_proc_getattr(struct nfs_server *server, struct nfs_fh *fhandle, struct nfs_fattr *fattr); 77 static int nfs4_do_setattr(struct inode *inode, struct rpc_cred *cred, 78 struct nfs_fattr *fattr, struct iattr *sattr, 79 struct nfs4_state *state); 80 #ifdef CONFIG_NFS_V4_1 81 static int nfs41_test_stateid(struct nfs_server *, struct nfs4_state *); 82 static int nfs41_free_stateid(struct nfs_server *, struct nfs4_state *); 83 #endif 84 /* Prevent leaks of NFSv4 errors into userland */ 85 static int nfs4_map_errors(int err) 86 { 87 if (err >= -1000) 88 return err; 89 switch (err) { 90 case -NFS4ERR_RESOURCE: 91 return -EREMOTEIO; 92 case -NFS4ERR_WRONGSEC: 93 return -EPERM; 94 case -NFS4ERR_BADOWNER: 95 case -NFS4ERR_BADNAME: 96 return -EINVAL; 97 default: 98 dprintk("%s could not handle NFSv4 error %d\n", 99 __func__, -err); 100 break; 101 } 102 return -EIO; 103 } 104 105 /* 106 * This is our standard bitmap for GETATTR requests. 107 */ 108 const u32 nfs4_fattr_bitmap[2] = { 109 FATTR4_WORD0_TYPE 110 | FATTR4_WORD0_CHANGE 111 | FATTR4_WORD0_SIZE 112 | FATTR4_WORD0_FSID 113 | FATTR4_WORD0_FILEID, 114 FATTR4_WORD1_MODE 115 | FATTR4_WORD1_NUMLINKS 116 | FATTR4_WORD1_OWNER 117 | FATTR4_WORD1_OWNER_GROUP 118 | FATTR4_WORD1_RAWDEV 119 | FATTR4_WORD1_SPACE_USED 120 | FATTR4_WORD1_TIME_ACCESS 121 | FATTR4_WORD1_TIME_METADATA 122 | FATTR4_WORD1_TIME_MODIFY 123 }; 124 125 const u32 nfs4_statfs_bitmap[2] = { 126 FATTR4_WORD0_FILES_AVAIL 127 | FATTR4_WORD0_FILES_FREE 128 | FATTR4_WORD0_FILES_TOTAL, 129 FATTR4_WORD1_SPACE_AVAIL 130 | FATTR4_WORD1_SPACE_FREE 131 | FATTR4_WORD1_SPACE_TOTAL 132 }; 133 134 const u32 nfs4_pathconf_bitmap[2] = { 135 FATTR4_WORD0_MAXLINK 136 | FATTR4_WORD0_MAXNAME, 137 0 138 }; 139 140 const u32 nfs4_fsinfo_bitmap[3] = { FATTR4_WORD0_MAXFILESIZE 141 | FATTR4_WORD0_MAXREAD 142 | FATTR4_WORD0_MAXWRITE 143 | FATTR4_WORD0_LEASE_TIME, 144 FATTR4_WORD1_TIME_DELTA 145 | FATTR4_WORD1_FS_LAYOUT_TYPES, 146 FATTR4_WORD2_LAYOUT_BLKSIZE 147 }; 148 149 const u32 nfs4_fs_locations_bitmap[2] = { 150 FATTR4_WORD0_TYPE 151 | FATTR4_WORD0_CHANGE 152 | FATTR4_WORD0_SIZE 153 | FATTR4_WORD0_FSID 154 | FATTR4_WORD0_FILEID 155 | FATTR4_WORD0_FS_LOCATIONS, 156 FATTR4_WORD1_MODE 157 | FATTR4_WORD1_NUMLINKS 158 | FATTR4_WORD1_OWNER 159 | FATTR4_WORD1_OWNER_GROUP 160 | FATTR4_WORD1_RAWDEV 161 | FATTR4_WORD1_SPACE_USED 162 | FATTR4_WORD1_TIME_ACCESS 163 | FATTR4_WORD1_TIME_METADATA 164 | FATTR4_WORD1_TIME_MODIFY 165 | FATTR4_WORD1_MOUNTED_ON_FILEID 166 }; 167 168 static void nfs4_setup_readdir(u64 cookie, __be32 *verifier, struct dentry *dentry, 169 struct nfs4_readdir_arg *readdir) 170 { 171 __be32 *start, *p; 172 173 BUG_ON(readdir->count < 80); 174 if (cookie > 2) { 175 readdir->cookie = cookie; 176 memcpy(&readdir->verifier, verifier, sizeof(readdir->verifier)); 177 return; 178 } 179 180 readdir->cookie = 0; 181 memset(&readdir->verifier, 0, sizeof(readdir->verifier)); 182 if (cookie == 2) 183 return; 184 185 /* 186 * NFSv4 servers do not return entries for '.' and '..' 187 * Therefore, we fake these entries here. We let '.' 188 * have cookie 0 and '..' have cookie 1. Note that 189 * when talking to the server, we always send cookie 0 190 * instead of 1 or 2. 191 */ 192 start = p = kmap_atomic(*readdir->pages, KM_USER0); 193 194 if (cookie == 0) { 195 *p++ = xdr_one; /* next */ 196 *p++ = xdr_zero; /* cookie, first word */ 197 *p++ = xdr_one; /* cookie, second word */ 198 *p++ = xdr_one; /* entry len */ 199 memcpy(p, ".\0\0\0", 4); /* entry */ 200 p++; 201 *p++ = xdr_one; /* bitmap length */ 202 *p++ = htonl(FATTR4_WORD0_FILEID); /* bitmap */ 203 *p++ = htonl(8); /* attribute buffer length */ 204 p = xdr_encode_hyper(p, NFS_FILEID(dentry->d_inode)); 205 } 206 207 *p++ = xdr_one; /* next */ 208 *p++ = xdr_zero; /* cookie, first word */ 209 *p++ = xdr_two; /* cookie, second word */ 210 *p++ = xdr_two; /* entry len */ 211 memcpy(p, "..\0\0", 4); /* entry */ 212 p++; 213 *p++ = xdr_one; /* bitmap length */ 214 *p++ = htonl(FATTR4_WORD0_FILEID); /* bitmap */ 215 *p++ = htonl(8); /* attribute buffer length */ 216 p = xdr_encode_hyper(p, NFS_FILEID(dentry->d_parent->d_inode)); 217 218 readdir->pgbase = (char *)p - (char *)start; 219 readdir->count -= readdir->pgbase; 220 kunmap_atomic(start, KM_USER0); 221 } 222 223 static int nfs4_wait_clnt_recover(struct nfs_client *clp) 224 { 225 int res; 226 227 might_sleep(); 228 229 res = wait_on_bit(&clp->cl_state, NFS4CLNT_MANAGER_RUNNING, 230 nfs_wait_bit_killable, TASK_KILLABLE); 231 return res; 232 } 233 234 static int nfs4_delay(struct rpc_clnt *clnt, long *timeout) 235 { 236 int res = 0; 237 238 might_sleep(); 239 240 if (*timeout <= 0) 241 *timeout = NFS4_POLL_RETRY_MIN; 242 if (*timeout > NFS4_POLL_RETRY_MAX) 243 *timeout = NFS4_POLL_RETRY_MAX; 244 schedule_timeout_killable(*timeout); 245 if (fatal_signal_pending(current)) 246 res = -ERESTARTSYS; 247 *timeout <<= 1; 248 return res; 249 } 250 251 /* This is the error handling routine for processes that are allowed 252 * to sleep. 253 */ 254 static int nfs4_handle_exception(struct nfs_server *server, int errorcode, struct nfs4_exception *exception) 255 { 256 struct nfs_client *clp = server->nfs_client; 257 struct nfs4_state *state = exception->state; 258 int ret = errorcode; 259 260 exception->retry = 0; 261 switch(errorcode) { 262 case 0: 263 return 0; 264 case -NFS4ERR_ADMIN_REVOKED: 265 case -NFS4ERR_BAD_STATEID: 266 case -NFS4ERR_OPENMODE: 267 if (state == NULL) 268 break; 269 nfs4_schedule_stateid_recovery(server, state); 270 goto wait_on_recovery; 271 case -NFS4ERR_EXPIRED: 272 if (state != NULL) 273 nfs4_schedule_stateid_recovery(server, state); 274 case -NFS4ERR_STALE_STATEID: 275 case -NFS4ERR_STALE_CLIENTID: 276 nfs4_schedule_lease_recovery(clp); 277 goto wait_on_recovery; 278 #if defined(CONFIG_NFS_V4_1) 279 case -NFS4ERR_BADSESSION: 280 case -NFS4ERR_BADSLOT: 281 case -NFS4ERR_BAD_HIGH_SLOT: 282 case -NFS4ERR_CONN_NOT_BOUND_TO_SESSION: 283 case -NFS4ERR_DEADSESSION: 284 case -NFS4ERR_SEQ_FALSE_RETRY: 285 case -NFS4ERR_SEQ_MISORDERED: 286 dprintk("%s ERROR: %d Reset session\n", __func__, 287 errorcode); 288 nfs4_schedule_session_recovery(clp->cl_session); 289 exception->retry = 1; 290 break; 291 #endif /* defined(CONFIG_NFS_V4_1) */ 292 case -NFS4ERR_FILE_OPEN: 293 if (exception->timeout > HZ) { 294 /* We have retried a decent amount, time to 295 * fail 296 */ 297 ret = -EBUSY; 298 break; 299 } 300 case -NFS4ERR_GRACE: 301 case -NFS4ERR_DELAY: 302 case -EKEYEXPIRED: 303 ret = nfs4_delay(server->client, &exception->timeout); 304 if (ret != 0) 305 break; 306 case -NFS4ERR_RETRY_UNCACHED_REP: 307 case -NFS4ERR_OLD_STATEID: 308 exception->retry = 1; 309 break; 310 case -NFS4ERR_BADOWNER: 311 /* The following works around a Linux server bug! */ 312 case -NFS4ERR_BADNAME: 313 if (server->caps & NFS_CAP_UIDGID_NOMAP) { 314 server->caps &= ~NFS_CAP_UIDGID_NOMAP; 315 exception->retry = 1; 316 printk(KERN_WARNING "NFS: v4 server %s " 317 "does not accept raw " 318 "uid/gids. " 319 "Reenabling the idmapper.\n", 320 server->nfs_client->cl_hostname); 321 } 322 } 323 /* We failed to handle the error */ 324 return nfs4_map_errors(ret); 325 wait_on_recovery: 326 ret = nfs4_wait_clnt_recover(clp); 327 if (ret == 0) 328 exception->retry = 1; 329 return ret; 330 } 331 332 333 static void do_renew_lease(struct nfs_client *clp, unsigned long timestamp) 334 { 335 spin_lock(&clp->cl_lock); 336 if (time_before(clp->cl_last_renewal,timestamp)) 337 clp->cl_last_renewal = timestamp; 338 spin_unlock(&clp->cl_lock); 339 } 340 341 static void renew_lease(const struct nfs_server *server, unsigned long timestamp) 342 { 343 do_renew_lease(server->nfs_client, timestamp); 344 } 345 346 #if defined(CONFIG_NFS_V4_1) 347 348 /* 349 * nfs4_free_slot - free a slot and efficiently update slot table. 350 * 351 * freeing a slot is trivially done by clearing its respective bit 352 * in the bitmap. 353 * If the freed slotid equals highest_used_slotid we want to update it 354 * so that the server would be able to size down the slot table if needed, 355 * otherwise we know that the highest_used_slotid is still in use. 356 * When updating highest_used_slotid there may be "holes" in the bitmap 357 * so we need to scan down from highest_used_slotid to 0 looking for the now 358 * highest slotid in use. 359 * If none found, highest_used_slotid is set to -1. 360 * 361 * Must be called while holding tbl->slot_tbl_lock 362 */ 363 static void 364 nfs4_free_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *free_slot) 365 { 366 int free_slotid = free_slot - tbl->slots; 367 int slotid = free_slotid; 368 369 BUG_ON(slotid < 0 || slotid >= NFS4_MAX_SLOT_TABLE); 370 /* clear used bit in bitmap */ 371 __clear_bit(slotid, tbl->used_slots); 372 373 /* update highest_used_slotid when it is freed */ 374 if (slotid == tbl->highest_used_slotid) { 375 slotid = find_last_bit(tbl->used_slots, tbl->max_slots); 376 if (slotid < tbl->max_slots) 377 tbl->highest_used_slotid = slotid; 378 else 379 tbl->highest_used_slotid = -1; 380 } 381 dprintk("%s: free_slotid %u highest_used_slotid %d\n", __func__, 382 free_slotid, tbl->highest_used_slotid); 383 } 384 385 /* 386 * Signal state manager thread if session fore channel is drained 387 */ 388 static void nfs4_check_drain_fc_complete(struct nfs4_session *ses) 389 { 390 struct rpc_task *task; 391 392 if (!test_bit(NFS4_SESSION_DRAINING, &ses->session_state)) { 393 task = rpc_wake_up_next(&ses->fc_slot_table.slot_tbl_waitq); 394 if (task) 395 rpc_task_set_priority(task, RPC_PRIORITY_PRIVILEGED); 396 return; 397 } 398 399 if (ses->fc_slot_table.highest_used_slotid != -1) 400 return; 401 402 dprintk("%s COMPLETE: Session Fore Channel Drained\n", __func__); 403 complete(&ses->fc_slot_table.complete); 404 } 405 406 /* 407 * Signal state manager thread if session back channel is drained 408 */ 409 void nfs4_check_drain_bc_complete(struct nfs4_session *ses) 410 { 411 if (!test_bit(NFS4_SESSION_DRAINING, &ses->session_state) || 412 ses->bc_slot_table.highest_used_slotid != -1) 413 return; 414 dprintk("%s COMPLETE: Session Back Channel Drained\n", __func__); 415 complete(&ses->bc_slot_table.complete); 416 } 417 418 static void nfs41_sequence_free_slot(struct nfs4_sequence_res *res) 419 { 420 struct nfs4_slot_table *tbl; 421 422 tbl = &res->sr_session->fc_slot_table; 423 if (!res->sr_slot) { 424 /* just wake up the next guy waiting since 425 * we may have not consumed a slot after all */ 426 dprintk("%s: No slot\n", __func__); 427 return; 428 } 429 430 spin_lock(&tbl->slot_tbl_lock); 431 nfs4_free_slot(tbl, res->sr_slot); 432 nfs4_check_drain_fc_complete(res->sr_session); 433 spin_unlock(&tbl->slot_tbl_lock); 434 res->sr_slot = NULL; 435 } 436 437 static int nfs41_sequence_done(struct rpc_task *task, struct nfs4_sequence_res *res) 438 { 439 unsigned long timestamp; 440 struct nfs_client *clp; 441 442 /* 443 * sr_status remains 1 if an RPC level error occurred. The server 444 * may or may not have processed the sequence operation.. 445 * Proceed as if the server received and processed the sequence 446 * operation. 447 */ 448 if (res->sr_status == 1) 449 res->sr_status = NFS_OK; 450 451 /* don't increment the sequence number if the task wasn't sent */ 452 if (!RPC_WAS_SENT(task)) 453 goto out; 454 455 /* Check the SEQUENCE operation status */ 456 switch (res->sr_status) { 457 case 0: 458 /* Update the slot's sequence and clientid lease timer */ 459 ++res->sr_slot->seq_nr; 460 timestamp = res->sr_renewal_time; 461 clp = res->sr_session->clp; 462 do_renew_lease(clp, timestamp); 463 /* Check sequence flags */ 464 if (res->sr_status_flags != 0) 465 nfs4_schedule_lease_recovery(clp); 466 break; 467 case -NFS4ERR_DELAY: 468 /* The server detected a resend of the RPC call and 469 * returned NFS4ERR_DELAY as per Section 2.10.6.2 470 * of RFC5661. 471 */ 472 dprintk("%s: slot=%td seq=%d: Operation in progress\n", 473 __func__, 474 res->sr_slot - res->sr_session->fc_slot_table.slots, 475 res->sr_slot->seq_nr); 476 goto out_retry; 477 default: 478 /* Just update the slot sequence no. */ 479 ++res->sr_slot->seq_nr; 480 } 481 out: 482 /* The session may be reset by one of the error handlers. */ 483 dprintk("%s: Error %d free the slot \n", __func__, res->sr_status); 484 nfs41_sequence_free_slot(res); 485 return 1; 486 out_retry: 487 if (!rpc_restart_call(task)) 488 goto out; 489 rpc_delay(task, NFS4_POLL_RETRY_MAX); 490 return 0; 491 } 492 493 static int nfs4_sequence_done(struct rpc_task *task, 494 struct nfs4_sequence_res *res) 495 { 496 if (res->sr_session == NULL) 497 return 1; 498 return nfs41_sequence_done(task, res); 499 } 500 501 /* 502 * nfs4_find_slot - efficiently look for a free slot 503 * 504 * nfs4_find_slot looks for an unset bit in the used_slots bitmap. 505 * If found, we mark the slot as used, update the highest_used_slotid, 506 * and respectively set up the sequence operation args. 507 * The slot number is returned if found, or NFS4_MAX_SLOT_TABLE otherwise. 508 * 509 * Note: must be called with under the slot_tbl_lock. 510 */ 511 static u8 512 nfs4_find_slot(struct nfs4_slot_table *tbl) 513 { 514 int slotid; 515 u8 ret_id = NFS4_MAX_SLOT_TABLE; 516 BUILD_BUG_ON((u8)NFS4_MAX_SLOT_TABLE != (int)NFS4_MAX_SLOT_TABLE); 517 518 dprintk("--> %s used_slots=%04lx highest_used=%d max_slots=%d\n", 519 __func__, tbl->used_slots[0], tbl->highest_used_slotid, 520 tbl->max_slots); 521 slotid = find_first_zero_bit(tbl->used_slots, tbl->max_slots); 522 if (slotid >= tbl->max_slots) 523 goto out; 524 __set_bit(slotid, tbl->used_slots); 525 if (slotid > tbl->highest_used_slotid) 526 tbl->highest_used_slotid = slotid; 527 ret_id = slotid; 528 out: 529 dprintk("<-- %s used_slots=%04lx highest_used=%d slotid=%d \n", 530 __func__, tbl->used_slots[0], tbl->highest_used_slotid, ret_id); 531 return ret_id; 532 } 533 534 int nfs41_setup_sequence(struct nfs4_session *session, 535 struct nfs4_sequence_args *args, 536 struct nfs4_sequence_res *res, 537 int cache_reply, 538 struct rpc_task *task) 539 { 540 struct nfs4_slot *slot; 541 struct nfs4_slot_table *tbl; 542 u8 slotid; 543 544 dprintk("--> %s\n", __func__); 545 /* slot already allocated? */ 546 if (res->sr_slot != NULL) 547 return 0; 548 549 tbl = &session->fc_slot_table; 550 551 spin_lock(&tbl->slot_tbl_lock); 552 if (test_bit(NFS4_SESSION_DRAINING, &session->session_state) && 553 !rpc_task_has_priority(task, RPC_PRIORITY_PRIVILEGED)) { 554 /* 555 * The state manager will wait until the slot table is empty. 556 * Schedule the reset thread 557 */ 558 rpc_sleep_on(&tbl->slot_tbl_waitq, task, NULL); 559 spin_unlock(&tbl->slot_tbl_lock); 560 dprintk("%s Schedule Session Reset\n", __func__); 561 return -EAGAIN; 562 } 563 564 if (!rpc_queue_empty(&tbl->slot_tbl_waitq) && 565 !rpc_task_has_priority(task, RPC_PRIORITY_PRIVILEGED)) { 566 rpc_sleep_on(&tbl->slot_tbl_waitq, task, NULL); 567 spin_unlock(&tbl->slot_tbl_lock); 568 dprintk("%s enforce FIFO order\n", __func__); 569 return -EAGAIN; 570 } 571 572 slotid = nfs4_find_slot(tbl); 573 if (slotid == NFS4_MAX_SLOT_TABLE) { 574 rpc_sleep_on(&tbl->slot_tbl_waitq, task, NULL); 575 spin_unlock(&tbl->slot_tbl_lock); 576 dprintk("<-- %s: no free slots\n", __func__); 577 return -EAGAIN; 578 } 579 spin_unlock(&tbl->slot_tbl_lock); 580 581 rpc_task_set_priority(task, RPC_PRIORITY_NORMAL); 582 slot = tbl->slots + slotid; 583 args->sa_session = session; 584 args->sa_slotid = slotid; 585 args->sa_cache_this = cache_reply; 586 587 dprintk("<-- %s slotid=%d seqid=%d\n", __func__, slotid, slot->seq_nr); 588 589 res->sr_session = session; 590 res->sr_slot = slot; 591 res->sr_renewal_time = jiffies; 592 res->sr_status_flags = 0; 593 /* 594 * sr_status is only set in decode_sequence, and so will remain 595 * set to 1 if an rpc level failure occurs. 596 */ 597 res->sr_status = 1; 598 return 0; 599 } 600 EXPORT_SYMBOL_GPL(nfs41_setup_sequence); 601 602 int nfs4_setup_sequence(const struct nfs_server *server, 603 struct nfs4_sequence_args *args, 604 struct nfs4_sequence_res *res, 605 int cache_reply, 606 struct rpc_task *task) 607 { 608 struct nfs4_session *session = nfs4_get_session(server); 609 int ret = 0; 610 611 if (session == NULL) { 612 args->sa_session = NULL; 613 res->sr_session = NULL; 614 goto out; 615 } 616 617 dprintk("--> %s clp %p session %p sr_slot %td\n", 618 __func__, session->clp, session, res->sr_slot ? 619 res->sr_slot - session->fc_slot_table.slots : -1); 620 621 ret = nfs41_setup_sequence(session, args, res, cache_reply, 622 task); 623 out: 624 dprintk("<-- %s status=%d\n", __func__, ret); 625 return ret; 626 } 627 628 struct nfs41_call_sync_data { 629 const struct nfs_server *seq_server; 630 struct nfs4_sequence_args *seq_args; 631 struct nfs4_sequence_res *seq_res; 632 int cache_reply; 633 }; 634 635 static void nfs41_call_sync_prepare(struct rpc_task *task, void *calldata) 636 { 637 struct nfs41_call_sync_data *data = calldata; 638 639 dprintk("--> %s data->seq_server %p\n", __func__, data->seq_server); 640 641 if (nfs4_setup_sequence(data->seq_server, data->seq_args, 642 data->seq_res, data->cache_reply, task)) 643 return; 644 rpc_call_start(task); 645 } 646 647 static void nfs41_call_priv_sync_prepare(struct rpc_task *task, void *calldata) 648 { 649 rpc_task_set_priority(task, RPC_PRIORITY_PRIVILEGED); 650 nfs41_call_sync_prepare(task, calldata); 651 } 652 653 static void nfs41_call_sync_done(struct rpc_task *task, void *calldata) 654 { 655 struct nfs41_call_sync_data *data = calldata; 656 657 nfs41_sequence_done(task, data->seq_res); 658 } 659 660 struct rpc_call_ops nfs41_call_sync_ops = { 661 .rpc_call_prepare = nfs41_call_sync_prepare, 662 .rpc_call_done = nfs41_call_sync_done, 663 }; 664 665 struct rpc_call_ops nfs41_call_priv_sync_ops = { 666 .rpc_call_prepare = nfs41_call_priv_sync_prepare, 667 .rpc_call_done = nfs41_call_sync_done, 668 }; 669 670 static int nfs4_call_sync_sequence(struct rpc_clnt *clnt, 671 struct nfs_server *server, 672 struct rpc_message *msg, 673 struct nfs4_sequence_args *args, 674 struct nfs4_sequence_res *res, 675 int cache_reply, 676 int privileged) 677 { 678 int ret; 679 struct rpc_task *task; 680 struct nfs41_call_sync_data data = { 681 .seq_server = server, 682 .seq_args = args, 683 .seq_res = res, 684 .cache_reply = cache_reply, 685 }; 686 struct rpc_task_setup task_setup = { 687 .rpc_client = clnt, 688 .rpc_message = msg, 689 .callback_ops = &nfs41_call_sync_ops, 690 .callback_data = &data 691 }; 692 693 res->sr_slot = NULL; 694 if (privileged) 695 task_setup.callback_ops = &nfs41_call_priv_sync_ops; 696 task = rpc_run_task(&task_setup); 697 if (IS_ERR(task)) 698 ret = PTR_ERR(task); 699 else { 700 ret = task->tk_status; 701 rpc_put_task(task); 702 } 703 return ret; 704 } 705 706 int _nfs4_call_sync_session(struct rpc_clnt *clnt, 707 struct nfs_server *server, 708 struct rpc_message *msg, 709 struct nfs4_sequence_args *args, 710 struct nfs4_sequence_res *res, 711 int cache_reply) 712 { 713 return nfs4_call_sync_sequence(clnt, server, msg, args, res, cache_reply, 0); 714 } 715 716 #else 717 static int nfs4_sequence_done(struct rpc_task *task, 718 struct nfs4_sequence_res *res) 719 { 720 return 1; 721 } 722 #endif /* CONFIG_NFS_V4_1 */ 723 724 int _nfs4_call_sync(struct rpc_clnt *clnt, 725 struct nfs_server *server, 726 struct rpc_message *msg, 727 struct nfs4_sequence_args *args, 728 struct nfs4_sequence_res *res, 729 int cache_reply) 730 { 731 args->sa_session = res->sr_session = NULL; 732 return rpc_call_sync(clnt, msg, 0); 733 } 734 735 static inline 736 int nfs4_call_sync(struct rpc_clnt *clnt, 737 struct nfs_server *server, 738 struct rpc_message *msg, 739 struct nfs4_sequence_args *args, 740 struct nfs4_sequence_res *res, 741 int cache_reply) 742 { 743 return server->nfs_client->cl_mvops->call_sync(clnt, server, msg, 744 args, res, cache_reply); 745 } 746 747 static void update_changeattr(struct inode *dir, struct nfs4_change_info *cinfo) 748 { 749 struct nfs_inode *nfsi = NFS_I(dir); 750 751 spin_lock(&dir->i_lock); 752 nfsi->cache_validity |= NFS_INO_INVALID_ATTR|NFS_INO_REVAL_PAGECACHE|NFS_INO_INVALID_DATA; 753 if (!cinfo->atomic || cinfo->before != dir->i_version) 754 nfs_force_lookup_revalidate(dir); 755 dir->i_version = cinfo->after; 756 spin_unlock(&dir->i_lock); 757 } 758 759 struct nfs4_opendata { 760 struct kref kref; 761 struct nfs_openargs o_arg; 762 struct nfs_openres o_res; 763 struct nfs_open_confirmargs c_arg; 764 struct nfs_open_confirmres c_res; 765 struct nfs_fattr f_attr; 766 struct nfs_fattr dir_attr; 767 struct dentry *dir; 768 struct dentry *dentry; 769 struct nfs4_state_owner *owner; 770 struct nfs4_state *state; 771 struct iattr attrs; 772 unsigned long timestamp; 773 unsigned int rpc_done : 1; 774 int rpc_status; 775 int cancelled; 776 }; 777 778 779 static void nfs4_init_opendata_res(struct nfs4_opendata *p) 780 { 781 p->o_res.f_attr = &p->f_attr; 782 p->o_res.dir_attr = &p->dir_attr; 783 p->o_res.seqid = p->o_arg.seqid; 784 p->c_res.seqid = p->c_arg.seqid; 785 p->o_res.server = p->o_arg.server; 786 nfs_fattr_init(&p->f_attr); 787 nfs_fattr_init(&p->dir_attr); 788 } 789 790 static struct nfs4_opendata *nfs4_opendata_alloc(struct dentry *dentry, 791 struct nfs4_state_owner *sp, fmode_t fmode, int flags, 792 const struct iattr *attrs, 793 gfp_t gfp_mask) 794 { 795 struct dentry *parent = dget_parent(dentry); 796 struct inode *dir = parent->d_inode; 797 struct nfs_server *server = NFS_SERVER(dir); 798 struct nfs4_opendata *p; 799 800 p = kzalloc(sizeof(*p), gfp_mask); 801 if (p == NULL) 802 goto err; 803 p->o_arg.seqid = nfs_alloc_seqid(&sp->so_seqid, gfp_mask); 804 if (p->o_arg.seqid == NULL) 805 goto err_free; 806 nfs_sb_active(dentry->d_sb); 807 p->dentry = dget(dentry); 808 p->dir = parent; 809 p->owner = sp; 810 atomic_inc(&sp->so_count); 811 p->o_arg.fh = NFS_FH(dir); 812 p->o_arg.open_flags = flags; 813 p->o_arg.fmode = fmode & (FMODE_READ|FMODE_WRITE); 814 p->o_arg.clientid = server->nfs_client->cl_clientid; 815 p->o_arg.id = sp->so_owner_id.id; 816 p->o_arg.name = &dentry->d_name; 817 p->o_arg.server = server; 818 p->o_arg.bitmask = server->attr_bitmask; 819 p->o_arg.claim = NFS4_OPEN_CLAIM_NULL; 820 if (flags & O_CREAT) { 821 u32 *s; 822 823 p->o_arg.u.attrs = &p->attrs; 824 memcpy(&p->attrs, attrs, sizeof(p->attrs)); 825 s = (u32 *) p->o_arg.u.verifier.data; 826 s[0] = jiffies; 827 s[1] = current->pid; 828 } 829 p->c_arg.fh = &p->o_res.fh; 830 p->c_arg.stateid = &p->o_res.stateid; 831 p->c_arg.seqid = p->o_arg.seqid; 832 nfs4_init_opendata_res(p); 833 kref_init(&p->kref); 834 return p; 835 err_free: 836 kfree(p); 837 err: 838 dput(parent); 839 return NULL; 840 } 841 842 static void nfs4_opendata_free(struct kref *kref) 843 { 844 struct nfs4_opendata *p = container_of(kref, 845 struct nfs4_opendata, kref); 846 struct super_block *sb = p->dentry->d_sb; 847 848 nfs_free_seqid(p->o_arg.seqid); 849 if (p->state != NULL) 850 nfs4_put_open_state(p->state); 851 nfs4_put_state_owner(p->owner); 852 dput(p->dir); 853 dput(p->dentry); 854 nfs_sb_deactive(sb); 855 kfree(p); 856 } 857 858 static void nfs4_opendata_put(struct nfs4_opendata *p) 859 { 860 if (p != NULL) 861 kref_put(&p->kref, nfs4_opendata_free); 862 } 863 864 static int nfs4_wait_for_completion_rpc_task(struct rpc_task *task) 865 { 866 int ret; 867 868 ret = rpc_wait_for_completion_task(task); 869 return ret; 870 } 871 872 static int can_open_cached(struct nfs4_state *state, fmode_t mode, int open_mode) 873 { 874 int ret = 0; 875 876 if (open_mode & O_EXCL) 877 goto out; 878 switch (mode & (FMODE_READ|FMODE_WRITE)) { 879 case FMODE_READ: 880 ret |= test_bit(NFS_O_RDONLY_STATE, &state->flags) != 0 881 && state->n_rdonly != 0; 882 break; 883 case FMODE_WRITE: 884 ret |= test_bit(NFS_O_WRONLY_STATE, &state->flags) != 0 885 && state->n_wronly != 0; 886 break; 887 case FMODE_READ|FMODE_WRITE: 888 ret |= test_bit(NFS_O_RDWR_STATE, &state->flags) != 0 889 && state->n_rdwr != 0; 890 } 891 out: 892 return ret; 893 } 894 895 static int can_open_delegated(struct nfs_delegation *delegation, fmode_t fmode) 896 { 897 if ((delegation->type & fmode) != fmode) 898 return 0; 899 if (test_bit(NFS_DELEGATION_NEED_RECLAIM, &delegation->flags)) 900 return 0; 901 nfs_mark_delegation_referenced(delegation); 902 return 1; 903 } 904 905 static void update_open_stateflags(struct nfs4_state *state, fmode_t fmode) 906 { 907 switch (fmode) { 908 case FMODE_WRITE: 909 state->n_wronly++; 910 break; 911 case FMODE_READ: 912 state->n_rdonly++; 913 break; 914 case FMODE_READ|FMODE_WRITE: 915 state->n_rdwr++; 916 } 917 nfs4_state_set_mode_locked(state, state->state | fmode); 918 } 919 920 static void nfs_set_open_stateid_locked(struct nfs4_state *state, nfs4_stateid *stateid, fmode_t fmode) 921 { 922 if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) 923 memcpy(state->stateid.data, stateid->data, sizeof(state->stateid.data)); 924 memcpy(state->open_stateid.data, stateid->data, sizeof(state->open_stateid.data)); 925 switch (fmode) { 926 case FMODE_READ: 927 set_bit(NFS_O_RDONLY_STATE, &state->flags); 928 break; 929 case FMODE_WRITE: 930 set_bit(NFS_O_WRONLY_STATE, &state->flags); 931 break; 932 case FMODE_READ|FMODE_WRITE: 933 set_bit(NFS_O_RDWR_STATE, &state->flags); 934 } 935 } 936 937 static void nfs_set_open_stateid(struct nfs4_state *state, nfs4_stateid *stateid, fmode_t fmode) 938 { 939 write_seqlock(&state->seqlock); 940 nfs_set_open_stateid_locked(state, stateid, fmode); 941 write_sequnlock(&state->seqlock); 942 } 943 944 static void __update_open_stateid(struct nfs4_state *state, nfs4_stateid *open_stateid, const nfs4_stateid *deleg_stateid, fmode_t fmode) 945 { 946 /* 947 * Protect the call to nfs4_state_set_mode_locked and 948 * serialise the stateid update 949 */ 950 write_seqlock(&state->seqlock); 951 if (deleg_stateid != NULL) { 952 memcpy(state->stateid.data, deleg_stateid->data, sizeof(state->stateid.data)); 953 set_bit(NFS_DELEGATED_STATE, &state->flags); 954 } 955 if (open_stateid != NULL) 956 nfs_set_open_stateid_locked(state, open_stateid, fmode); 957 write_sequnlock(&state->seqlock); 958 spin_lock(&state->owner->so_lock); 959 update_open_stateflags(state, fmode); 960 spin_unlock(&state->owner->so_lock); 961 } 962 963 static int update_open_stateid(struct nfs4_state *state, nfs4_stateid *open_stateid, nfs4_stateid *delegation, fmode_t fmode) 964 { 965 struct nfs_inode *nfsi = NFS_I(state->inode); 966 struct nfs_delegation *deleg_cur; 967 int ret = 0; 968 969 fmode &= (FMODE_READ|FMODE_WRITE); 970 971 rcu_read_lock(); 972 deleg_cur = rcu_dereference(nfsi->delegation); 973 if (deleg_cur == NULL) 974 goto no_delegation; 975 976 spin_lock(&deleg_cur->lock); 977 if (nfsi->delegation != deleg_cur || 978 (deleg_cur->type & fmode) != fmode) 979 goto no_delegation_unlock; 980 981 if (delegation == NULL) 982 delegation = &deleg_cur->stateid; 983 else if (memcmp(deleg_cur->stateid.data, delegation->data, NFS4_STATEID_SIZE) != 0) 984 goto no_delegation_unlock; 985 986 nfs_mark_delegation_referenced(deleg_cur); 987 __update_open_stateid(state, open_stateid, &deleg_cur->stateid, fmode); 988 ret = 1; 989 no_delegation_unlock: 990 spin_unlock(&deleg_cur->lock); 991 no_delegation: 992 rcu_read_unlock(); 993 994 if (!ret && open_stateid != NULL) { 995 __update_open_stateid(state, open_stateid, NULL, fmode); 996 ret = 1; 997 } 998 999 return ret; 1000 } 1001 1002 1003 static void nfs4_return_incompatible_delegation(struct inode *inode, fmode_t fmode) 1004 { 1005 struct nfs_delegation *delegation; 1006 1007 rcu_read_lock(); 1008 delegation = rcu_dereference(NFS_I(inode)->delegation); 1009 if (delegation == NULL || (delegation->type & fmode) == fmode) { 1010 rcu_read_unlock(); 1011 return; 1012 } 1013 rcu_read_unlock(); 1014 nfs_inode_return_delegation(inode); 1015 } 1016 1017 static struct nfs4_state *nfs4_try_open_cached(struct nfs4_opendata *opendata) 1018 { 1019 struct nfs4_state *state = opendata->state; 1020 struct nfs_inode *nfsi = NFS_I(state->inode); 1021 struct nfs_delegation *delegation; 1022 int open_mode = opendata->o_arg.open_flags & O_EXCL; 1023 fmode_t fmode = opendata->o_arg.fmode; 1024 nfs4_stateid stateid; 1025 int ret = -EAGAIN; 1026 1027 for (;;) { 1028 if (can_open_cached(state, fmode, open_mode)) { 1029 spin_lock(&state->owner->so_lock); 1030 if (can_open_cached(state, fmode, open_mode)) { 1031 update_open_stateflags(state, fmode); 1032 spin_unlock(&state->owner->so_lock); 1033 goto out_return_state; 1034 } 1035 spin_unlock(&state->owner->so_lock); 1036 } 1037 rcu_read_lock(); 1038 delegation = rcu_dereference(nfsi->delegation); 1039 if (delegation == NULL || 1040 !can_open_delegated(delegation, fmode)) { 1041 rcu_read_unlock(); 1042 break; 1043 } 1044 /* Save the delegation */ 1045 memcpy(stateid.data, delegation->stateid.data, sizeof(stateid.data)); 1046 rcu_read_unlock(); 1047 ret = nfs_may_open(state->inode, state->owner->so_cred, open_mode); 1048 if (ret != 0) 1049 goto out; 1050 ret = -EAGAIN; 1051 1052 /* Try to update the stateid using the delegation */ 1053 if (update_open_stateid(state, NULL, &stateid, fmode)) 1054 goto out_return_state; 1055 } 1056 out: 1057 return ERR_PTR(ret); 1058 out_return_state: 1059 atomic_inc(&state->count); 1060 return state; 1061 } 1062 1063 static struct nfs4_state *nfs4_opendata_to_nfs4_state(struct nfs4_opendata *data) 1064 { 1065 struct inode *inode; 1066 struct nfs4_state *state = NULL; 1067 struct nfs_delegation *delegation; 1068 int ret; 1069 1070 if (!data->rpc_done) { 1071 state = nfs4_try_open_cached(data); 1072 goto out; 1073 } 1074 1075 ret = -EAGAIN; 1076 if (!(data->f_attr.valid & NFS_ATTR_FATTR)) 1077 goto err; 1078 inode = nfs_fhget(data->dir->d_sb, &data->o_res.fh, &data->f_attr); 1079 ret = PTR_ERR(inode); 1080 if (IS_ERR(inode)) 1081 goto err; 1082 ret = -ENOMEM; 1083 state = nfs4_get_open_state(inode, data->owner); 1084 if (state == NULL) 1085 goto err_put_inode; 1086 if (data->o_res.delegation_type != 0) { 1087 int delegation_flags = 0; 1088 1089 rcu_read_lock(); 1090 delegation = rcu_dereference(NFS_I(inode)->delegation); 1091 if (delegation) 1092 delegation_flags = delegation->flags; 1093 rcu_read_unlock(); 1094 if ((delegation_flags & 1UL<<NFS_DELEGATION_NEED_RECLAIM) == 0) 1095 nfs_inode_set_delegation(state->inode, 1096 data->owner->so_cred, 1097 &data->o_res); 1098 else 1099 nfs_inode_reclaim_delegation(state->inode, 1100 data->owner->so_cred, 1101 &data->o_res); 1102 } 1103 1104 update_open_stateid(state, &data->o_res.stateid, NULL, 1105 data->o_arg.fmode); 1106 iput(inode); 1107 out: 1108 return state; 1109 err_put_inode: 1110 iput(inode); 1111 err: 1112 return ERR_PTR(ret); 1113 } 1114 1115 static struct nfs_open_context *nfs4_state_find_open_context(struct nfs4_state *state) 1116 { 1117 struct nfs_inode *nfsi = NFS_I(state->inode); 1118 struct nfs_open_context *ctx; 1119 1120 spin_lock(&state->inode->i_lock); 1121 list_for_each_entry(ctx, &nfsi->open_files, list) { 1122 if (ctx->state != state) 1123 continue; 1124 get_nfs_open_context(ctx); 1125 spin_unlock(&state->inode->i_lock); 1126 return ctx; 1127 } 1128 spin_unlock(&state->inode->i_lock); 1129 return ERR_PTR(-ENOENT); 1130 } 1131 1132 static struct nfs4_opendata *nfs4_open_recoverdata_alloc(struct nfs_open_context *ctx, struct nfs4_state *state) 1133 { 1134 struct nfs4_opendata *opendata; 1135 1136 opendata = nfs4_opendata_alloc(ctx->dentry, state->owner, 0, 0, NULL, GFP_NOFS); 1137 if (opendata == NULL) 1138 return ERR_PTR(-ENOMEM); 1139 opendata->state = state; 1140 atomic_inc(&state->count); 1141 return opendata; 1142 } 1143 1144 static int nfs4_open_recover_helper(struct nfs4_opendata *opendata, fmode_t fmode, struct nfs4_state **res) 1145 { 1146 struct nfs4_state *newstate; 1147 int ret; 1148 1149 opendata->o_arg.open_flags = 0; 1150 opendata->o_arg.fmode = fmode; 1151 memset(&opendata->o_res, 0, sizeof(opendata->o_res)); 1152 memset(&opendata->c_res, 0, sizeof(opendata->c_res)); 1153 nfs4_init_opendata_res(opendata); 1154 ret = _nfs4_recover_proc_open(opendata); 1155 if (ret != 0) 1156 return ret; 1157 newstate = nfs4_opendata_to_nfs4_state(opendata); 1158 if (IS_ERR(newstate)) 1159 return PTR_ERR(newstate); 1160 nfs4_close_state(newstate, fmode); 1161 *res = newstate; 1162 return 0; 1163 } 1164 1165 static int nfs4_open_recover(struct nfs4_opendata *opendata, struct nfs4_state *state) 1166 { 1167 struct nfs4_state *newstate; 1168 int ret; 1169 1170 /* memory barrier prior to reading state->n_* */ 1171 clear_bit(NFS_DELEGATED_STATE, &state->flags); 1172 smp_rmb(); 1173 if (state->n_rdwr != 0) { 1174 clear_bit(NFS_O_RDWR_STATE, &state->flags); 1175 ret = nfs4_open_recover_helper(opendata, FMODE_READ|FMODE_WRITE, &newstate); 1176 if (ret != 0) 1177 return ret; 1178 if (newstate != state) 1179 return -ESTALE; 1180 } 1181 if (state->n_wronly != 0) { 1182 clear_bit(NFS_O_WRONLY_STATE, &state->flags); 1183 ret = nfs4_open_recover_helper(opendata, FMODE_WRITE, &newstate); 1184 if (ret != 0) 1185 return ret; 1186 if (newstate != state) 1187 return -ESTALE; 1188 } 1189 if (state->n_rdonly != 0) { 1190 clear_bit(NFS_O_RDONLY_STATE, &state->flags); 1191 ret = nfs4_open_recover_helper(opendata, FMODE_READ, &newstate); 1192 if (ret != 0) 1193 return ret; 1194 if (newstate != state) 1195 return -ESTALE; 1196 } 1197 /* 1198 * We may have performed cached opens for all three recoveries. 1199 * Check if we need to update the current stateid. 1200 */ 1201 if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0 && 1202 memcmp(state->stateid.data, state->open_stateid.data, sizeof(state->stateid.data)) != 0) { 1203 write_seqlock(&state->seqlock); 1204 if (test_bit(NFS_DELEGATED_STATE, &state->flags) == 0) 1205 memcpy(state->stateid.data, state->open_stateid.data, sizeof(state->stateid.data)); 1206 write_sequnlock(&state->seqlock); 1207 } 1208 return 0; 1209 } 1210 1211 /* 1212 * OPEN_RECLAIM: 1213 * reclaim state on the server after a reboot. 1214 */ 1215 static int _nfs4_do_open_reclaim(struct nfs_open_context *ctx, struct nfs4_state *state) 1216 { 1217 struct nfs_delegation *delegation; 1218 struct nfs4_opendata *opendata; 1219 fmode_t delegation_type = 0; 1220 int status; 1221 1222 opendata = nfs4_open_recoverdata_alloc(ctx, state); 1223 if (IS_ERR(opendata)) 1224 return PTR_ERR(opendata); 1225 opendata->o_arg.claim = NFS4_OPEN_CLAIM_PREVIOUS; 1226 opendata->o_arg.fh = NFS_FH(state->inode); 1227 rcu_read_lock(); 1228 delegation = rcu_dereference(NFS_I(state->inode)->delegation); 1229 if (delegation != NULL && test_bit(NFS_DELEGATION_NEED_RECLAIM, &delegation->flags) != 0) 1230 delegation_type = delegation->type; 1231 rcu_read_unlock(); 1232 opendata->o_arg.u.delegation_type = delegation_type; 1233 status = nfs4_open_recover(opendata, state); 1234 nfs4_opendata_put(opendata); 1235 return status; 1236 } 1237 1238 static int nfs4_do_open_reclaim(struct nfs_open_context *ctx, struct nfs4_state *state) 1239 { 1240 struct nfs_server *server = NFS_SERVER(state->inode); 1241 struct nfs4_exception exception = { }; 1242 int err; 1243 do { 1244 err = _nfs4_do_open_reclaim(ctx, state); 1245 if (err != -NFS4ERR_DELAY) 1246 break; 1247 nfs4_handle_exception(server, err, &exception); 1248 } while (exception.retry); 1249 return err; 1250 } 1251 1252 static int nfs4_open_reclaim(struct nfs4_state_owner *sp, struct nfs4_state *state) 1253 { 1254 struct nfs_open_context *ctx; 1255 int ret; 1256 1257 ctx = nfs4_state_find_open_context(state); 1258 if (IS_ERR(ctx)) 1259 return PTR_ERR(ctx); 1260 ret = nfs4_do_open_reclaim(ctx, state); 1261 put_nfs_open_context(ctx); 1262 return ret; 1263 } 1264 1265 static int _nfs4_open_delegation_recall(struct nfs_open_context *ctx, struct nfs4_state *state, const nfs4_stateid *stateid) 1266 { 1267 struct nfs4_opendata *opendata; 1268 int ret; 1269 1270 opendata = nfs4_open_recoverdata_alloc(ctx, state); 1271 if (IS_ERR(opendata)) 1272 return PTR_ERR(opendata); 1273 opendata->o_arg.claim = NFS4_OPEN_CLAIM_DELEGATE_CUR; 1274 memcpy(opendata->o_arg.u.delegation.data, stateid->data, 1275 sizeof(opendata->o_arg.u.delegation.data)); 1276 ret = nfs4_open_recover(opendata, state); 1277 nfs4_opendata_put(opendata); 1278 return ret; 1279 } 1280 1281 int nfs4_open_delegation_recall(struct nfs_open_context *ctx, struct nfs4_state *state, const nfs4_stateid *stateid) 1282 { 1283 struct nfs4_exception exception = { }; 1284 struct nfs_server *server = NFS_SERVER(state->inode); 1285 int err; 1286 do { 1287 err = _nfs4_open_delegation_recall(ctx, state, stateid); 1288 switch (err) { 1289 case 0: 1290 case -ENOENT: 1291 case -ESTALE: 1292 goto out; 1293 case -NFS4ERR_BADSESSION: 1294 case -NFS4ERR_BADSLOT: 1295 case -NFS4ERR_BAD_HIGH_SLOT: 1296 case -NFS4ERR_CONN_NOT_BOUND_TO_SESSION: 1297 case -NFS4ERR_DEADSESSION: 1298 nfs4_schedule_session_recovery(server->nfs_client->cl_session); 1299 goto out; 1300 case -NFS4ERR_STALE_CLIENTID: 1301 case -NFS4ERR_STALE_STATEID: 1302 case -NFS4ERR_EXPIRED: 1303 /* Don't recall a delegation if it was lost */ 1304 nfs4_schedule_lease_recovery(server->nfs_client); 1305 goto out; 1306 case -ERESTARTSYS: 1307 /* 1308 * The show must go on: exit, but mark the 1309 * stateid as needing recovery. 1310 */ 1311 case -NFS4ERR_ADMIN_REVOKED: 1312 case -NFS4ERR_BAD_STATEID: 1313 nfs4_schedule_stateid_recovery(server, state); 1314 case -EKEYEXPIRED: 1315 /* 1316 * User RPCSEC_GSS context has expired. 1317 * We cannot recover this stateid now, so 1318 * skip it and allow recovery thread to 1319 * proceed. 1320 */ 1321 case -ENOMEM: 1322 err = 0; 1323 goto out; 1324 } 1325 err = nfs4_handle_exception(server, err, &exception); 1326 } while (exception.retry); 1327 out: 1328 return err; 1329 } 1330 1331 static void nfs4_open_confirm_done(struct rpc_task *task, void *calldata) 1332 { 1333 struct nfs4_opendata *data = calldata; 1334 1335 data->rpc_status = task->tk_status; 1336 if (data->rpc_status == 0) { 1337 memcpy(data->o_res.stateid.data, data->c_res.stateid.data, 1338 sizeof(data->o_res.stateid.data)); 1339 nfs_confirm_seqid(&data->owner->so_seqid, 0); 1340 renew_lease(data->o_res.server, data->timestamp); 1341 data->rpc_done = 1; 1342 } 1343 } 1344 1345 static void nfs4_open_confirm_release(void *calldata) 1346 { 1347 struct nfs4_opendata *data = calldata; 1348 struct nfs4_state *state = NULL; 1349 1350 /* If this request hasn't been cancelled, do nothing */ 1351 if (data->cancelled == 0) 1352 goto out_free; 1353 /* In case of error, no cleanup! */ 1354 if (!data->rpc_done) 1355 goto out_free; 1356 state = nfs4_opendata_to_nfs4_state(data); 1357 if (!IS_ERR(state)) 1358 nfs4_close_state(state, data->o_arg.fmode); 1359 out_free: 1360 nfs4_opendata_put(data); 1361 } 1362 1363 static const struct rpc_call_ops nfs4_open_confirm_ops = { 1364 .rpc_call_done = nfs4_open_confirm_done, 1365 .rpc_release = nfs4_open_confirm_release, 1366 }; 1367 1368 /* 1369 * Note: On error, nfs4_proc_open_confirm will free the struct nfs4_opendata 1370 */ 1371 static int _nfs4_proc_open_confirm(struct nfs4_opendata *data) 1372 { 1373 struct nfs_server *server = NFS_SERVER(data->dir->d_inode); 1374 struct rpc_task *task; 1375 struct rpc_message msg = { 1376 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_OPEN_CONFIRM], 1377 .rpc_argp = &data->c_arg, 1378 .rpc_resp = &data->c_res, 1379 .rpc_cred = data->owner->so_cred, 1380 }; 1381 struct rpc_task_setup task_setup_data = { 1382 .rpc_client = server->client, 1383 .rpc_message = &msg, 1384 .callback_ops = &nfs4_open_confirm_ops, 1385 .callback_data = data, 1386 .workqueue = nfsiod_workqueue, 1387 .flags = RPC_TASK_ASYNC, 1388 }; 1389 int status; 1390 1391 kref_get(&data->kref); 1392 data->rpc_done = 0; 1393 data->rpc_status = 0; 1394 data->timestamp = jiffies; 1395 task = rpc_run_task(&task_setup_data); 1396 if (IS_ERR(task)) 1397 return PTR_ERR(task); 1398 status = nfs4_wait_for_completion_rpc_task(task); 1399 if (status != 0) { 1400 data->cancelled = 1; 1401 smp_wmb(); 1402 } else 1403 status = data->rpc_status; 1404 rpc_put_task(task); 1405 return status; 1406 } 1407 1408 static void nfs4_open_prepare(struct rpc_task *task, void *calldata) 1409 { 1410 struct nfs4_opendata *data = calldata; 1411 struct nfs4_state_owner *sp = data->owner; 1412 1413 if (nfs_wait_on_sequence(data->o_arg.seqid, task) != 0) 1414 return; 1415 /* 1416 * Check if we still need to send an OPEN call, or if we can use 1417 * a delegation instead. 1418 */ 1419 if (data->state != NULL) { 1420 struct nfs_delegation *delegation; 1421 1422 if (can_open_cached(data->state, data->o_arg.fmode, data->o_arg.open_flags)) 1423 goto out_no_action; 1424 rcu_read_lock(); 1425 delegation = rcu_dereference(NFS_I(data->state->inode)->delegation); 1426 if (delegation != NULL && 1427 test_bit(NFS_DELEGATION_NEED_RECLAIM, &delegation->flags) == 0) { 1428 rcu_read_unlock(); 1429 goto out_no_action; 1430 } 1431 rcu_read_unlock(); 1432 } 1433 /* Update sequence id. */ 1434 data->o_arg.id = sp->so_owner_id.id; 1435 data->o_arg.clientid = sp->so_server->nfs_client->cl_clientid; 1436 if (data->o_arg.claim == NFS4_OPEN_CLAIM_PREVIOUS) { 1437 task->tk_msg.rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_OPEN_NOATTR]; 1438 nfs_copy_fh(&data->o_res.fh, data->o_arg.fh); 1439 } 1440 data->timestamp = jiffies; 1441 if (nfs4_setup_sequence(data->o_arg.server, 1442 &data->o_arg.seq_args, 1443 &data->o_res.seq_res, 1, task)) 1444 return; 1445 rpc_call_start(task); 1446 return; 1447 out_no_action: 1448 task->tk_action = NULL; 1449 1450 } 1451 1452 static void nfs4_recover_open_prepare(struct rpc_task *task, void *calldata) 1453 { 1454 rpc_task_set_priority(task, RPC_PRIORITY_PRIVILEGED); 1455 nfs4_open_prepare(task, calldata); 1456 } 1457 1458 static void nfs4_open_done(struct rpc_task *task, void *calldata) 1459 { 1460 struct nfs4_opendata *data = calldata; 1461 1462 data->rpc_status = task->tk_status; 1463 1464 if (!nfs4_sequence_done(task, &data->o_res.seq_res)) 1465 return; 1466 1467 if (task->tk_status == 0) { 1468 switch (data->o_res.f_attr->mode & S_IFMT) { 1469 case S_IFREG: 1470 break; 1471 case S_IFLNK: 1472 data->rpc_status = -ELOOP; 1473 break; 1474 case S_IFDIR: 1475 data->rpc_status = -EISDIR; 1476 break; 1477 default: 1478 data->rpc_status = -ENOTDIR; 1479 } 1480 renew_lease(data->o_res.server, data->timestamp); 1481 if (!(data->o_res.rflags & NFS4_OPEN_RESULT_CONFIRM)) 1482 nfs_confirm_seqid(&data->owner->so_seqid, 0); 1483 } 1484 data->rpc_done = 1; 1485 } 1486 1487 static void nfs4_open_release(void *calldata) 1488 { 1489 struct nfs4_opendata *data = calldata; 1490 struct nfs4_state *state = NULL; 1491 1492 /* If this request hasn't been cancelled, do nothing */ 1493 if (data->cancelled == 0) 1494 goto out_free; 1495 /* In case of error, no cleanup! */ 1496 if (data->rpc_status != 0 || !data->rpc_done) 1497 goto out_free; 1498 /* In case we need an open_confirm, no cleanup! */ 1499 if (data->o_res.rflags & NFS4_OPEN_RESULT_CONFIRM) 1500 goto out_free; 1501 state = nfs4_opendata_to_nfs4_state(data); 1502 if (!IS_ERR(state)) 1503 nfs4_close_state(state, data->o_arg.fmode); 1504 out_free: 1505 nfs4_opendata_put(data); 1506 } 1507 1508 static const struct rpc_call_ops nfs4_open_ops = { 1509 .rpc_call_prepare = nfs4_open_prepare, 1510 .rpc_call_done = nfs4_open_done, 1511 .rpc_release = nfs4_open_release, 1512 }; 1513 1514 static const struct rpc_call_ops nfs4_recover_open_ops = { 1515 .rpc_call_prepare = nfs4_recover_open_prepare, 1516 .rpc_call_done = nfs4_open_done, 1517 .rpc_release = nfs4_open_release, 1518 }; 1519 1520 static int nfs4_run_open_task(struct nfs4_opendata *data, int isrecover) 1521 { 1522 struct inode *dir = data->dir->d_inode; 1523 struct nfs_server *server = NFS_SERVER(dir); 1524 struct nfs_openargs *o_arg = &data->o_arg; 1525 struct nfs_openres *o_res = &data->o_res; 1526 struct rpc_task *task; 1527 struct rpc_message msg = { 1528 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_OPEN], 1529 .rpc_argp = o_arg, 1530 .rpc_resp = o_res, 1531 .rpc_cred = data->owner->so_cred, 1532 }; 1533 struct rpc_task_setup task_setup_data = { 1534 .rpc_client = server->client, 1535 .rpc_message = &msg, 1536 .callback_ops = &nfs4_open_ops, 1537 .callback_data = data, 1538 .workqueue = nfsiod_workqueue, 1539 .flags = RPC_TASK_ASYNC, 1540 }; 1541 int status; 1542 1543 kref_get(&data->kref); 1544 data->rpc_done = 0; 1545 data->rpc_status = 0; 1546 data->cancelled = 0; 1547 if (isrecover) 1548 task_setup_data.callback_ops = &nfs4_recover_open_ops; 1549 task = rpc_run_task(&task_setup_data); 1550 if (IS_ERR(task)) 1551 return PTR_ERR(task); 1552 status = nfs4_wait_for_completion_rpc_task(task); 1553 if (status != 0) { 1554 data->cancelled = 1; 1555 smp_wmb(); 1556 } else 1557 status = data->rpc_status; 1558 rpc_put_task(task); 1559 1560 return status; 1561 } 1562 1563 static int _nfs4_recover_proc_open(struct nfs4_opendata *data) 1564 { 1565 struct inode *dir = data->dir->d_inode; 1566 struct nfs_openres *o_res = &data->o_res; 1567 int status; 1568 1569 status = nfs4_run_open_task(data, 1); 1570 if (status != 0 || !data->rpc_done) 1571 return status; 1572 1573 nfs_refresh_inode(dir, o_res->dir_attr); 1574 1575 if (o_res->rflags & NFS4_OPEN_RESULT_CONFIRM) { 1576 status = _nfs4_proc_open_confirm(data); 1577 if (status != 0) 1578 return status; 1579 } 1580 1581 return status; 1582 } 1583 1584 /* 1585 * Note: On error, nfs4_proc_open will free the struct nfs4_opendata 1586 */ 1587 static int _nfs4_proc_open(struct nfs4_opendata *data) 1588 { 1589 struct inode *dir = data->dir->d_inode; 1590 struct nfs_server *server = NFS_SERVER(dir); 1591 struct nfs_openargs *o_arg = &data->o_arg; 1592 struct nfs_openres *o_res = &data->o_res; 1593 int status; 1594 1595 status = nfs4_run_open_task(data, 0); 1596 if (!data->rpc_done) 1597 return status; 1598 if (status != 0) { 1599 if (status == -NFS4ERR_BADNAME && 1600 !(o_arg->open_flags & O_CREAT)) 1601 return -ENOENT; 1602 return status; 1603 } 1604 1605 if (o_arg->open_flags & O_CREAT) { 1606 update_changeattr(dir, &o_res->cinfo); 1607 nfs_post_op_update_inode(dir, o_res->dir_attr); 1608 } else 1609 nfs_refresh_inode(dir, o_res->dir_attr); 1610 if ((o_res->rflags & NFS4_OPEN_RESULT_LOCKTYPE_POSIX) == 0) 1611 server->caps &= ~NFS_CAP_POSIX_LOCK; 1612 if(o_res->rflags & NFS4_OPEN_RESULT_CONFIRM) { 1613 status = _nfs4_proc_open_confirm(data); 1614 if (status != 0) 1615 return status; 1616 } 1617 if (!(o_res->f_attr->valid & NFS_ATTR_FATTR)) 1618 _nfs4_proc_getattr(server, &o_res->fh, o_res->f_attr); 1619 return 0; 1620 } 1621 1622 static int nfs4_client_recover_expired_lease(struct nfs_client *clp) 1623 { 1624 unsigned int loop; 1625 int ret; 1626 1627 for (loop = NFS4_MAX_LOOP_ON_RECOVER; loop != 0; loop--) { 1628 ret = nfs4_wait_clnt_recover(clp); 1629 if (ret != 0) 1630 break; 1631 if (!test_bit(NFS4CLNT_LEASE_EXPIRED, &clp->cl_state) && 1632 !test_bit(NFS4CLNT_CHECK_LEASE,&clp->cl_state)) 1633 break; 1634 nfs4_schedule_state_manager(clp); 1635 ret = -EIO; 1636 } 1637 return ret; 1638 } 1639 1640 static int nfs4_recover_expired_lease(struct nfs_server *server) 1641 { 1642 return nfs4_client_recover_expired_lease(server->nfs_client); 1643 } 1644 1645 /* 1646 * OPEN_EXPIRED: 1647 * reclaim state on the server after a network partition. 1648 * Assumes caller holds the appropriate lock 1649 */ 1650 static int _nfs4_open_expired(struct nfs_open_context *ctx, struct nfs4_state *state) 1651 { 1652 struct nfs4_opendata *opendata; 1653 int ret; 1654 1655 opendata = nfs4_open_recoverdata_alloc(ctx, state); 1656 if (IS_ERR(opendata)) 1657 return PTR_ERR(opendata); 1658 ret = nfs4_open_recover(opendata, state); 1659 if (ret == -ESTALE) 1660 d_drop(ctx->dentry); 1661 nfs4_opendata_put(opendata); 1662 return ret; 1663 } 1664 1665 static int nfs4_do_open_expired(struct nfs_open_context *ctx, struct nfs4_state *state) 1666 { 1667 struct nfs_server *server = NFS_SERVER(state->inode); 1668 struct nfs4_exception exception = { }; 1669 int err; 1670 1671 do { 1672 err = _nfs4_open_expired(ctx, state); 1673 switch (err) { 1674 default: 1675 goto out; 1676 case -NFS4ERR_GRACE: 1677 case -NFS4ERR_DELAY: 1678 nfs4_handle_exception(server, err, &exception); 1679 err = 0; 1680 } 1681 } while (exception.retry); 1682 out: 1683 return err; 1684 } 1685 1686 static int nfs4_open_expired(struct nfs4_state_owner *sp, struct nfs4_state *state) 1687 { 1688 struct nfs_open_context *ctx; 1689 int ret; 1690 1691 ctx = nfs4_state_find_open_context(state); 1692 if (IS_ERR(ctx)) 1693 return PTR_ERR(ctx); 1694 ret = nfs4_do_open_expired(ctx, state); 1695 put_nfs_open_context(ctx); 1696 return ret; 1697 } 1698 1699 #if defined(CONFIG_NFS_V4_1) 1700 static int nfs41_open_expired(struct nfs4_state_owner *sp, struct nfs4_state *state) 1701 { 1702 int status; 1703 struct nfs_server *server = NFS_SERVER(state->inode); 1704 1705 status = nfs41_test_stateid(server, state); 1706 if (status == NFS_OK) 1707 return 0; 1708 nfs41_free_stateid(server, state); 1709 return nfs4_open_expired(sp, state); 1710 } 1711 #endif 1712 1713 /* 1714 * on an EXCLUSIVE create, the server should send back a bitmask with FATTR4-* 1715 * fields corresponding to attributes that were used to store the verifier. 1716 * Make sure we clobber those fields in the later setattr call 1717 */ 1718 static inline void nfs4_exclusive_attrset(struct nfs4_opendata *opendata, struct iattr *sattr) 1719 { 1720 if ((opendata->o_res.attrset[1] & FATTR4_WORD1_TIME_ACCESS) && 1721 !(sattr->ia_valid & ATTR_ATIME_SET)) 1722 sattr->ia_valid |= ATTR_ATIME; 1723 1724 if ((opendata->o_res.attrset[1] & FATTR4_WORD1_TIME_MODIFY) && 1725 !(sattr->ia_valid & ATTR_MTIME_SET)) 1726 sattr->ia_valid |= ATTR_MTIME; 1727 } 1728 1729 /* 1730 * Returns a referenced nfs4_state 1731 */ 1732 static int _nfs4_do_open(struct inode *dir, struct dentry *dentry, fmode_t fmode, int flags, struct iattr *sattr, struct rpc_cred *cred, struct nfs4_state **res) 1733 { 1734 struct nfs4_state_owner *sp; 1735 struct nfs4_state *state = NULL; 1736 struct nfs_server *server = NFS_SERVER(dir); 1737 struct nfs4_opendata *opendata; 1738 int status; 1739 1740 /* Protect against reboot recovery conflicts */ 1741 status = -ENOMEM; 1742 if (!(sp = nfs4_get_state_owner(server, cred))) { 1743 dprintk("nfs4_do_open: nfs4_get_state_owner failed!\n"); 1744 goto out_err; 1745 } 1746 status = nfs4_recover_expired_lease(server); 1747 if (status != 0) 1748 goto err_put_state_owner; 1749 if (dentry->d_inode != NULL) 1750 nfs4_return_incompatible_delegation(dentry->d_inode, fmode); 1751 status = -ENOMEM; 1752 opendata = nfs4_opendata_alloc(dentry, sp, fmode, flags, sattr, GFP_KERNEL); 1753 if (opendata == NULL) 1754 goto err_put_state_owner; 1755 1756 if (dentry->d_inode != NULL) 1757 opendata->state = nfs4_get_open_state(dentry->d_inode, sp); 1758 1759 status = _nfs4_proc_open(opendata); 1760 if (status != 0) 1761 goto err_opendata_put; 1762 1763 state = nfs4_opendata_to_nfs4_state(opendata); 1764 status = PTR_ERR(state); 1765 if (IS_ERR(state)) 1766 goto err_opendata_put; 1767 if (server->caps & NFS_CAP_POSIX_LOCK) 1768 set_bit(NFS_STATE_POSIX_LOCKS, &state->flags); 1769 1770 if (opendata->o_arg.open_flags & O_EXCL) { 1771 nfs4_exclusive_attrset(opendata, sattr); 1772 1773 nfs_fattr_init(opendata->o_res.f_attr); 1774 status = nfs4_do_setattr(state->inode, cred, 1775 opendata->o_res.f_attr, sattr, 1776 state); 1777 if (status == 0) 1778 nfs_setattr_update_inode(state->inode, sattr); 1779 nfs_post_op_update_inode(state->inode, opendata->o_res.f_attr); 1780 } 1781 nfs4_opendata_put(opendata); 1782 nfs4_put_state_owner(sp); 1783 *res = state; 1784 return 0; 1785 err_opendata_put: 1786 nfs4_opendata_put(opendata); 1787 err_put_state_owner: 1788 nfs4_put_state_owner(sp); 1789 out_err: 1790 *res = NULL; 1791 return status; 1792 } 1793 1794 1795 static struct nfs4_state *nfs4_do_open(struct inode *dir, struct dentry *dentry, fmode_t fmode, int flags, struct iattr *sattr, struct rpc_cred *cred) 1796 { 1797 struct nfs4_exception exception = { }; 1798 struct nfs4_state *res; 1799 int status; 1800 1801 do { 1802 status = _nfs4_do_open(dir, dentry, fmode, flags, sattr, cred, &res); 1803 if (status == 0) 1804 break; 1805 /* NOTE: BAD_SEQID means the server and client disagree about the 1806 * book-keeping w.r.t. state-changing operations 1807 * (OPEN/CLOSE/LOCK/LOCKU...) 1808 * It is actually a sign of a bug on the client or on the server. 1809 * 1810 * If we receive a BAD_SEQID error in the particular case of 1811 * doing an OPEN, we assume that nfs_increment_open_seqid() will 1812 * have unhashed the old state_owner for us, and that we can 1813 * therefore safely retry using a new one. We should still warn 1814 * the user though... 1815 */ 1816 if (status == -NFS4ERR_BAD_SEQID) { 1817 printk(KERN_WARNING "NFS: v4 server %s " 1818 " returned a bad sequence-id error!\n", 1819 NFS_SERVER(dir)->nfs_client->cl_hostname); 1820 exception.retry = 1; 1821 continue; 1822 } 1823 /* 1824 * BAD_STATEID on OPEN means that the server cancelled our 1825 * state before it received the OPEN_CONFIRM. 1826 * Recover by retrying the request as per the discussion 1827 * on Page 181 of RFC3530. 1828 */ 1829 if (status == -NFS4ERR_BAD_STATEID) { 1830 exception.retry = 1; 1831 continue; 1832 } 1833 if (status == -EAGAIN) { 1834 /* We must have found a delegation */ 1835 exception.retry = 1; 1836 continue; 1837 } 1838 res = ERR_PTR(nfs4_handle_exception(NFS_SERVER(dir), 1839 status, &exception)); 1840 } while (exception.retry); 1841 return res; 1842 } 1843 1844 static int _nfs4_do_setattr(struct inode *inode, struct rpc_cred *cred, 1845 struct nfs_fattr *fattr, struct iattr *sattr, 1846 struct nfs4_state *state) 1847 { 1848 struct nfs_server *server = NFS_SERVER(inode); 1849 struct nfs_setattrargs arg = { 1850 .fh = NFS_FH(inode), 1851 .iap = sattr, 1852 .server = server, 1853 .bitmask = server->attr_bitmask, 1854 }; 1855 struct nfs_setattrres res = { 1856 .fattr = fattr, 1857 .server = server, 1858 }; 1859 struct rpc_message msg = { 1860 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SETATTR], 1861 .rpc_argp = &arg, 1862 .rpc_resp = &res, 1863 .rpc_cred = cred, 1864 }; 1865 unsigned long timestamp = jiffies; 1866 int status; 1867 1868 nfs_fattr_init(fattr); 1869 1870 if (nfs4_copy_delegation_stateid(&arg.stateid, inode)) { 1871 /* Use that stateid */ 1872 } else if (state != NULL) { 1873 nfs4_copy_stateid(&arg.stateid, state, current->files, current->tgid); 1874 } else 1875 memcpy(&arg.stateid, &zero_stateid, sizeof(arg.stateid)); 1876 1877 status = nfs4_call_sync(server->client, server, &msg, &arg.seq_args, &res.seq_res, 1); 1878 if (status == 0 && state != NULL) 1879 renew_lease(server, timestamp); 1880 return status; 1881 } 1882 1883 static int nfs4_do_setattr(struct inode *inode, struct rpc_cred *cred, 1884 struct nfs_fattr *fattr, struct iattr *sattr, 1885 struct nfs4_state *state) 1886 { 1887 struct nfs_server *server = NFS_SERVER(inode); 1888 struct nfs4_exception exception = { }; 1889 int err; 1890 do { 1891 err = nfs4_handle_exception(server, 1892 _nfs4_do_setattr(inode, cred, fattr, sattr, state), 1893 &exception); 1894 } while (exception.retry); 1895 return err; 1896 } 1897 1898 struct nfs4_closedata { 1899 struct inode *inode; 1900 struct nfs4_state *state; 1901 struct nfs_closeargs arg; 1902 struct nfs_closeres res; 1903 struct nfs_fattr fattr; 1904 unsigned long timestamp; 1905 bool roc; 1906 u32 roc_barrier; 1907 }; 1908 1909 static void nfs4_free_closedata(void *data) 1910 { 1911 struct nfs4_closedata *calldata = data; 1912 struct nfs4_state_owner *sp = calldata->state->owner; 1913 struct super_block *sb = calldata->state->inode->i_sb; 1914 1915 if (calldata->roc) 1916 pnfs_roc_release(calldata->state->inode); 1917 nfs4_put_open_state(calldata->state); 1918 nfs_free_seqid(calldata->arg.seqid); 1919 nfs4_put_state_owner(sp); 1920 nfs_sb_deactive(sb); 1921 kfree(calldata); 1922 } 1923 1924 static void nfs4_close_clear_stateid_flags(struct nfs4_state *state, 1925 fmode_t fmode) 1926 { 1927 spin_lock(&state->owner->so_lock); 1928 if (!(fmode & FMODE_READ)) 1929 clear_bit(NFS_O_RDONLY_STATE, &state->flags); 1930 if (!(fmode & FMODE_WRITE)) 1931 clear_bit(NFS_O_WRONLY_STATE, &state->flags); 1932 clear_bit(NFS_O_RDWR_STATE, &state->flags); 1933 spin_unlock(&state->owner->so_lock); 1934 } 1935 1936 static void nfs4_close_done(struct rpc_task *task, void *data) 1937 { 1938 struct nfs4_closedata *calldata = data; 1939 struct nfs4_state *state = calldata->state; 1940 struct nfs_server *server = NFS_SERVER(calldata->inode); 1941 1942 if (!nfs4_sequence_done(task, &calldata->res.seq_res)) 1943 return; 1944 /* hmm. we are done with the inode, and in the process of freeing 1945 * the state_owner. we keep this around to process errors 1946 */ 1947 switch (task->tk_status) { 1948 case 0: 1949 if (calldata->roc) 1950 pnfs_roc_set_barrier(state->inode, 1951 calldata->roc_barrier); 1952 nfs_set_open_stateid(state, &calldata->res.stateid, 0); 1953 renew_lease(server, calldata->timestamp); 1954 nfs4_close_clear_stateid_flags(state, 1955 calldata->arg.fmode); 1956 break; 1957 case -NFS4ERR_STALE_STATEID: 1958 case -NFS4ERR_OLD_STATEID: 1959 case -NFS4ERR_BAD_STATEID: 1960 case -NFS4ERR_EXPIRED: 1961 if (calldata->arg.fmode == 0) 1962 break; 1963 default: 1964 if (nfs4_async_handle_error(task, server, state) == -EAGAIN) 1965 rpc_restart_call_prepare(task); 1966 } 1967 nfs_release_seqid(calldata->arg.seqid); 1968 nfs_refresh_inode(calldata->inode, calldata->res.fattr); 1969 } 1970 1971 static void nfs4_close_prepare(struct rpc_task *task, void *data) 1972 { 1973 struct nfs4_closedata *calldata = data; 1974 struct nfs4_state *state = calldata->state; 1975 int call_close = 0; 1976 1977 if (nfs_wait_on_sequence(calldata->arg.seqid, task) != 0) 1978 return; 1979 1980 task->tk_msg.rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_OPEN_DOWNGRADE]; 1981 calldata->arg.fmode = FMODE_READ|FMODE_WRITE; 1982 spin_lock(&state->owner->so_lock); 1983 /* Calculate the change in open mode */ 1984 if (state->n_rdwr == 0) { 1985 if (state->n_rdonly == 0) { 1986 call_close |= test_bit(NFS_O_RDONLY_STATE, &state->flags); 1987 call_close |= test_bit(NFS_O_RDWR_STATE, &state->flags); 1988 calldata->arg.fmode &= ~FMODE_READ; 1989 } 1990 if (state->n_wronly == 0) { 1991 call_close |= test_bit(NFS_O_WRONLY_STATE, &state->flags); 1992 call_close |= test_bit(NFS_O_RDWR_STATE, &state->flags); 1993 calldata->arg.fmode &= ~FMODE_WRITE; 1994 } 1995 } 1996 spin_unlock(&state->owner->so_lock); 1997 1998 if (!call_close) { 1999 /* Note: exit _without_ calling nfs4_close_done */ 2000 task->tk_action = NULL; 2001 return; 2002 } 2003 2004 if (calldata->arg.fmode == 0) { 2005 task->tk_msg.rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_CLOSE]; 2006 if (calldata->roc && 2007 pnfs_roc_drain(calldata->inode, &calldata->roc_barrier)) { 2008 rpc_sleep_on(&NFS_SERVER(calldata->inode)->roc_rpcwaitq, 2009 task, NULL); 2010 return; 2011 } 2012 } 2013 2014 nfs_fattr_init(calldata->res.fattr); 2015 calldata->timestamp = jiffies; 2016 if (nfs4_setup_sequence(NFS_SERVER(calldata->inode), 2017 &calldata->arg.seq_args, &calldata->res.seq_res, 2018 1, task)) 2019 return; 2020 rpc_call_start(task); 2021 } 2022 2023 static const struct rpc_call_ops nfs4_close_ops = { 2024 .rpc_call_prepare = nfs4_close_prepare, 2025 .rpc_call_done = nfs4_close_done, 2026 .rpc_release = nfs4_free_closedata, 2027 }; 2028 2029 /* 2030 * It is possible for data to be read/written from a mem-mapped file 2031 * after the sys_close call (which hits the vfs layer as a flush). 2032 * This means that we can't safely call nfsv4 close on a file until 2033 * the inode is cleared. This in turn means that we are not good 2034 * NFSv4 citizens - we do not indicate to the server to update the file's 2035 * share state even when we are done with one of the three share 2036 * stateid's in the inode. 2037 * 2038 * NOTE: Caller must be holding the sp->so_owner semaphore! 2039 */ 2040 int nfs4_do_close(struct nfs4_state *state, gfp_t gfp_mask, int wait, bool roc) 2041 { 2042 struct nfs_server *server = NFS_SERVER(state->inode); 2043 struct nfs4_closedata *calldata; 2044 struct nfs4_state_owner *sp = state->owner; 2045 struct rpc_task *task; 2046 struct rpc_message msg = { 2047 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_CLOSE], 2048 .rpc_cred = state->owner->so_cred, 2049 }; 2050 struct rpc_task_setup task_setup_data = { 2051 .rpc_client = server->client, 2052 .rpc_message = &msg, 2053 .callback_ops = &nfs4_close_ops, 2054 .workqueue = nfsiod_workqueue, 2055 .flags = RPC_TASK_ASYNC, 2056 }; 2057 int status = -ENOMEM; 2058 2059 calldata = kzalloc(sizeof(*calldata), gfp_mask); 2060 if (calldata == NULL) 2061 goto out; 2062 calldata->inode = state->inode; 2063 calldata->state = state; 2064 calldata->arg.fh = NFS_FH(state->inode); 2065 calldata->arg.stateid = &state->open_stateid; 2066 /* Serialization for the sequence id */ 2067 calldata->arg.seqid = nfs_alloc_seqid(&state->owner->so_seqid, gfp_mask); 2068 if (calldata->arg.seqid == NULL) 2069 goto out_free_calldata; 2070 calldata->arg.fmode = 0; 2071 calldata->arg.bitmask = server->cache_consistency_bitmask; 2072 calldata->res.fattr = &calldata->fattr; 2073 calldata->res.seqid = calldata->arg.seqid; 2074 calldata->res.server = server; 2075 calldata->roc = roc; 2076 nfs_sb_active(calldata->inode->i_sb); 2077 2078 msg.rpc_argp = &calldata->arg; 2079 msg.rpc_resp = &calldata->res; 2080 task_setup_data.callback_data = calldata; 2081 task = rpc_run_task(&task_setup_data); 2082 if (IS_ERR(task)) 2083 return PTR_ERR(task); 2084 status = 0; 2085 if (wait) 2086 status = rpc_wait_for_completion_task(task); 2087 rpc_put_task(task); 2088 return status; 2089 out_free_calldata: 2090 kfree(calldata); 2091 out: 2092 if (roc) 2093 pnfs_roc_release(state->inode); 2094 nfs4_put_open_state(state); 2095 nfs4_put_state_owner(sp); 2096 return status; 2097 } 2098 2099 static struct inode * 2100 nfs4_atomic_open(struct inode *dir, struct nfs_open_context *ctx, int open_flags, struct iattr *attr) 2101 { 2102 struct nfs4_state *state; 2103 2104 /* Protect against concurrent sillydeletes */ 2105 state = nfs4_do_open(dir, ctx->dentry, ctx->mode, open_flags, attr, ctx->cred); 2106 if (IS_ERR(state)) 2107 return ERR_CAST(state); 2108 ctx->state = state; 2109 return igrab(state->inode); 2110 } 2111 2112 static void nfs4_close_context(struct nfs_open_context *ctx, int is_sync) 2113 { 2114 if (ctx->state == NULL) 2115 return; 2116 if (is_sync) 2117 nfs4_close_sync(ctx->state, ctx->mode); 2118 else 2119 nfs4_close_state(ctx->state, ctx->mode); 2120 } 2121 2122 static int _nfs4_server_capabilities(struct nfs_server *server, struct nfs_fh *fhandle) 2123 { 2124 struct nfs4_server_caps_arg args = { 2125 .fhandle = fhandle, 2126 }; 2127 struct nfs4_server_caps_res res = {}; 2128 struct rpc_message msg = { 2129 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SERVER_CAPS], 2130 .rpc_argp = &args, 2131 .rpc_resp = &res, 2132 }; 2133 int status; 2134 2135 status = nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 2136 if (status == 0) { 2137 memcpy(server->attr_bitmask, res.attr_bitmask, sizeof(server->attr_bitmask)); 2138 server->caps &= ~(NFS_CAP_ACLS|NFS_CAP_HARDLINKS| 2139 NFS_CAP_SYMLINKS|NFS_CAP_FILEID| 2140 NFS_CAP_MODE|NFS_CAP_NLINK|NFS_CAP_OWNER| 2141 NFS_CAP_OWNER_GROUP|NFS_CAP_ATIME| 2142 NFS_CAP_CTIME|NFS_CAP_MTIME); 2143 if (res.attr_bitmask[0] & FATTR4_WORD0_ACL) 2144 server->caps |= NFS_CAP_ACLS; 2145 if (res.has_links != 0) 2146 server->caps |= NFS_CAP_HARDLINKS; 2147 if (res.has_symlinks != 0) 2148 server->caps |= NFS_CAP_SYMLINKS; 2149 if (res.attr_bitmask[0] & FATTR4_WORD0_FILEID) 2150 server->caps |= NFS_CAP_FILEID; 2151 if (res.attr_bitmask[1] & FATTR4_WORD1_MODE) 2152 server->caps |= NFS_CAP_MODE; 2153 if (res.attr_bitmask[1] & FATTR4_WORD1_NUMLINKS) 2154 server->caps |= NFS_CAP_NLINK; 2155 if (res.attr_bitmask[1] & FATTR4_WORD1_OWNER) 2156 server->caps |= NFS_CAP_OWNER; 2157 if (res.attr_bitmask[1] & FATTR4_WORD1_OWNER_GROUP) 2158 server->caps |= NFS_CAP_OWNER_GROUP; 2159 if (res.attr_bitmask[1] & FATTR4_WORD1_TIME_ACCESS) 2160 server->caps |= NFS_CAP_ATIME; 2161 if (res.attr_bitmask[1] & FATTR4_WORD1_TIME_METADATA) 2162 server->caps |= NFS_CAP_CTIME; 2163 if (res.attr_bitmask[1] & FATTR4_WORD1_TIME_MODIFY) 2164 server->caps |= NFS_CAP_MTIME; 2165 2166 memcpy(server->cache_consistency_bitmask, res.attr_bitmask, sizeof(server->cache_consistency_bitmask)); 2167 server->cache_consistency_bitmask[0] &= FATTR4_WORD0_CHANGE|FATTR4_WORD0_SIZE; 2168 server->cache_consistency_bitmask[1] &= FATTR4_WORD1_TIME_METADATA|FATTR4_WORD1_TIME_MODIFY; 2169 server->acl_bitmask = res.acl_bitmask; 2170 } 2171 2172 return status; 2173 } 2174 2175 int nfs4_server_capabilities(struct nfs_server *server, struct nfs_fh *fhandle) 2176 { 2177 struct nfs4_exception exception = { }; 2178 int err; 2179 do { 2180 err = nfs4_handle_exception(server, 2181 _nfs4_server_capabilities(server, fhandle), 2182 &exception); 2183 } while (exception.retry); 2184 return err; 2185 } 2186 2187 static int _nfs4_lookup_root(struct nfs_server *server, struct nfs_fh *fhandle, 2188 struct nfs_fsinfo *info) 2189 { 2190 struct nfs4_lookup_root_arg args = { 2191 .bitmask = nfs4_fattr_bitmap, 2192 }; 2193 struct nfs4_lookup_res res = { 2194 .server = server, 2195 .fattr = info->fattr, 2196 .fh = fhandle, 2197 }; 2198 struct rpc_message msg = { 2199 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LOOKUP_ROOT], 2200 .rpc_argp = &args, 2201 .rpc_resp = &res, 2202 }; 2203 2204 nfs_fattr_init(info->fattr); 2205 return nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 2206 } 2207 2208 static int nfs4_lookup_root(struct nfs_server *server, struct nfs_fh *fhandle, 2209 struct nfs_fsinfo *info) 2210 { 2211 struct nfs4_exception exception = { }; 2212 int err; 2213 do { 2214 err = _nfs4_lookup_root(server, fhandle, info); 2215 switch (err) { 2216 case 0: 2217 case -NFS4ERR_WRONGSEC: 2218 break; 2219 default: 2220 err = nfs4_handle_exception(server, err, &exception); 2221 } 2222 } while (exception.retry); 2223 return err; 2224 } 2225 2226 static int nfs4_lookup_root_sec(struct nfs_server *server, struct nfs_fh *fhandle, 2227 struct nfs_fsinfo *info, rpc_authflavor_t flavor) 2228 { 2229 struct rpc_auth *auth; 2230 int ret; 2231 2232 auth = rpcauth_create(flavor, server->client); 2233 if (!auth) { 2234 ret = -EIO; 2235 goto out; 2236 } 2237 ret = nfs4_lookup_root(server, fhandle, info); 2238 out: 2239 return ret; 2240 } 2241 2242 static int nfs4_find_root_sec(struct nfs_server *server, struct nfs_fh *fhandle, 2243 struct nfs_fsinfo *info) 2244 { 2245 int i, len, status = 0; 2246 rpc_authflavor_t flav_array[NFS_MAX_SECFLAVORS]; 2247 2248 len = gss_mech_list_pseudoflavors(&flav_array[0]); 2249 flav_array[len] = RPC_AUTH_NULL; 2250 len += 1; 2251 2252 for (i = 0; i < len; i++) { 2253 status = nfs4_lookup_root_sec(server, fhandle, info, flav_array[i]); 2254 if (status == -NFS4ERR_WRONGSEC || status == -EACCES) 2255 continue; 2256 break; 2257 } 2258 /* 2259 * -EACCESS could mean that the user doesn't have correct permissions 2260 * to access the mount. It could also mean that we tried to mount 2261 * with a gss auth flavor, but rpc.gssd isn't running. Either way, 2262 * existing mount programs don't handle -EACCES very well so it should 2263 * be mapped to -EPERM instead. 2264 */ 2265 if (status == -EACCES) 2266 status = -EPERM; 2267 return status; 2268 } 2269 2270 /* 2271 * get the file handle for the "/" directory on the server 2272 */ 2273 static int nfs4_proc_get_root(struct nfs_server *server, struct nfs_fh *fhandle, 2274 struct nfs_fsinfo *info) 2275 { 2276 int minor_version = server->nfs_client->cl_minorversion; 2277 int status = nfs4_lookup_root(server, fhandle, info); 2278 if ((status == -NFS4ERR_WRONGSEC) && !(server->flags & NFS_MOUNT_SECFLAVOUR)) 2279 /* 2280 * A status of -NFS4ERR_WRONGSEC will be mapped to -EPERM 2281 * by nfs4_map_errors() as this function exits. 2282 */ 2283 status = nfs_v4_minor_ops[minor_version]->find_root_sec(server, fhandle, info); 2284 if (status == 0) 2285 status = nfs4_server_capabilities(server, fhandle); 2286 if (status == 0) 2287 status = nfs4_do_fsinfo(server, fhandle, info); 2288 return nfs4_map_errors(status); 2289 } 2290 2291 static void nfs_fixup_referral_attributes(struct nfs_fattr *fattr); 2292 /* 2293 * Get locations and (maybe) other attributes of a referral. 2294 * Note that we'll actually follow the referral later when 2295 * we detect fsid mismatch in inode revalidation 2296 */ 2297 static int nfs4_get_referral(struct inode *dir, const struct qstr *name, 2298 struct nfs_fattr *fattr, struct nfs_fh *fhandle) 2299 { 2300 int status = -ENOMEM; 2301 struct page *page = NULL; 2302 struct nfs4_fs_locations *locations = NULL; 2303 2304 page = alloc_page(GFP_KERNEL); 2305 if (page == NULL) 2306 goto out; 2307 locations = kmalloc(sizeof(struct nfs4_fs_locations), GFP_KERNEL); 2308 if (locations == NULL) 2309 goto out; 2310 2311 status = nfs4_proc_fs_locations(dir, name, locations, page); 2312 if (status != 0) 2313 goto out; 2314 /* Make sure server returned a different fsid for the referral */ 2315 if (nfs_fsid_equal(&NFS_SERVER(dir)->fsid, &locations->fattr.fsid)) { 2316 dprintk("%s: server did not return a different fsid for" 2317 " a referral at %s\n", __func__, name->name); 2318 status = -EIO; 2319 goto out; 2320 } 2321 /* Fixup attributes for the nfs_lookup() call to nfs_fhget() */ 2322 nfs_fixup_referral_attributes(&locations->fattr); 2323 2324 /* replace the lookup nfs_fattr with the locations nfs_fattr */ 2325 memcpy(fattr, &locations->fattr, sizeof(struct nfs_fattr)); 2326 memset(fhandle, 0, sizeof(struct nfs_fh)); 2327 out: 2328 if (page) 2329 __free_page(page); 2330 kfree(locations); 2331 return status; 2332 } 2333 2334 static int _nfs4_proc_getattr(struct nfs_server *server, struct nfs_fh *fhandle, struct nfs_fattr *fattr) 2335 { 2336 struct nfs4_getattr_arg args = { 2337 .fh = fhandle, 2338 .bitmask = server->attr_bitmask, 2339 }; 2340 struct nfs4_getattr_res res = { 2341 .fattr = fattr, 2342 .server = server, 2343 }; 2344 struct rpc_message msg = { 2345 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_GETATTR], 2346 .rpc_argp = &args, 2347 .rpc_resp = &res, 2348 }; 2349 2350 nfs_fattr_init(fattr); 2351 return nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 2352 } 2353 2354 static int nfs4_proc_getattr(struct nfs_server *server, struct nfs_fh *fhandle, struct nfs_fattr *fattr) 2355 { 2356 struct nfs4_exception exception = { }; 2357 int err; 2358 do { 2359 err = nfs4_handle_exception(server, 2360 _nfs4_proc_getattr(server, fhandle, fattr), 2361 &exception); 2362 } while (exception.retry); 2363 return err; 2364 } 2365 2366 /* 2367 * The file is not closed if it is opened due to the a request to change 2368 * the size of the file. The open call will not be needed once the 2369 * VFS layer lookup-intents are implemented. 2370 * 2371 * Close is called when the inode is destroyed. 2372 * If we haven't opened the file for O_WRONLY, we 2373 * need to in the size_change case to obtain a stateid. 2374 * 2375 * Got race? 2376 * Because OPEN is always done by name in nfsv4, it is 2377 * possible that we opened a different file by the same 2378 * name. We can recognize this race condition, but we 2379 * can't do anything about it besides returning an error. 2380 * 2381 * This will be fixed with VFS changes (lookup-intent). 2382 */ 2383 static int 2384 nfs4_proc_setattr(struct dentry *dentry, struct nfs_fattr *fattr, 2385 struct iattr *sattr) 2386 { 2387 struct inode *inode = dentry->d_inode; 2388 struct rpc_cred *cred = NULL; 2389 struct nfs4_state *state = NULL; 2390 int status; 2391 2392 if (pnfs_ld_layoutret_on_setattr(inode)) 2393 pnfs_return_layout(inode); 2394 2395 nfs_fattr_init(fattr); 2396 2397 /* Search for an existing open(O_WRITE) file */ 2398 if (sattr->ia_valid & ATTR_FILE) { 2399 struct nfs_open_context *ctx; 2400 2401 ctx = nfs_file_open_context(sattr->ia_file); 2402 if (ctx) { 2403 cred = ctx->cred; 2404 state = ctx->state; 2405 } 2406 } 2407 2408 status = nfs4_do_setattr(inode, cred, fattr, sattr, state); 2409 if (status == 0) 2410 nfs_setattr_update_inode(inode, sattr); 2411 return status; 2412 } 2413 2414 static int _nfs4_proc_lookup(struct rpc_clnt *clnt, struct inode *dir, 2415 const struct qstr *name, struct nfs_fh *fhandle, 2416 struct nfs_fattr *fattr) 2417 { 2418 struct nfs_server *server = NFS_SERVER(dir); 2419 int status; 2420 struct nfs4_lookup_arg args = { 2421 .bitmask = server->attr_bitmask, 2422 .dir_fh = NFS_FH(dir), 2423 .name = name, 2424 }; 2425 struct nfs4_lookup_res res = { 2426 .server = server, 2427 .fattr = fattr, 2428 .fh = fhandle, 2429 }; 2430 struct rpc_message msg = { 2431 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LOOKUP], 2432 .rpc_argp = &args, 2433 .rpc_resp = &res, 2434 }; 2435 2436 nfs_fattr_init(fattr); 2437 2438 dprintk("NFS call lookup %s\n", name->name); 2439 status = nfs4_call_sync(clnt, server, &msg, &args.seq_args, &res.seq_res, 0); 2440 dprintk("NFS reply lookup: %d\n", status); 2441 return status; 2442 } 2443 2444 void nfs_fixup_secinfo_attributes(struct nfs_fattr *fattr, struct nfs_fh *fh) 2445 { 2446 memset(fh, 0, sizeof(struct nfs_fh)); 2447 fattr->fsid.major = 1; 2448 fattr->valid |= NFS_ATTR_FATTR_TYPE | NFS_ATTR_FATTR_MODE | 2449 NFS_ATTR_FATTR_NLINK | NFS_ATTR_FATTR_FSID | NFS_ATTR_FATTR_MOUNTPOINT; 2450 fattr->mode = S_IFDIR | S_IRUGO | S_IXUGO; 2451 fattr->nlink = 2; 2452 } 2453 2454 static int nfs4_proc_lookup(struct rpc_clnt *clnt, struct inode *dir, struct qstr *name, 2455 struct nfs_fh *fhandle, struct nfs_fattr *fattr) 2456 { 2457 struct nfs4_exception exception = { }; 2458 int err; 2459 do { 2460 int status; 2461 2462 status = _nfs4_proc_lookup(clnt, dir, name, fhandle, fattr); 2463 switch (status) { 2464 case -NFS4ERR_BADNAME: 2465 return -ENOENT; 2466 case -NFS4ERR_MOVED: 2467 err = nfs4_get_referral(dir, name, fattr, fhandle); 2468 break; 2469 case -NFS4ERR_WRONGSEC: 2470 nfs_fixup_secinfo_attributes(fattr, fhandle); 2471 } 2472 err = nfs4_handle_exception(NFS_SERVER(dir), 2473 status, &exception); 2474 } while (exception.retry); 2475 return err; 2476 } 2477 2478 static int _nfs4_proc_access(struct inode *inode, struct nfs_access_entry *entry) 2479 { 2480 struct nfs_server *server = NFS_SERVER(inode); 2481 struct nfs4_accessargs args = { 2482 .fh = NFS_FH(inode), 2483 .bitmask = server->attr_bitmask, 2484 }; 2485 struct nfs4_accessres res = { 2486 .server = server, 2487 }; 2488 struct rpc_message msg = { 2489 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_ACCESS], 2490 .rpc_argp = &args, 2491 .rpc_resp = &res, 2492 .rpc_cred = entry->cred, 2493 }; 2494 int mode = entry->mask; 2495 int status; 2496 2497 /* 2498 * Determine which access bits we want to ask for... 2499 */ 2500 if (mode & MAY_READ) 2501 args.access |= NFS4_ACCESS_READ; 2502 if (S_ISDIR(inode->i_mode)) { 2503 if (mode & MAY_WRITE) 2504 args.access |= NFS4_ACCESS_MODIFY | NFS4_ACCESS_EXTEND | NFS4_ACCESS_DELETE; 2505 if (mode & MAY_EXEC) 2506 args.access |= NFS4_ACCESS_LOOKUP; 2507 } else { 2508 if (mode & MAY_WRITE) 2509 args.access |= NFS4_ACCESS_MODIFY | NFS4_ACCESS_EXTEND; 2510 if (mode & MAY_EXEC) 2511 args.access |= NFS4_ACCESS_EXECUTE; 2512 } 2513 2514 res.fattr = nfs_alloc_fattr(); 2515 if (res.fattr == NULL) 2516 return -ENOMEM; 2517 2518 status = nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 2519 if (!status) { 2520 entry->mask = 0; 2521 if (res.access & NFS4_ACCESS_READ) 2522 entry->mask |= MAY_READ; 2523 if (res.access & (NFS4_ACCESS_MODIFY | NFS4_ACCESS_EXTEND | NFS4_ACCESS_DELETE)) 2524 entry->mask |= MAY_WRITE; 2525 if (res.access & (NFS4_ACCESS_LOOKUP|NFS4_ACCESS_EXECUTE)) 2526 entry->mask |= MAY_EXEC; 2527 nfs_refresh_inode(inode, res.fattr); 2528 } 2529 nfs_free_fattr(res.fattr); 2530 return status; 2531 } 2532 2533 static int nfs4_proc_access(struct inode *inode, struct nfs_access_entry *entry) 2534 { 2535 struct nfs4_exception exception = { }; 2536 int err; 2537 do { 2538 err = nfs4_handle_exception(NFS_SERVER(inode), 2539 _nfs4_proc_access(inode, entry), 2540 &exception); 2541 } while (exception.retry); 2542 return err; 2543 } 2544 2545 /* 2546 * TODO: For the time being, we don't try to get any attributes 2547 * along with any of the zero-copy operations READ, READDIR, 2548 * READLINK, WRITE. 2549 * 2550 * In the case of the first three, we want to put the GETATTR 2551 * after the read-type operation -- this is because it is hard 2552 * to predict the length of a GETATTR response in v4, and thus 2553 * align the READ data correctly. This means that the GETATTR 2554 * may end up partially falling into the page cache, and we should 2555 * shift it into the 'tail' of the xdr_buf before processing. 2556 * To do this efficiently, we need to know the total length 2557 * of data received, which doesn't seem to be available outside 2558 * of the RPC layer. 2559 * 2560 * In the case of WRITE, we also want to put the GETATTR after 2561 * the operation -- in this case because we want to make sure 2562 * we get the post-operation mtime and size. This means that 2563 * we can't use xdr_encode_pages() as written: we need a variant 2564 * of it which would leave room in the 'tail' iovec. 2565 * 2566 * Both of these changes to the XDR layer would in fact be quite 2567 * minor, but I decided to leave them for a subsequent patch. 2568 */ 2569 static int _nfs4_proc_readlink(struct inode *inode, struct page *page, 2570 unsigned int pgbase, unsigned int pglen) 2571 { 2572 struct nfs4_readlink args = { 2573 .fh = NFS_FH(inode), 2574 .pgbase = pgbase, 2575 .pglen = pglen, 2576 .pages = &page, 2577 }; 2578 struct nfs4_readlink_res res; 2579 struct rpc_message msg = { 2580 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_READLINK], 2581 .rpc_argp = &args, 2582 .rpc_resp = &res, 2583 }; 2584 2585 return nfs4_call_sync(NFS_SERVER(inode)->client, NFS_SERVER(inode), &msg, &args.seq_args, &res.seq_res, 0); 2586 } 2587 2588 static int nfs4_proc_readlink(struct inode *inode, struct page *page, 2589 unsigned int pgbase, unsigned int pglen) 2590 { 2591 struct nfs4_exception exception = { }; 2592 int err; 2593 do { 2594 err = nfs4_handle_exception(NFS_SERVER(inode), 2595 _nfs4_proc_readlink(inode, page, pgbase, pglen), 2596 &exception); 2597 } while (exception.retry); 2598 return err; 2599 } 2600 2601 /* 2602 * Got race? 2603 * We will need to arrange for the VFS layer to provide an atomic open. 2604 * Until then, this create/open method is prone to inefficiency and race 2605 * conditions due to the lookup, create, and open VFS calls from sys_open() 2606 * placed on the wire. 2607 * 2608 * Given the above sorry state of affairs, I'm simply sending an OPEN. 2609 * The file will be opened again in the subsequent VFS open call 2610 * (nfs4_proc_file_open). 2611 * 2612 * The open for read will just hang around to be used by any process that 2613 * opens the file O_RDONLY. This will all be resolved with the VFS changes. 2614 */ 2615 2616 static int 2617 nfs4_proc_create(struct inode *dir, struct dentry *dentry, struct iattr *sattr, 2618 int flags, struct nfs_open_context *ctx) 2619 { 2620 struct dentry *de = dentry; 2621 struct nfs4_state *state; 2622 struct rpc_cred *cred = NULL; 2623 fmode_t fmode = 0; 2624 int status = 0; 2625 2626 if (ctx != NULL) { 2627 cred = ctx->cred; 2628 de = ctx->dentry; 2629 fmode = ctx->mode; 2630 } 2631 sattr->ia_mode &= ~current_umask(); 2632 state = nfs4_do_open(dir, de, fmode, flags, sattr, cred); 2633 d_drop(dentry); 2634 if (IS_ERR(state)) { 2635 status = PTR_ERR(state); 2636 goto out; 2637 } 2638 d_add(dentry, igrab(state->inode)); 2639 nfs_set_verifier(dentry, nfs_save_change_attribute(dir)); 2640 if (ctx != NULL) 2641 ctx->state = state; 2642 else 2643 nfs4_close_sync(state, fmode); 2644 out: 2645 return status; 2646 } 2647 2648 static int _nfs4_proc_remove(struct inode *dir, struct qstr *name) 2649 { 2650 struct nfs_server *server = NFS_SERVER(dir); 2651 struct nfs_removeargs args = { 2652 .fh = NFS_FH(dir), 2653 .name.len = name->len, 2654 .name.name = name->name, 2655 .bitmask = server->attr_bitmask, 2656 }; 2657 struct nfs_removeres res = { 2658 .server = server, 2659 }; 2660 struct rpc_message msg = { 2661 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_REMOVE], 2662 .rpc_argp = &args, 2663 .rpc_resp = &res, 2664 }; 2665 int status = -ENOMEM; 2666 2667 res.dir_attr = nfs_alloc_fattr(); 2668 if (res.dir_attr == NULL) 2669 goto out; 2670 2671 status = nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 1); 2672 if (status == 0) { 2673 update_changeattr(dir, &res.cinfo); 2674 nfs_post_op_update_inode(dir, res.dir_attr); 2675 } 2676 nfs_free_fattr(res.dir_attr); 2677 out: 2678 return status; 2679 } 2680 2681 static int nfs4_proc_remove(struct inode *dir, struct qstr *name) 2682 { 2683 struct nfs4_exception exception = { }; 2684 int err; 2685 do { 2686 err = nfs4_handle_exception(NFS_SERVER(dir), 2687 _nfs4_proc_remove(dir, name), 2688 &exception); 2689 } while (exception.retry); 2690 return err; 2691 } 2692 2693 static void nfs4_proc_unlink_setup(struct rpc_message *msg, struct inode *dir) 2694 { 2695 struct nfs_server *server = NFS_SERVER(dir); 2696 struct nfs_removeargs *args = msg->rpc_argp; 2697 struct nfs_removeres *res = msg->rpc_resp; 2698 2699 args->bitmask = server->cache_consistency_bitmask; 2700 res->server = server; 2701 res->seq_res.sr_slot = NULL; 2702 msg->rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_REMOVE]; 2703 } 2704 2705 static int nfs4_proc_unlink_done(struct rpc_task *task, struct inode *dir) 2706 { 2707 struct nfs_removeres *res = task->tk_msg.rpc_resp; 2708 2709 if (!nfs4_sequence_done(task, &res->seq_res)) 2710 return 0; 2711 if (nfs4_async_handle_error(task, res->server, NULL) == -EAGAIN) 2712 return 0; 2713 update_changeattr(dir, &res->cinfo); 2714 nfs_post_op_update_inode(dir, res->dir_attr); 2715 return 1; 2716 } 2717 2718 static void nfs4_proc_rename_setup(struct rpc_message *msg, struct inode *dir) 2719 { 2720 struct nfs_server *server = NFS_SERVER(dir); 2721 struct nfs_renameargs *arg = msg->rpc_argp; 2722 struct nfs_renameres *res = msg->rpc_resp; 2723 2724 msg->rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_RENAME]; 2725 arg->bitmask = server->attr_bitmask; 2726 res->server = server; 2727 } 2728 2729 static int nfs4_proc_rename_done(struct rpc_task *task, struct inode *old_dir, 2730 struct inode *new_dir) 2731 { 2732 struct nfs_renameres *res = task->tk_msg.rpc_resp; 2733 2734 if (!nfs4_sequence_done(task, &res->seq_res)) 2735 return 0; 2736 if (nfs4_async_handle_error(task, res->server, NULL) == -EAGAIN) 2737 return 0; 2738 2739 update_changeattr(old_dir, &res->old_cinfo); 2740 nfs_post_op_update_inode(old_dir, res->old_fattr); 2741 update_changeattr(new_dir, &res->new_cinfo); 2742 nfs_post_op_update_inode(new_dir, res->new_fattr); 2743 return 1; 2744 } 2745 2746 static int _nfs4_proc_rename(struct inode *old_dir, struct qstr *old_name, 2747 struct inode *new_dir, struct qstr *new_name) 2748 { 2749 struct nfs_server *server = NFS_SERVER(old_dir); 2750 struct nfs_renameargs arg = { 2751 .old_dir = NFS_FH(old_dir), 2752 .new_dir = NFS_FH(new_dir), 2753 .old_name = old_name, 2754 .new_name = new_name, 2755 .bitmask = server->attr_bitmask, 2756 }; 2757 struct nfs_renameres res = { 2758 .server = server, 2759 }; 2760 struct rpc_message msg = { 2761 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_RENAME], 2762 .rpc_argp = &arg, 2763 .rpc_resp = &res, 2764 }; 2765 int status = -ENOMEM; 2766 2767 res.old_fattr = nfs_alloc_fattr(); 2768 res.new_fattr = nfs_alloc_fattr(); 2769 if (res.old_fattr == NULL || res.new_fattr == NULL) 2770 goto out; 2771 2772 status = nfs4_call_sync(server->client, server, &msg, &arg.seq_args, &res.seq_res, 1); 2773 if (!status) { 2774 update_changeattr(old_dir, &res.old_cinfo); 2775 nfs_post_op_update_inode(old_dir, res.old_fattr); 2776 update_changeattr(new_dir, &res.new_cinfo); 2777 nfs_post_op_update_inode(new_dir, res.new_fattr); 2778 } 2779 out: 2780 nfs_free_fattr(res.new_fattr); 2781 nfs_free_fattr(res.old_fattr); 2782 return status; 2783 } 2784 2785 static int nfs4_proc_rename(struct inode *old_dir, struct qstr *old_name, 2786 struct inode *new_dir, struct qstr *new_name) 2787 { 2788 struct nfs4_exception exception = { }; 2789 int err; 2790 do { 2791 err = nfs4_handle_exception(NFS_SERVER(old_dir), 2792 _nfs4_proc_rename(old_dir, old_name, 2793 new_dir, new_name), 2794 &exception); 2795 } while (exception.retry); 2796 return err; 2797 } 2798 2799 static int _nfs4_proc_link(struct inode *inode, struct inode *dir, struct qstr *name) 2800 { 2801 struct nfs_server *server = NFS_SERVER(inode); 2802 struct nfs4_link_arg arg = { 2803 .fh = NFS_FH(inode), 2804 .dir_fh = NFS_FH(dir), 2805 .name = name, 2806 .bitmask = server->attr_bitmask, 2807 }; 2808 struct nfs4_link_res res = { 2809 .server = server, 2810 }; 2811 struct rpc_message msg = { 2812 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LINK], 2813 .rpc_argp = &arg, 2814 .rpc_resp = &res, 2815 }; 2816 int status = -ENOMEM; 2817 2818 res.fattr = nfs_alloc_fattr(); 2819 res.dir_attr = nfs_alloc_fattr(); 2820 if (res.fattr == NULL || res.dir_attr == NULL) 2821 goto out; 2822 2823 status = nfs4_call_sync(server->client, server, &msg, &arg.seq_args, &res.seq_res, 1); 2824 if (!status) { 2825 update_changeattr(dir, &res.cinfo); 2826 nfs_post_op_update_inode(dir, res.dir_attr); 2827 nfs_post_op_update_inode(inode, res.fattr); 2828 } 2829 out: 2830 nfs_free_fattr(res.dir_attr); 2831 nfs_free_fattr(res.fattr); 2832 return status; 2833 } 2834 2835 static int nfs4_proc_link(struct inode *inode, struct inode *dir, struct qstr *name) 2836 { 2837 struct nfs4_exception exception = { }; 2838 int err; 2839 do { 2840 err = nfs4_handle_exception(NFS_SERVER(inode), 2841 _nfs4_proc_link(inode, dir, name), 2842 &exception); 2843 } while (exception.retry); 2844 return err; 2845 } 2846 2847 struct nfs4_createdata { 2848 struct rpc_message msg; 2849 struct nfs4_create_arg arg; 2850 struct nfs4_create_res res; 2851 struct nfs_fh fh; 2852 struct nfs_fattr fattr; 2853 struct nfs_fattr dir_fattr; 2854 }; 2855 2856 static struct nfs4_createdata *nfs4_alloc_createdata(struct inode *dir, 2857 struct qstr *name, struct iattr *sattr, u32 ftype) 2858 { 2859 struct nfs4_createdata *data; 2860 2861 data = kzalloc(sizeof(*data), GFP_KERNEL); 2862 if (data != NULL) { 2863 struct nfs_server *server = NFS_SERVER(dir); 2864 2865 data->msg.rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_CREATE]; 2866 data->msg.rpc_argp = &data->arg; 2867 data->msg.rpc_resp = &data->res; 2868 data->arg.dir_fh = NFS_FH(dir); 2869 data->arg.server = server; 2870 data->arg.name = name; 2871 data->arg.attrs = sattr; 2872 data->arg.ftype = ftype; 2873 data->arg.bitmask = server->attr_bitmask; 2874 data->res.server = server; 2875 data->res.fh = &data->fh; 2876 data->res.fattr = &data->fattr; 2877 data->res.dir_fattr = &data->dir_fattr; 2878 nfs_fattr_init(data->res.fattr); 2879 nfs_fattr_init(data->res.dir_fattr); 2880 } 2881 return data; 2882 } 2883 2884 static int nfs4_do_create(struct inode *dir, struct dentry *dentry, struct nfs4_createdata *data) 2885 { 2886 int status = nfs4_call_sync(NFS_SERVER(dir)->client, NFS_SERVER(dir), &data->msg, 2887 &data->arg.seq_args, &data->res.seq_res, 1); 2888 if (status == 0) { 2889 update_changeattr(dir, &data->res.dir_cinfo); 2890 nfs_post_op_update_inode(dir, data->res.dir_fattr); 2891 status = nfs_instantiate(dentry, data->res.fh, data->res.fattr); 2892 } 2893 return status; 2894 } 2895 2896 static void nfs4_free_createdata(struct nfs4_createdata *data) 2897 { 2898 kfree(data); 2899 } 2900 2901 static int _nfs4_proc_symlink(struct inode *dir, struct dentry *dentry, 2902 struct page *page, unsigned int len, struct iattr *sattr) 2903 { 2904 struct nfs4_createdata *data; 2905 int status = -ENAMETOOLONG; 2906 2907 if (len > NFS4_MAXPATHLEN) 2908 goto out; 2909 2910 status = -ENOMEM; 2911 data = nfs4_alloc_createdata(dir, &dentry->d_name, sattr, NF4LNK); 2912 if (data == NULL) 2913 goto out; 2914 2915 data->msg.rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SYMLINK]; 2916 data->arg.u.symlink.pages = &page; 2917 data->arg.u.symlink.len = len; 2918 2919 status = nfs4_do_create(dir, dentry, data); 2920 2921 nfs4_free_createdata(data); 2922 out: 2923 return status; 2924 } 2925 2926 static int nfs4_proc_symlink(struct inode *dir, struct dentry *dentry, 2927 struct page *page, unsigned int len, struct iattr *sattr) 2928 { 2929 struct nfs4_exception exception = { }; 2930 int err; 2931 do { 2932 err = nfs4_handle_exception(NFS_SERVER(dir), 2933 _nfs4_proc_symlink(dir, dentry, page, 2934 len, sattr), 2935 &exception); 2936 } while (exception.retry); 2937 return err; 2938 } 2939 2940 static int _nfs4_proc_mkdir(struct inode *dir, struct dentry *dentry, 2941 struct iattr *sattr) 2942 { 2943 struct nfs4_createdata *data; 2944 int status = -ENOMEM; 2945 2946 data = nfs4_alloc_createdata(dir, &dentry->d_name, sattr, NF4DIR); 2947 if (data == NULL) 2948 goto out; 2949 2950 status = nfs4_do_create(dir, dentry, data); 2951 2952 nfs4_free_createdata(data); 2953 out: 2954 return status; 2955 } 2956 2957 static int nfs4_proc_mkdir(struct inode *dir, struct dentry *dentry, 2958 struct iattr *sattr) 2959 { 2960 struct nfs4_exception exception = { }; 2961 int err; 2962 2963 sattr->ia_mode &= ~current_umask(); 2964 do { 2965 err = nfs4_handle_exception(NFS_SERVER(dir), 2966 _nfs4_proc_mkdir(dir, dentry, sattr), 2967 &exception); 2968 } while (exception.retry); 2969 return err; 2970 } 2971 2972 static int _nfs4_proc_readdir(struct dentry *dentry, struct rpc_cred *cred, 2973 u64 cookie, struct page **pages, unsigned int count, int plus) 2974 { 2975 struct inode *dir = dentry->d_inode; 2976 struct nfs4_readdir_arg args = { 2977 .fh = NFS_FH(dir), 2978 .pages = pages, 2979 .pgbase = 0, 2980 .count = count, 2981 .bitmask = NFS_SERVER(dentry->d_inode)->attr_bitmask, 2982 .plus = plus, 2983 }; 2984 struct nfs4_readdir_res res; 2985 struct rpc_message msg = { 2986 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_READDIR], 2987 .rpc_argp = &args, 2988 .rpc_resp = &res, 2989 .rpc_cred = cred, 2990 }; 2991 int status; 2992 2993 dprintk("%s: dentry = %s/%s, cookie = %Lu\n", __func__, 2994 dentry->d_parent->d_name.name, 2995 dentry->d_name.name, 2996 (unsigned long long)cookie); 2997 nfs4_setup_readdir(cookie, NFS_COOKIEVERF(dir), dentry, &args); 2998 res.pgbase = args.pgbase; 2999 status = nfs4_call_sync(NFS_SERVER(dir)->client, NFS_SERVER(dir), &msg, &args.seq_args, &res.seq_res, 0); 3000 if (status >= 0) { 3001 memcpy(NFS_COOKIEVERF(dir), res.verifier.data, NFS4_VERIFIER_SIZE); 3002 status += args.pgbase; 3003 } 3004 3005 nfs_invalidate_atime(dir); 3006 3007 dprintk("%s: returns %d\n", __func__, status); 3008 return status; 3009 } 3010 3011 static int nfs4_proc_readdir(struct dentry *dentry, struct rpc_cred *cred, 3012 u64 cookie, struct page **pages, unsigned int count, int plus) 3013 { 3014 struct nfs4_exception exception = { }; 3015 int err; 3016 do { 3017 err = nfs4_handle_exception(NFS_SERVER(dentry->d_inode), 3018 _nfs4_proc_readdir(dentry, cred, cookie, 3019 pages, count, plus), 3020 &exception); 3021 } while (exception.retry); 3022 return err; 3023 } 3024 3025 static int _nfs4_proc_mknod(struct inode *dir, struct dentry *dentry, 3026 struct iattr *sattr, dev_t rdev) 3027 { 3028 struct nfs4_createdata *data; 3029 int mode = sattr->ia_mode; 3030 int status = -ENOMEM; 3031 3032 BUG_ON(!(sattr->ia_valid & ATTR_MODE)); 3033 BUG_ON(!S_ISFIFO(mode) && !S_ISBLK(mode) && !S_ISCHR(mode) && !S_ISSOCK(mode)); 3034 3035 data = nfs4_alloc_createdata(dir, &dentry->d_name, sattr, NF4SOCK); 3036 if (data == NULL) 3037 goto out; 3038 3039 if (S_ISFIFO(mode)) 3040 data->arg.ftype = NF4FIFO; 3041 else if (S_ISBLK(mode)) { 3042 data->arg.ftype = NF4BLK; 3043 data->arg.u.device.specdata1 = MAJOR(rdev); 3044 data->arg.u.device.specdata2 = MINOR(rdev); 3045 } 3046 else if (S_ISCHR(mode)) { 3047 data->arg.ftype = NF4CHR; 3048 data->arg.u.device.specdata1 = MAJOR(rdev); 3049 data->arg.u.device.specdata2 = MINOR(rdev); 3050 } 3051 3052 status = nfs4_do_create(dir, dentry, data); 3053 3054 nfs4_free_createdata(data); 3055 out: 3056 return status; 3057 } 3058 3059 static int nfs4_proc_mknod(struct inode *dir, struct dentry *dentry, 3060 struct iattr *sattr, dev_t rdev) 3061 { 3062 struct nfs4_exception exception = { }; 3063 int err; 3064 3065 sattr->ia_mode &= ~current_umask(); 3066 do { 3067 err = nfs4_handle_exception(NFS_SERVER(dir), 3068 _nfs4_proc_mknod(dir, dentry, sattr, rdev), 3069 &exception); 3070 } while (exception.retry); 3071 return err; 3072 } 3073 3074 static int _nfs4_proc_statfs(struct nfs_server *server, struct nfs_fh *fhandle, 3075 struct nfs_fsstat *fsstat) 3076 { 3077 struct nfs4_statfs_arg args = { 3078 .fh = fhandle, 3079 .bitmask = server->attr_bitmask, 3080 }; 3081 struct nfs4_statfs_res res = { 3082 .fsstat = fsstat, 3083 }; 3084 struct rpc_message msg = { 3085 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_STATFS], 3086 .rpc_argp = &args, 3087 .rpc_resp = &res, 3088 }; 3089 3090 nfs_fattr_init(fsstat->fattr); 3091 return nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 3092 } 3093 3094 static int nfs4_proc_statfs(struct nfs_server *server, struct nfs_fh *fhandle, struct nfs_fsstat *fsstat) 3095 { 3096 struct nfs4_exception exception = { }; 3097 int err; 3098 do { 3099 err = nfs4_handle_exception(server, 3100 _nfs4_proc_statfs(server, fhandle, fsstat), 3101 &exception); 3102 } while (exception.retry); 3103 return err; 3104 } 3105 3106 static int _nfs4_do_fsinfo(struct nfs_server *server, struct nfs_fh *fhandle, 3107 struct nfs_fsinfo *fsinfo) 3108 { 3109 struct nfs4_fsinfo_arg args = { 3110 .fh = fhandle, 3111 .bitmask = server->attr_bitmask, 3112 }; 3113 struct nfs4_fsinfo_res res = { 3114 .fsinfo = fsinfo, 3115 }; 3116 struct rpc_message msg = { 3117 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_FSINFO], 3118 .rpc_argp = &args, 3119 .rpc_resp = &res, 3120 }; 3121 3122 return nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 3123 } 3124 3125 static int nfs4_do_fsinfo(struct nfs_server *server, struct nfs_fh *fhandle, struct nfs_fsinfo *fsinfo) 3126 { 3127 struct nfs4_exception exception = { }; 3128 int err; 3129 3130 do { 3131 err = nfs4_handle_exception(server, 3132 _nfs4_do_fsinfo(server, fhandle, fsinfo), 3133 &exception); 3134 } while (exception.retry); 3135 return err; 3136 } 3137 3138 static int nfs4_proc_fsinfo(struct nfs_server *server, struct nfs_fh *fhandle, struct nfs_fsinfo *fsinfo) 3139 { 3140 nfs_fattr_init(fsinfo->fattr); 3141 return nfs4_do_fsinfo(server, fhandle, fsinfo); 3142 } 3143 3144 static int _nfs4_proc_pathconf(struct nfs_server *server, struct nfs_fh *fhandle, 3145 struct nfs_pathconf *pathconf) 3146 { 3147 struct nfs4_pathconf_arg args = { 3148 .fh = fhandle, 3149 .bitmask = server->attr_bitmask, 3150 }; 3151 struct nfs4_pathconf_res res = { 3152 .pathconf = pathconf, 3153 }; 3154 struct rpc_message msg = { 3155 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_PATHCONF], 3156 .rpc_argp = &args, 3157 .rpc_resp = &res, 3158 }; 3159 3160 /* None of the pathconf attributes are mandatory to implement */ 3161 if ((args.bitmask[0] & nfs4_pathconf_bitmap[0]) == 0) { 3162 memset(pathconf, 0, sizeof(*pathconf)); 3163 return 0; 3164 } 3165 3166 nfs_fattr_init(pathconf->fattr); 3167 return nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 3168 } 3169 3170 static int nfs4_proc_pathconf(struct nfs_server *server, struct nfs_fh *fhandle, 3171 struct nfs_pathconf *pathconf) 3172 { 3173 struct nfs4_exception exception = { }; 3174 int err; 3175 3176 do { 3177 err = nfs4_handle_exception(server, 3178 _nfs4_proc_pathconf(server, fhandle, pathconf), 3179 &exception); 3180 } while (exception.retry); 3181 return err; 3182 } 3183 3184 void __nfs4_read_done_cb(struct nfs_read_data *data) 3185 { 3186 nfs_invalidate_atime(data->inode); 3187 } 3188 3189 static int nfs4_read_done_cb(struct rpc_task *task, struct nfs_read_data *data) 3190 { 3191 struct nfs_server *server = NFS_SERVER(data->inode); 3192 3193 if (nfs4_async_handle_error(task, server, data->args.context->state) == -EAGAIN) { 3194 rpc_restart_call_prepare(task); 3195 return -EAGAIN; 3196 } 3197 3198 __nfs4_read_done_cb(data); 3199 if (task->tk_status > 0) 3200 renew_lease(server, data->timestamp); 3201 return 0; 3202 } 3203 3204 static int nfs4_read_done(struct rpc_task *task, struct nfs_read_data *data) 3205 { 3206 3207 dprintk("--> %s\n", __func__); 3208 3209 if (!nfs4_sequence_done(task, &data->res.seq_res)) 3210 return -EAGAIN; 3211 3212 return data->read_done_cb ? data->read_done_cb(task, data) : 3213 nfs4_read_done_cb(task, data); 3214 } 3215 3216 static void nfs4_proc_read_setup(struct nfs_read_data *data, struct rpc_message *msg) 3217 { 3218 data->timestamp = jiffies; 3219 data->read_done_cb = nfs4_read_done_cb; 3220 msg->rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_READ]; 3221 } 3222 3223 /* Reset the the nfs_read_data to send the read to the MDS. */ 3224 void nfs4_reset_read(struct rpc_task *task, struct nfs_read_data *data) 3225 { 3226 dprintk("%s Reset task for i/o through\n", __func__); 3227 put_lseg(data->lseg); 3228 data->lseg = NULL; 3229 /* offsets will differ in the dense stripe case */ 3230 data->args.offset = data->mds_offset; 3231 data->ds_clp = NULL; 3232 data->args.fh = NFS_FH(data->inode); 3233 data->read_done_cb = nfs4_read_done_cb; 3234 task->tk_ops = data->mds_ops; 3235 rpc_task_reset_client(task, NFS_CLIENT(data->inode)); 3236 } 3237 EXPORT_SYMBOL_GPL(nfs4_reset_read); 3238 3239 static int nfs4_write_done_cb(struct rpc_task *task, struct nfs_write_data *data) 3240 { 3241 struct inode *inode = data->inode; 3242 3243 if (nfs4_async_handle_error(task, NFS_SERVER(inode), data->args.context->state) == -EAGAIN) { 3244 rpc_restart_call_prepare(task); 3245 return -EAGAIN; 3246 } 3247 if (task->tk_status >= 0) { 3248 renew_lease(NFS_SERVER(inode), data->timestamp); 3249 nfs_post_op_update_inode_force_wcc(inode, data->res.fattr); 3250 } 3251 return 0; 3252 } 3253 3254 static int nfs4_write_done(struct rpc_task *task, struct nfs_write_data *data) 3255 { 3256 if (!nfs4_sequence_done(task, &data->res.seq_res)) 3257 return -EAGAIN; 3258 return data->write_done_cb ? data->write_done_cb(task, data) : 3259 nfs4_write_done_cb(task, data); 3260 } 3261 3262 /* Reset the the nfs_write_data to send the write to the MDS. */ 3263 void nfs4_reset_write(struct rpc_task *task, struct nfs_write_data *data) 3264 { 3265 dprintk("%s Reset task for i/o through\n", __func__); 3266 put_lseg(data->lseg); 3267 data->lseg = NULL; 3268 data->ds_clp = NULL; 3269 data->write_done_cb = nfs4_write_done_cb; 3270 data->args.fh = NFS_FH(data->inode); 3271 data->args.bitmask = data->res.server->cache_consistency_bitmask; 3272 data->args.offset = data->mds_offset; 3273 data->res.fattr = &data->fattr; 3274 task->tk_ops = data->mds_ops; 3275 rpc_task_reset_client(task, NFS_CLIENT(data->inode)); 3276 } 3277 EXPORT_SYMBOL_GPL(nfs4_reset_write); 3278 3279 static void nfs4_proc_write_setup(struct nfs_write_data *data, struct rpc_message *msg) 3280 { 3281 struct nfs_server *server = NFS_SERVER(data->inode); 3282 3283 if (data->lseg) { 3284 data->args.bitmask = NULL; 3285 data->res.fattr = NULL; 3286 } else 3287 data->args.bitmask = server->cache_consistency_bitmask; 3288 if (!data->write_done_cb) 3289 data->write_done_cb = nfs4_write_done_cb; 3290 data->res.server = server; 3291 data->timestamp = jiffies; 3292 3293 msg->rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_WRITE]; 3294 } 3295 3296 static int nfs4_commit_done_cb(struct rpc_task *task, struct nfs_write_data *data) 3297 { 3298 struct inode *inode = data->inode; 3299 3300 if (nfs4_async_handle_error(task, NFS_SERVER(inode), NULL) == -EAGAIN) { 3301 rpc_restart_call_prepare(task); 3302 return -EAGAIN; 3303 } 3304 nfs_refresh_inode(inode, data->res.fattr); 3305 return 0; 3306 } 3307 3308 static int nfs4_commit_done(struct rpc_task *task, struct nfs_write_data *data) 3309 { 3310 if (!nfs4_sequence_done(task, &data->res.seq_res)) 3311 return -EAGAIN; 3312 return data->write_done_cb(task, data); 3313 } 3314 3315 static void nfs4_proc_commit_setup(struct nfs_write_data *data, struct rpc_message *msg) 3316 { 3317 struct nfs_server *server = NFS_SERVER(data->inode); 3318 3319 if (data->lseg) { 3320 data->args.bitmask = NULL; 3321 data->res.fattr = NULL; 3322 } else 3323 data->args.bitmask = server->cache_consistency_bitmask; 3324 if (!data->write_done_cb) 3325 data->write_done_cb = nfs4_commit_done_cb; 3326 data->res.server = server; 3327 msg->rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_COMMIT]; 3328 } 3329 3330 struct nfs4_renewdata { 3331 struct nfs_client *client; 3332 unsigned long timestamp; 3333 }; 3334 3335 /* 3336 * nfs4_proc_async_renew(): This is not one of the nfs_rpc_ops; it is a special 3337 * standalone procedure for queueing an asynchronous RENEW. 3338 */ 3339 static void nfs4_renew_release(void *calldata) 3340 { 3341 struct nfs4_renewdata *data = calldata; 3342 struct nfs_client *clp = data->client; 3343 3344 if (atomic_read(&clp->cl_count) > 1) 3345 nfs4_schedule_state_renewal(clp); 3346 nfs_put_client(clp); 3347 kfree(data); 3348 } 3349 3350 static void nfs4_renew_done(struct rpc_task *task, void *calldata) 3351 { 3352 struct nfs4_renewdata *data = calldata; 3353 struct nfs_client *clp = data->client; 3354 unsigned long timestamp = data->timestamp; 3355 3356 if (task->tk_status < 0) { 3357 /* Unless we're shutting down, schedule state recovery! */ 3358 if (test_bit(NFS_CS_RENEWD, &clp->cl_res_state) == 0) 3359 return; 3360 if (task->tk_status != NFS4ERR_CB_PATH_DOWN) { 3361 nfs4_schedule_lease_recovery(clp); 3362 return; 3363 } 3364 nfs4_schedule_path_down_recovery(clp); 3365 } 3366 do_renew_lease(clp, timestamp); 3367 } 3368 3369 static const struct rpc_call_ops nfs4_renew_ops = { 3370 .rpc_call_done = nfs4_renew_done, 3371 .rpc_release = nfs4_renew_release, 3372 }; 3373 3374 static int nfs4_proc_async_renew(struct nfs_client *clp, struct rpc_cred *cred, unsigned renew_flags) 3375 { 3376 struct rpc_message msg = { 3377 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_RENEW], 3378 .rpc_argp = clp, 3379 .rpc_cred = cred, 3380 }; 3381 struct nfs4_renewdata *data; 3382 3383 if (renew_flags == 0) 3384 return 0; 3385 if (!atomic_inc_not_zero(&clp->cl_count)) 3386 return -EIO; 3387 data = kmalloc(sizeof(*data), GFP_NOFS); 3388 if (data == NULL) 3389 return -ENOMEM; 3390 data->client = clp; 3391 data->timestamp = jiffies; 3392 return rpc_call_async(clp->cl_rpcclient, &msg, RPC_TASK_SOFT, 3393 &nfs4_renew_ops, data); 3394 } 3395 3396 static int nfs4_proc_renew(struct nfs_client *clp, struct rpc_cred *cred) 3397 { 3398 struct rpc_message msg = { 3399 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_RENEW], 3400 .rpc_argp = clp, 3401 .rpc_cred = cred, 3402 }; 3403 unsigned long now = jiffies; 3404 int status; 3405 3406 status = rpc_call_sync(clp->cl_rpcclient, &msg, 0); 3407 if (status < 0) 3408 return status; 3409 do_renew_lease(clp, now); 3410 return 0; 3411 } 3412 3413 static inline int nfs4_server_supports_acls(struct nfs_server *server) 3414 { 3415 return (server->caps & NFS_CAP_ACLS) 3416 && (server->acl_bitmask & ACL4_SUPPORT_ALLOW_ACL) 3417 && (server->acl_bitmask & ACL4_SUPPORT_DENY_ACL); 3418 } 3419 3420 /* Assuming that XATTR_SIZE_MAX is a multiple of PAGE_CACHE_SIZE, and that 3421 * it's OK to put sizeof(void) * (XATTR_SIZE_MAX/PAGE_CACHE_SIZE) bytes on 3422 * the stack. 3423 */ 3424 #define NFS4ACL_MAXPAGES (XATTR_SIZE_MAX >> PAGE_CACHE_SHIFT) 3425 3426 static void buf_to_pages(const void *buf, size_t buflen, 3427 struct page **pages, unsigned int *pgbase) 3428 { 3429 const void *p = buf; 3430 3431 *pgbase = offset_in_page(buf); 3432 p -= *pgbase; 3433 while (p < buf + buflen) { 3434 *(pages++) = virt_to_page(p); 3435 p += PAGE_CACHE_SIZE; 3436 } 3437 } 3438 3439 static int buf_to_pages_noslab(const void *buf, size_t buflen, 3440 struct page **pages, unsigned int *pgbase) 3441 { 3442 struct page *newpage, **spages; 3443 int rc = 0; 3444 size_t len; 3445 spages = pages; 3446 3447 do { 3448 len = min_t(size_t, PAGE_CACHE_SIZE, buflen); 3449 newpage = alloc_page(GFP_KERNEL); 3450 3451 if (newpage == NULL) 3452 goto unwind; 3453 memcpy(page_address(newpage), buf, len); 3454 buf += len; 3455 buflen -= len; 3456 *pages++ = newpage; 3457 rc++; 3458 } while (buflen != 0); 3459 3460 return rc; 3461 3462 unwind: 3463 for(; rc > 0; rc--) 3464 __free_page(spages[rc-1]); 3465 return -ENOMEM; 3466 } 3467 3468 struct nfs4_cached_acl { 3469 int cached; 3470 size_t len; 3471 char data[0]; 3472 }; 3473 3474 static void nfs4_set_cached_acl(struct inode *inode, struct nfs4_cached_acl *acl) 3475 { 3476 struct nfs_inode *nfsi = NFS_I(inode); 3477 3478 spin_lock(&inode->i_lock); 3479 kfree(nfsi->nfs4_acl); 3480 nfsi->nfs4_acl = acl; 3481 spin_unlock(&inode->i_lock); 3482 } 3483 3484 static void nfs4_zap_acl_attr(struct inode *inode) 3485 { 3486 nfs4_set_cached_acl(inode, NULL); 3487 } 3488 3489 static inline ssize_t nfs4_read_cached_acl(struct inode *inode, char *buf, size_t buflen) 3490 { 3491 struct nfs_inode *nfsi = NFS_I(inode); 3492 struct nfs4_cached_acl *acl; 3493 int ret = -ENOENT; 3494 3495 spin_lock(&inode->i_lock); 3496 acl = nfsi->nfs4_acl; 3497 if (acl == NULL) 3498 goto out; 3499 if (buf == NULL) /* user is just asking for length */ 3500 goto out_len; 3501 if (acl->cached == 0) 3502 goto out; 3503 ret = -ERANGE; /* see getxattr(2) man page */ 3504 if (acl->len > buflen) 3505 goto out; 3506 memcpy(buf, acl->data, acl->len); 3507 out_len: 3508 ret = acl->len; 3509 out: 3510 spin_unlock(&inode->i_lock); 3511 return ret; 3512 } 3513 3514 static void nfs4_write_cached_acl(struct inode *inode, const char *buf, size_t acl_len) 3515 { 3516 struct nfs4_cached_acl *acl; 3517 3518 if (buf && acl_len <= PAGE_SIZE) { 3519 acl = kmalloc(sizeof(*acl) + acl_len, GFP_KERNEL); 3520 if (acl == NULL) 3521 goto out; 3522 acl->cached = 1; 3523 memcpy(acl->data, buf, acl_len); 3524 } else { 3525 acl = kmalloc(sizeof(*acl), GFP_KERNEL); 3526 if (acl == NULL) 3527 goto out; 3528 acl->cached = 0; 3529 } 3530 acl->len = acl_len; 3531 out: 3532 nfs4_set_cached_acl(inode, acl); 3533 } 3534 3535 static ssize_t __nfs4_get_acl_uncached(struct inode *inode, void *buf, size_t buflen) 3536 { 3537 struct page *pages[NFS4ACL_MAXPAGES]; 3538 struct nfs_getaclargs args = { 3539 .fh = NFS_FH(inode), 3540 .acl_pages = pages, 3541 .acl_len = buflen, 3542 }; 3543 struct nfs_getaclres res = { 3544 .acl_len = buflen, 3545 }; 3546 void *resp_buf; 3547 struct rpc_message msg = { 3548 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_GETACL], 3549 .rpc_argp = &args, 3550 .rpc_resp = &res, 3551 }; 3552 struct page *localpage = NULL; 3553 int ret; 3554 3555 if (buflen < PAGE_SIZE) { 3556 /* As long as we're doing a round trip to the server anyway, 3557 * let's be prepared for a page of acl data. */ 3558 localpage = alloc_page(GFP_KERNEL); 3559 resp_buf = page_address(localpage); 3560 if (localpage == NULL) 3561 return -ENOMEM; 3562 args.acl_pages[0] = localpage; 3563 args.acl_pgbase = 0; 3564 args.acl_len = PAGE_SIZE; 3565 } else { 3566 resp_buf = buf; 3567 buf_to_pages(buf, buflen, args.acl_pages, &args.acl_pgbase); 3568 } 3569 ret = nfs4_call_sync(NFS_SERVER(inode)->client, NFS_SERVER(inode), &msg, &args.seq_args, &res.seq_res, 0); 3570 if (ret) 3571 goto out_free; 3572 if (res.acl_len > args.acl_len) 3573 nfs4_write_cached_acl(inode, NULL, res.acl_len); 3574 else 3575 nfs4_write_cached_acl(inode, resp_buf, res.acl_len); 3576 if (buf) { 3577 ret = -ERANGE; 3578 if (res.acl_len > buflen) 3579 goto out_free; 3580 if (localpage) 3581 memcpy(buf, resp_buf, res.acl_len); 3582 } 3583 ret = res.acl_len; 3584 out_free: 3585 if (localpage) 3586 __free_page(localpage); 3587 return ret; 3588 } 3589 3590 static ssize_t nfs4_get_acl_uncached(struct inode *inode, void *buf, size_t buflen) 3591 { 3592 struct nfs4_exception exception = { }; 3593 ssize_t ret; 3594 do { 3595 ret = __nfs4_get_acl_uncached(inode, buf, buflen); 3596 if (ret >= 0) 3597 break; 3598 ret = nfs4_handle_exception(NFS_SERVER(inode), ret, &exception); 3599 } while (exception.retry); 3600 return ret; 3601 } 3602 3603 static ssize_t nfs4_proc_get_acl(struct inode *inode, void *buf, size_t buflen) 3604 { 3605 struct nfs_server *server = NFS_SERVER(inode); 3606 int ret; 3607 3608 if (!nfs4_server_supports_acls(server)) 3609 return -EOPNOTSUPP; 3610 ret = nfs_revalidate_inode(server, inode); 3611 if (ret < 0) 3612 return ret; 3613 if (NFS_I(inode)->cache_validity & NFS_INO_INVALID_ACL) 3614 nfs_zap_acl_cache(inode); 3615 ret = nfs4_read_cached_acl(inode, buf, buflen); 3616 if (ret != -ENOENT) 3617 return ret; 3618 return nfs4_get_acl_uncached(inode, buf, buflen); 3619 } 3620 3621 static int __nfs4_proc_set_acl(struct inode *inode, const void *buf, size_t buflen) 3622 { 3623 struct nfs_server *server = NFS_SERVER(inode); 3624 struct page *pages[NFS4ACL_MAXPAGES]; 3625 struct nfs_setaclargs arg = { 3626 .fh = NFS_FH(inode), 3627 .acl_pages = pages, 3628 .acl_len = buflen, 3629 }; 3630 struct nfs_setaclres res; 3631 struct rpc_message msg = { 3632 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SETACL], 3633 .rpc_argp = &arg, 3634 .rpc_resp = &res, 3635 }; 3636 int ret, i; 3637 3638 if (!nfs4_server_supports_acls(server)) 3639 return -EOPNOTSUPP; 3640 i = buf_to_pages_noslab(buf, buflen, arg.acl_pages, &arg.acl_pgbase); 3641 if (i < 0) 3642 return i; 3643 nfs_inode_return_delegation(inode); 3644 ret = nfs4_call_sync(server->client, server, &msg, &arg.seq_args, &res.seq_res, 1); 3645 3646 /* 3647 * Free each page after tx, so the only ref left is 3648 * held by the network stack 3649 */ 3650 for (; i > 0; i--) 3651 put_page(pages[i-1]); 3652 3653 /* 3654 * Acl update can result in inode attribute update. 3655 * so mark the attribute cache invalid. 3656 */ 3657 spin_lock(&inode->i_lock); 3658 NFS_I(inode)->cache_validity |= NFS_INO_INVALID_ATTR; 3659 spin_unlock(&inode->i_lock); 3660 nfs_access_zap_cache(inode); 3661 nfs_zap_acl_cache(inode); 3662 return ret; 3663 } 3664 3665 static int nfs4_proc_set_acl(struct inode *inode, const void *buf, size_t buflen) 3666 { 3667 struct nfs4_exception exception = { }; 3668 int err; 3669 do { 3670 err = nfs4_handle_exception(NFS_SERVER(inode), 3671 __nfs4_proc_set_acl(inode, buf, buflen), 3672 &exception); 3673 } while (exception.retry); 3674 return err; 3675 } 3676 3677 static int 3678 nfs4_async_handle_error(struct rpc_task *task, const struct nfs_server *server, struct nfs4_state *state) 3679 { 3680 struct nfs_client *clp = server->nfs_client; 3681 3682 if (task->tk_status >= 0) 3683 return 0; 3684 switch(task->tk_status) { 3685 case -NFS4ERR_ADMIN_REVOKED: 3686 case -NFS4ERR_BAD_STATEID: 3687 case -NFS4ERR_OPENMODE: 3688 if (state == NULL) 3689 break; 3690 nfs4_schedule_stateid_recovery(server, state); 3691 goto wait_on_recovery; 3692 case -NFS4ERR_EXPIRED: 3693 if (state != NULL) 3694 nfs4_schedule_stateid_recovery(server, state); 3695 case -NFS4ERR_STALE_STATEID: 3696 case -NFS4ERR_STALE_CLIENTID: 3697 nfs4_schedule_lease_recovery(clp); 3698 goto wait_on_recovery; 3699 #if defined(CONFIG_NFS_V4_1) 3700 case -NFS4ERR_BADSESSION: 3701 case -NFS4ERR_BADSLOT: 3702 case -NFS4ERR_BAD_HIGH_SLOT: 3703 case -NFS4ERR_DEADSESSION: 3704 case -NFS4ERR_CONN_NOT_BOUND_TO_SESSION: 3705 case -NFS4ERR_SEQ_FALSE_RETRY: 3706 case -NFS4ERR_SEQ_MISORDERED: 3707 dprintk("%s ERROR %d, Reset session\n", __func__, 3708 task->tk_status); 3709 nfs4_schedule_session_recovery(clp->cl_session); 3710 task->tk_status = 0; 3711 return -EAGAIN; 3712 #endif /* CONFIG_NFS_V4_1 */ 3713 case -NFS4ERR_DELAY: 3714 nfs_inc_server_stats(server, NFSIOS_DELAY); 3715 case -NFS4ERR_GRACE: 3716 case -EKEYEXPIRED: 3717 rpc_delay(task, NFS4_POLL_RETRY_MAX); 3718 task->tk_status = 0; 3719 return -EAGAIN; 3720 case -NFS4ERR_RETRY_UNCACHED_REP: 3721 case -NFS4ERR_OLD_STATEID: 3722 task->tk_status = 0; 3723 return -EAGAIN; 3724 } 3725 task->tk_status = nfs4_map_errors(task->tk_status); 3726 return 0; 3727 wait_on_recovery: 3728 rpc_sleep_on(&clp->cl_rpcwaitq, task, NULL); 3729 if (test_bit(NFS4CLNT_MANAGER_RUNNING, &clp->cl_state) == 0) 3730 rpc_wake_up_queued_task(&clp->cl_rpcwaitq, task); 3731 task->tk_status = 0; 3732 return -EAGAIN; 3733 } 3734 3735 int nfs4_proc_setclientid(struct nfs_client *clp, u32 program, 3736 unsigned short port, struct rpc_cred *cred, 3737 struct nfs4_setclientid_res *res) 3738 { 3739 nfs4_verifier sc_verifier; 3740 struct nfs4_setclientid setclientid = { 3741 .sc_verifier = &sc_verifier, 3742 .sc_prog = program, 3743 .sc_cb_ident = clp->cl_cb_ident, 3744 }; 3745 struct rpc_message msg = { 3746 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SETCLIENTID], 3747 .rpc_argp = &setclientid, 3748 .rpc_resp = res, 3749 .rpc_cred = cred, 3750 }; 3751 __be32 *p; 3752 int loop = 0; 3753 int status; 3754 3755 p = (__be32*)sc_verifier.data; 3756 *p++ = htonl((u32)clp->cl_boot_time.tv_sec); 3757 *p = htonl((u32)clp->cl_boot_time.tv_nsec); 3758 3759 for(;;) { 3760 setclientid.sc_name_len = scnprintf(setclientid.sc_name, 3761 sizeof(setclientid.sc_name), "%s/%s %s %s %u", 3762 clp->cl_ipaddr, 3763 rpc_peeraddr2str(clp->cl_rpcclient, 3764 RPC_DISPLAY_ADDR), 3765 rpc_peeraddr2str(clp->cl_rpcclient, 3766 RPC_DISPLAY_PROTO), 3767 clp->cl_rpcclient->cl_auth->au_ops->au_name, 3768 clp->cl_id_uniquifier); 3769 setclientid.sc_netid_len = scnprintf(setclientid.sc_netid, 3770 sizeof(setclientid.sc_netid), 3771 rpc_peeraddr2str(clp->cl_rpcclient, 3772 RPC_DISPLAY_NETID)); 3773 setclientid.sc_uaddr_len = scnprintf(setclientid.sc_uaddr, 3774 sizeof(setclientid.sc_uaddr), "%s.%u.%u", 3775 clp->cl_ipaddr, port >> 8, port & 255); 3776 3777 status = rpc_call_sync(clp->cl_rpcclient, &msg, RPC_TASK_TIMEOUT); 3778 if (status != -NFS4ERR_CLID_INUSE) 3779 break; 3780 if (loop != 0) { 3781 ++clp->cl_id_uniquifier; 3782 break; 3783 } 3784 ++loop; 3785 ssleep(clp->cl_lease_time / HZ + 1); 3786 } 3787 return status; 3788 } 3789 3790 int nfs4_proc_setclientid_confirm(struct nfs_client *clp, 3791 struct nfs4_setclientid_res *arg, 3792 struct rpc_cred *cred) 3793 { 3794 struct nfs_fsinfo fsinfo; 3795 struct rpc_message msg = { 3796 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SETCLIENTID_CONFIRM], 3797 .rpc_argp = arg, 3798 .rpc_resp = &fsinfo, 3799 .rpc_cred = cred, 3800 }; 3801 unsigned long now; 3802 int status; 3803 3804 now = jiffies; 3805 status = rpc_call_sync(clp->cl_rpcclient, &msg, RPC_TASK_TIMEOUT); 3806 if (status == 0) { 3807 spin_lock(&clp->cl_lock); 3808 clp->cl_lease_time = fsinfo.lease_time * HZ; 3809 clp->cl_last_renewal = now; 3810 spin_unlock(&clp->cl_lock); 3811 } 3812 return status; 3813 } 3814 3815 struct nfs4_delegreturndata { 3816 struct nfs4_delegreturnargs args; 3817 struct nfs4_delegreturnres res; 3818 struct nfs_fh fh; 3819 nfs4_stateid stateid; 3820 unsigned long timestamp; 3821 struct nfs_fattr fattr; 3822 int rpc_status; 3823 }; 3824 3825 static void nfs4_delegreturn_done(struct rpc_task *task, void *calldata) 3826 { 3827 struct nfs4_delegreturndata *data = calldata; 3828 3829 if (!nfs4_sequence_done(task, &data->res.seq_res)) 3830 return; 3831 3832 switch (task->tk_status) { 3833 case -NFS4ERR_STALE_STATEID: 3834 case -NFS4ERR_EXPIRED: 3835 case 0: 3836 renew_lease(data->res.server, data->timestamp); 3837 break; 3838 default: 3839 if (nfs4_async_handle_error(task, data->res.server, NULL) == 3840 -EAGAIN) { 3841 rpc_restart_call_prepare(task); 3842 return; 3843 } 3844 } 3845 data->rpc_status = task->tk_status; 3846 } 3847 3848 static void nfs4_delegreturn_release(void *calldata) 3849 { 3850 kfree(calldata); 3851 } 3852 3853 #if defined(CONFIG_NFS_V4_1) 3854 static void nfs4_delegreturn_prepare(struct rpc_task *task, void *data) 3855 { 3856 struct nfs4_delegreturndata *d_data; 3857 3858 d_data = (struct nfs4_delegreturndata *)data; 3859 3860 if (nfs4_setup_sequence(d_data->res.server, 3861 &d_data->args.seq_args, 3862 &d_data->res.seq_res, 1, task)) 3863 return; 3864 rpc_call_start(task); 3865 } 3866 #endif /* CONFIG_NFS_V4_1 */ 3867 3868 static const struct rpc_call_ops nfs4_delegreturn_ops = { 3869 #if defined(CONFIG_NFS_V4_1) 3870 .rpc_call_prepare = nfs4_delegreturn_prepare, 3871 #endif /* CONFIG_NFS_V4_1 */ 3872 .rpc_call_done = nfs4_delegreturn_done, 3873 .rpc_release = nfs4_delegreturn_release, 3874 }; 3875 3876 static int _nfs4_proc_delegreturn(struct inode *inode, struct rpc_cred *cred, const nfs4_stateid *stateid, int issync) 3877 { 3878 struct nfs4_delegreturndata *data; 3879 struct nfs_server *server = NFS_SERVER(inode); 3880 struct rpc_task *task; 3881 struct rpc_message msg = { 3882 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_DELEGRETURN], 3883 .rpc_cred = cred, 3884 }; 3885 struct rpc_task_setup task_setup_data = { 3886 .rpc_client = server->client, 3887 .rpc_message = &msg, 3888 .callback_ops = &nfs4_delegreturn_ops, 3889 .flags = RPC_TASK_ASYNC, 3890 }; 3891 int status = 0; 3892 3893 data = kzalloc(sizeof(*data), GFP_NOFS); 3894 if (data == NULL) 3895 return -ENOMEM; 3896 data->args.fhandle = &data->fh; 3897 data->args.stateid = &data->stateid; 3898 data->args.bitmask = server->attr_bitmask; 3899 nfs_copy_fh(&data->fh, NFS_FH(inode)); 3900 memcpy(&data->stateid, stateid, sizeof(data->stateid)); 3901 data->res.fattr = &data->fattr; 3902 data->res.server = server; 3903 nfs_fattr_init(data->res.fattr); 3904 data->timestamp = jiffies; 3905 data->rpc_status = 0; 3906 3907 task_setup_data.callback_data = data; 3908 msg.rpc_argp = &data->args; 3909 msg.rpc_resp = &data->res; 3910 task = rpc_run_task(&task_setup_data); 3911 if (IS_ERR(task)) 3912 return PTR_ERR(task); 3913 if (!issync) 3914 goto out; 3915 status = nfs4_wait_for_completion_rpc_task(task); 3916 if (status != 0) 3917 goto out; 3918 status = data->rpc_status; 3919 if (status != 0) 3920 goto out; 3921 nfs_refresh_inode(inode, &data->fattr); 3922 out: 3923 rpc_put_task(task); 3924 return status; 3925 } 3926 3927 int nfs4_proc_delegreturn(struct inode *inode, struct rpc_cred *cred, const nfs4_stateid *stateid, int issync) 3928 { 3929 struct nfs_server *server = NFS_SERVER(inode); 3930 struct nfs4_exception exception = { }; 3931 int err; 3932 do { 3933 err = _nfs4_proc_delegreturn(inode, cred, stateid, issync); 3934 switch (err) { 3935 case -NFS4ERR_STALE_STATEID: 3936 case -NFS4ERR_EXPIRED: 3937 case 0: 3938 return 0; 3939 } 3940 err = nfs4_handle_exception(server, err, &exception); 3941 } while (exception.retry); 3942 return err; 3943 } 3944 3945 #define NFS4_LOCK_MINTIMEOUT (1 * HZ) 3946 #define NFS4_LOCK_MAXTIMEOUT (30 * HZ) 3947 3948 /* 3949 * sleep, with exponential backoff, and retry the LOCK operation. 3950 */ 3951 static unsigned long 3952 nfs4_set_lock_task_retry(unsigned long timeout) 3953 { 3954 schedule_timeout_killable(timeout); 3955 timeout <<= 1; 3956 if (timeout > NFS4_LOCK_MAXTIMEOUT) 3957 return NFS4_LOCK_MAXTIMEOUT; 3958 return timeout; 3959 } 3960 3961 static int _nfs4_proc_getlk(struct nfs4_state *state, int cmd, struct file_lock *request) 3962 { 3963 struct inode *inode = state->inode; 3964 struct nfs_server *server = NFS_SERVER(inode); 3965 struct nfs_client *clp = server->nfs_client; 3966 struct nfs_lockt_args arg = { 3967 .fh = NFS_FH(inode), 3968 .fl = request, 3969 }; 3970 struct nfs_lockt_res res = { 3971 .denied = request, 3972 }; 3973 struct rpc_message msg = { 3974 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LOCKT], 3975 .rpc_argp = &arg, 3976 .rpc_resp = &res, 3977 .rpc_cred = state->owner->so_cred, 3978 }; 3979 struct nfs4_lock_state *lsp; 3980 int status; 3981 3982 arg.lock_owner.clientid = clp->cl_clientid; 3983 status = nfs4_set_lock_state(state, request); 3984 if (status != 0) 3985 goto out; 3986 lsp = request->fl_u.nfs4_fl.owner; 3987 arg.lock_owner.id = lsp->ls_id.id; 3988 arg.lock_owner.s_dev = server->s_dev; 3989 status = nfs4_call_sync(server->client, server, &msg, &arg.seq_args, &res.seq_res, 1); 3990 switch (status) { 3991 case 0: 3992 request->fl_type = F_UNLCK; 3993 break; 3994 case -NFS4ERR_DENIED: 3995 status = 0; 3996 } 3997 request->fl_ops->fl_release_private(request); 3998 out: 3999 return status; 4000 } 4001 4002 static int nfs4_proc_getlk(struct nfs4_state *state, int cmd, struct file_lock *request) 4003 { 4004 struct nfs4_exception exception = { }; 4005 int err; 4006 4007 do { 4008 err = nfs4_handle_exception(NFS_SERVER(state->inode), 4009 _nfs4_proc_getlk(state, cmd, request), 4010 &exception); 4011 } while (exception.retry); 4012 return err; 4013 } 4014 4015 static int do_vfs_lock(struct file *file, struct file_lock *fl) 4016 { 4017 int res = 0; 4018 switch (fl->fl_flags & (FL_POSIX|FL_FLOCK)) { 4019 case FL_POSIX: 4020 res = posix_lock_file_wait(file, fl); 4021 break; 4022 case FL_FLOCK: 4023 res = flock_lock_file_wait(file, fl); 4024 break; 4025 default: 4026 BUG(); 4027 } 4028 return res; 4029 } 4030 4031 struct nfs4_unlockdata { 4032 struct nfs_locku_args arg; 4033 struct nfs_locku_res res; 4034 struct nfs4_lock_state *lsp; 4035 struct nfs_open_context *ctx; 4036 struct file_lock fl; 4037 const struct nfs_server *server; 4038 unsigned long timestamp; 4039 }; 4040 4041 static struct nfs4_unlockdata *nfs4_alloc_unlockdata(struct file_lock *fl, 4042 struct nfs_open_context *ctx, 4043 struct nfs4_lock_state *lsp, 4044 struct nfs_seqid *seqid) 4045 { 4046 struct nfs4_unlockdata *p; 4047 struct inode *inode = lsp->ls_state->inode; 4048 4049 p = kzalloc(sizeof(*p), GFP_NOFS); 4050 if (p == NULL) 4051 return NULL; 4052 p->arg.fh = NFS_FH(inode); 4053 p->arg.fl = &p->fl; 4054 p->arg.seqid = seqid; 4055 p->res.seqid = seqid; 4056 p->arg.stateid = &lsp->ls_stateid; 4057 p->lsp = lsp; 4058 atomic_inc(&lsp->ls_count); 4059 /* Ensure we don't close file until we're done freeing locks! */ 4060 p->ctx = get_nfs_open_context(ctx); 4061 memcpy(&p->fl, fl, sizeof(p->fl)); 4062 p->server = NFS_SERVER(inode); 4063 return p; 4064 } 4065 4066 static void nfs4_locku_release_calldata(void *data) 4067 { 4068 struct nfs4_unlockdata *calldata = data; 4069 nfs_free_seqid(calldata->arg.seqid); 4070 nfs4_put_lock_state(calldata->lsp); 4071 put_nfs_open_context(calldata->ctx); 4072 kfree(calldata); 4073 } 4074 4075 static void nfs4_locku_done(struct rpc_task *task, void *data) 4076 { 4077 struct nfs4_unlockdata *calldata = data; 4078 4079 if (!nfs4_sequence_done(task, &calldata->res.seq_res)) 4080 return; 4081 switch (task->tk_status) { 4082 case 0: 4083 memcpy(calldata->lsp->ls_stateid.data, 4084 calldata->res.stateid.data, 4085 sizeof(calldata->lsp->ls_stateid.data)); 4086 renew_lease(calldata->server, calldata->timestamp); 4087 break; 4088 case -NFS4ERR_BAD_STATEID: 4089 case -NFS4ERR_OLD_STATEID: 4090 case -NFS4ERR_STALE_STATEID: 4091 case -NFS4ERR_EXPIRED: 4092 break; 4093 default: 4094 if (nfs4_async_handle_error(task, calldata->server, NULL) == -EAGAIN) 4095 rpc_restart_call_prepare(task); 4096 } 4097 } 4098 4099 static void nfs4_locku_prepare(struct rpc_task *task, void *data) 4100 { 4101 struct nfs4_unlockdata *calldata = data; 4102 4103 if (nfs_wait_on_sequence(calldata->arg.seqid, task) != 0) 4104 return; 4105 if ((calldata->lsp->ls_flags & NFS_LOCK_INITIALIZED) == 0) { 4106 /* Note: exit _without_ running nfs4_locku_done */ 4107 task->tk_action = NULL; 4108 return; 4109 } 4110 calldata->timestamp = jiffies; 4111 if (nfs4_setup_sequence(calldata->server, 4112 &calldata->arg.seq_args, 4113 &calldata->res.seq_res, 1, task)) 4114 return; 4115 rpc_call_start(task); 4116 } 4117 4118 static const struct rpc_call_ops nfs4_locku_ops = { 4119 .rpc_call_prepare = nfs4_locku_prepare, 4120 .rpc_call_done = nfs4_locku_done, 4121 .rpc_release = nfs4_locku_release_calldata, 4122 }; 4123 4124 static struct rpc_task *nfs4_do_unlck(struct file_lock *fl, 4125 struct nfs_open_context *ctx, 4126 struct nfs4_lock_state *lsp, 4127 struct nfs_seqid *seqid) 4128 { 4129 struct nfs4_unlockdata *data; 4130 struct rpc_message msg = { 4131 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LOCKU], 4132 .rpc_cred = ctx->cred, 4133 }; 4134 struct rpc_task_setup task_setup_data = { 4135 .rpc_client = NFS_CLIENT(lsp->ls_state->inode), 4136 .rpc_message = &msg, 4137 .callback_ops = &nfs4_locku_ops, 4138 .workqueue = nfsiod_workqueue, 4139 .flags = RPC_TASK_ASYNC, 4140 }; 4141 4142 /* Ensure this is an unlock - when canceling a lock, the 4143 * canceled lock is passed in, and it won't be an unlock. 4144 */ 4145 fl->fl_type = F_UNLCK; 4146 4147 data = nfs4_alloc_unlockdata(fl, ctx, lsp, seqid); 4148 if (data == NULL) { 4149 nfs_free_seqid(seqid); 4150 return ERR_PTR(-ENOMEM); 4151 } 4152 4153 msg.rpc_argp = &data->arg; 4154 msg.rpc_resp = &data->res; 4155 task_setup_data.callback_data = data; 4156 return rpc_run_task(&task_setup_data); 4157 } 4158 4159 static int nfs4_proc_unlck(struct nfs4_state *state, int cmd, struct file_lock *request) 4160 { 4161 struct nfs_inode *nfsi = NFS_I(state->inode); 4162 struct nfs_seqid *seqid; 4163 struct nfs4_lock_state *lsp; 4164 struct rpc_task *task; 4165 int status = 0; 4166 unsigned char fl_flags = request->fl_flags; 4167 4168 status = nfs4_set_lock_state(state, request); 4169 /* Unlock _before_ we do the RPC call */ 4170 request->fl_flags |= FL_EXISTS; 4171 down_read(&nfsi->rwsem); 4172 if (do_vfs_lock(request->fl_file, request) == -ENOENT) { 4173 up_read(&nfsi->rwsem); 4174 goto out; 4175 } 4176 up_read(&nfsi->rwsem); 4177 if (status != 0) 4178 goto out; 4179 /* Is this a delegated lock? */ 4180 if (test_bit(NFS_DELEGATED_STATE, &state->flags)) 4181 goto out; 4182 lsp = request->fl_u.nfs4_fl.owner; 4183 seqid = nfs_alloc_seqid(&lsp->ls_seqid, GFP_KERNEL); 4184 status = -ENOMEM; 4185 if (seqid == NULL) 4186 goto out; 4187 task = nfs4_do_unlck(request, nfs_file_open_context(request->fl_file), lsp, seqid); 4188 status = PTR_ERR(task); 4189 if (IS_ERR(task)) 4190 goto out; 4191 status = nfs4_wait_for_completion_rpc_task(task); 4192 rpc_put_task(task); 4193 out: 4194 request->fl_flags = fl_flags; 4195 return status; 4196 } 4197 4198 struct nfs4_lockdata { 4199 struct nfs_lock_args arg; 4200 struct nfs_lock_res res; 4201 struct nfs4_lock_state *lsp; 4202 struct nfs_open_context *ctx; 4203 struct file_lock fl; 4204 unsigned long timestamp; 4205 int rpc_status; 4206 int cancelled; 4207 struct nfs_server *server; 4208 }; 4209 4210 static struct nfs4_lockdata *nfs4_alloc_lockdata(struct file_lock *fl, 4211 struct nfs_open_context *ctx, struct nfs4_lock_state *lsp, 4212 gfp_t gfp_mask) 4213 { 4214 struct nfs4_lockdata *p; 4215 struct inode *inode = lsp->ls_state->inode; 4216 struct nfs_server *server = NFS_SERVER(inode); 4217 4218 p = kzalloc(sizeof(*p), gfp_mask); 4219 if (p == NULL) 4220 return NULL; 4221 4222 p->arg.fh = NFS_FH(inode); 4223 p->arg.fl = &p->fl; 4224 p->arg.open_seqid = nfs_alloc_seqid(&lsp->ls_state->owner->so_seqid, gfp_mask); 4225 if (p->arg.open_seqid == NULL) 4226 goto out_free; 4227 p->arg.lock_seqid = nfs_alloc_seqid(&lsp->ls_seqid, gfp_mask); 4228 if (p->arg.lock_seqid == NULL) 4229 goto out_free_seqid; 4230 p->arg.lock_stateid = &lsp->ls_stateid; 4231 p->arg.lock_owner.clientid = server->nfs_client->cl_clientid; 4232 p->arg.lock_owner.id = lsp->ls_id.id; 4233 p->arg.lock_owner.s_dev = server->s_dev; 4234 p->res.lock_seqid = p->arg.lock_seqid; 4235 p->lsp = lsp; 4236 p->server = server; 4237 atomic_inc(&lsp->ls_count); 4238 p->ctx = get_nfs_open_context(ctx); 4239 memcpy(&p->fl, fl, sizeof(p->fl)); 4240 return p; 4241 out_free_seqid: 4242 nfs_free_seqid(p->arg.open_seqid); 4243 out_free: 4244 kfree(p); 4245 return NULL; 4246 } 4247 4248 static void nfs4_lock_prepare(struct rpc_task *task, void *calldata) 4249 { 4250 struct nfs4_lockdata *data = calldata; 4251 struct nfs4_state *state = data->lsp->ls_state; 4252 4253 dprintk("%s: begin!\n", __func__); 4254 if (nfs_wait_on_sequence(data->arg.lock_seqid, task) != 0) 4255 return; 4256 /* Do we need to do an open_to_lock_owner? */ 4257 if (!(data->arg.lock_seqid->sequence->flags & NFS_SEQID_CONFIRMED)) { 4258 if (nfs_wait_on_sequence(data->arg.open_seqid, task) != 0) 4259 return; 4260 data->arg.open_stateid = &state->stateid; 4261 data->arg.new_lock_owner = 1; 4262 data->res.open_seqid = data->arg.open_seqid; 4263 } else 4264 data->arg.new_lock_owner = 0; 4265 data->timestamp = jiffies; 4266 if (nfs4_setup_sequence(data->server, 4267 &data->arg.seq_args, 4268 &data->res.seq_res, 1, task)) 4269 return; 4270 rpc_call_start(task); 4271 dprintk("%s: done!, ret = %d\n", __func__, data->rpc_status); 4272 } 4273 4274 static void nfs4_recover_lock_prepare(struct rpc_task *task, void *calldata) 4275 { 4276 rpc_task_set_priority(task, RPC_PRIORITY_PRIVILEGED); 4277 nfs4_lock_prepare(task, calldata); 4278 } 4279 4280 static void nfs4_lock_done(struct rpc_task *task, void *calldata) 4281 { 4282 struct nfs4_lockdata *data = calldata; 4283 4284 dprintk("%s: begin!\n", __func__); 4285 4286 if (!nfs4_sequence_done(task, &data->res.seq_res)) 4287 return; 4288 4289 data->rpc_status = task->tk_status; 4290 if (data->arg.new_lock_owner != 0) { 4291 if (data->rpc_status == 0) 4292 nfs_confirm_seqid(&data->lsp->ls_seqid, 0); 4293 else 4294 goto out; 4295 } 4296 if (data->rpc_status == 0) { 4297 memcpy(data->lsp->ls_stateid.data, data->res.stateid.data, 4298 sizeof(data->lsp->ls_stateid.data)); 4299 data->lsp->ls_flags |= NFS_LOCK_INITIALIZED; 4300 renew_lease(NFS_SERVER(data->ctx->dentry->d_inode), data->timestamp); 4301 } 4302 out: 4303 dprintk("%s: done, ret = %d!\n", __func__, data->rpc_status); 4304 } 4305 4306 static void nfs4_lock_release(void *calldata) 4307 { 4308 struct nfs4_lockdata *data = calldata; 4309 4310 dprintk("%s: begin!\n", __func__); 4311 nfs_free_seqid(data->arg.open_seqid); 4312 if (data->cancelled != 0) { 4313 struct rpc_task *task; 4314 task = nfs4_do_unlck(&data->fl, data->ctx, data->lsp, 4315 data->arg.lock_seqid); 4316 if (!IS_ERR(task)) 4317 rpc_put_task_async(task); 4318 dprintk("%s: cancelling lock!\n", __func__); 4319 } else 4320 nfs_free_seqid(data->arg.lock_seqid); 4321 nfs4_put_lock_state(data->lsp); 4322 put_nfs_open_context(data->ctx); 4323 kfree(data); 4324 dprintk("%s: done!\n", __func__); 4325 } 4326 4327 static const struct rpc_call_ops nfs4_lock_ops = { 4328 .rpc_call_prepare = nfs4_lock_prepare, 4329 .rpc_call_done = nfs4_lock_done, 4330 .rpc_release = nfs4_lock_release, 4331 }; 4332 4333 static const struct rpc_call_ops nfs4_recover_lock_ops = { 4334 .rpc_call_prepare = nfs4_recover_lock_prepare, 4335 .rpc_call_done = nfs4_lock_done, 4336 .rpc_release = nfs4_lock_release, 4337 }; 4338 4339 static void nfs4_handle_setlk_error(struct nfs_server *server, struct nfs4_lock_state *lsp, int new_lock_owner, int error) 4340 { 4341 switch (error) { 4342 case -NFS4ERR_ADMIN_REVOKED: 4343 case -NFS4ERR_BAD_STATEID: 4344 lsp->ls_seqid.flags &= ~NFS_SEQID_CONFIRMED; 4345 if (new_lock_owner != 0 || 4346 (lsp->ls_flags & NFS_LOCK_INITIALIZED) != 0) 4347 nfs4_schedule_stateid_recovery(server, lsp->ls_state); 4348 break; 4349 case -NFS4ERR_STALE_STATEID: 4350 lsp->ls_seqid.flags &= ~NFS_SEQID_CONFIRMED; 4351 case -NFS4ERR_EXPIRED: 4352 nfs4_schedule_lease_recovery(server->nfs_client); 4353 }; 4354 } 4355 4356 static int _nfs4_do_setlk(struct nfs4_state *state, int cmd, struct file_lock *fl, int recovery_type) 4357 { 4358 struct nfs4_lockdata *data; 4359 struct rpc_task *task; 4360 struct rpc_message msg = { 4361 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LOCK], 4362 .rpc_cred = state->owner->so_cred, 4363 }; 4364 struct rpc_task_setup task_setup_data = { 4365 .rpc_client = NFS_CLIENT(state->inode), 4366 .rpc_message = &msg, 4367 .callback_ops = &nfs4_lock_ops, 4368 .workqueue = nfsiod_workqueue, 4369 .flags = RPC_TASK_ASYNC, 4370 }; 4371 int ret; 4372 4373 dprintk("%s: begin!\n", __func__); 4374 data = nfs4_alloc_lockdata(fl, nfs_file_open_context(fl->fl_file), 4375 fl->fl_u.nfs4_fl.owner, 4376 recovery_type == NFS_LOCK_NEW ? GFP_KERNEL : GFP_NOFS); 4377 if (data == NULL) 4378 return -ENOMEM; 4379 if (IS_SETLKW(cmd)) 4380 data->arg.block = 1; 4381 if (recovery_type > NFS_LOCK_NEW) { 4382 if (recovery_type == NFS_LOCK_RECLAIM) 4383 data->arg.reclaim = NFS_LOCK_RECLAIM; 4384 task_setup_data.callback_ops = &nfs4_recover_lock_ops; 4385 } 4386 msg.rpc_argp = &data->arg; 4387 msg.rpc_resp = &data->res; 4388 task_setup_data.callback_data = data; 4389 task = rpc_run_task(&task_setup_data); 4390 if (IS_ERR(task)) 4391 return PTR_ERR(task); 4392 ret = nfs4_wait_for_completion_rpc_task(task); 4393 if (ret == 0) { 4394 ret = data->rpc_status; 4395 if (ret) 4396 nfs4_handle_setlk_error(data->server, data->lsp, 4397 data->arg.new_lock_owner, ret); 4398 } else 4399 data->cancelled = 1; 4400 rpc_put_task(task); 4401 dprintk("%s: done, ret = %d!\n", __func__, ret); 4402 return ret; 4403 } 4404 4405 static int nfs4_lock_reclaim(struct nfs4_state *state, struct file_lock *request) 4406 { 4407 struct nfs_server *server = NFS_SERVER(state->inode); 4408 struct nfs4_exception exception = { }; 4409 int err; 4410 4411 do { 4412 /* Cache the lock if possible... */ 4413 if (test_bit(NFS_DELEGATED_STATE, &state->flags) != 0) 4414 return 0; 4415 err = _nfs4_do_setlk(state, F_SETLK, request, NFS_LOCK_RECLAIM); 4416 if (err != -NFS4ERR_DELAY) 4417 break; 4418 nfs4_handle_exception(server, err, &exception); 4419 } while (exception.retry); 4420 return err; 4421 } 4422 4423 static int nfs4_lock_expired(struct nfs4_state *state, struct file_lock *request) 4424 { 4425 struct nfs_server *server = NFS_SERVER(state->inode); 4426 struct nfs4_exception exception = { }; 4427 int err; 4428 4429 err = nfs4_set_lock_state(state, request); 4430 if (err != 0) 4431 return err; 4432 do { 4433 if (test_bit(NFS_DELEGATED_STATE, &state->flags) != 0) 4434 return 0; 4435 err = _nfs4_do_setlk(state, F_SETLK, request, NFS_LOCK_EXPIRED); 4436 switch (err) { 4437 default: 4438 goto out; 4439 case -NFS4ERR_GRACE: 4440 case -NFS4ERR_DELAY: 4441 nfs4_handle_exception(server, err, &exception); 4442 err = 0; 4443 } 4444 } while (exception.retry); 4445 out: 4446 return err; 4447 } 4448 4449 #if defined(CONFIG_NFS_V4_1) 4450 static int nfs41_lock_expired(struct nfs4_state *state, struct file_lock *request) 4451 { 4452 int status; 4453 struct nfs_server *server = NFS_SERVER(state->inode); 4454 4455 status = nfs41_test_stateid(server, state); 4456 if (status == NFS_OK) 4457 return 0; 4458 nfs41_free_stateid(server, state); 4459 return nfs4_lock_expired(state, request); 4460 } 4461 #endif 4462 4463 static int _nfs4_proc_setlk(struct nfs4_state *state, int cmd, struct file_lock *request) 4464 { 4465 struct nfs_inode *nfsi = NFS_I(state->inode); 4466 unsigned char fl_flags = request->fl_flags; 4467 int status = -ENOLCK; 4468 4469 if ((fl_flags & FL_POSIX) && 4470 !test_bit(NFS_STATE_POSIX_LOCKS, &state->flags)) 4471 goto out; 4472 /* Is this a delegated open? */ 4473 status = nfs4_set_lock_state(state, request); 4474 if (status != 0) 4475 goto out; 4476 request->fl_flags |= FL_ACCESS; 4477 status = do_vfs_lock(request->fl_file, request); 4478 if (status < 0) 4479 goto out; 4480 down_read(&nfsi->rwsem); 4481 if (test_bit(NFS_DELEGATED_STATE, &state->flags)) { 4482 /* Yes: cache locks! */ 4483 /* ...but avoid races with delegation recall... */ 4484 request->fl_flags = fl_flags & ~FL_SLEEP; 4485 status = do_vfs_lock(request->fl_file, request); 4486 goto out_unlock; 4487 } 4488 status = _nfs4_do_setlk(state, cmd, request, NFS_LOCK_NEW); 4489 if (status != 0) 4490 goto out_unlock; 4491 /* Note: we always want to sleep here! */ 4492 request->fl_flags = fl_flags | FL_SLEEP; 4493 if (do_vfs_lock(request->fl_file, request) < 0) 4494 printk(KERN_WARNING "%s: VFS is out of sync with lock manager!\n", __func__); 4495 out_unlock: 4496 up_read(&nfsi->rwsem); 4497 out: 4498 request->fl_flags = fl_flags; 4499 return status; 4500 } 4501 4502 static int nfs4_proc_setlk(struct nfs4_state *state, int cmd, struct file_lock *request) 4503 { 4504 struct nfs4_exception exception = { }; 4505 int err; 4506 4507 do { 4508 err = _nfs4_proc_setlk(state, cmd, request); 4509 if (err == -NFS4ERR_DENIED) 4510 err = -EAGAIN; 4511 err = nfs4_handle_exception(NFS_SERVER(state->inode), 4512 err, &exception); 4513 } while (exception.retry); 4514 return err; 4515 } 4516 4517 static int 4518 nfs4_proc_lock(struct file *filp, int cmd, struct file_lock *request) 4519 { 4520 struct nfs_open_context *ctx; 4521 struct nfs4_state *state; 4522 unsigned long timeout = NFS4_LOCK_MINTIMEOUT; 4523 int status; 4524 4525 /* verify open state */ 4526 ctx = nfs_file_open_context(filp); 4527 state = ctx->state; 4528 4529 if (request->fl_start < 0 || request->fl_end < 0) 4530 return -EINVAL; 4531 4532 if (IS_GETLK(cmd)) { 4533 if (state != NULL) 4534 return nfs4_proc_getlk(state, F_GETLK, request); 4535 return 0; 4536 } 4537 4538 if (!(IS_SETLK(cmd) || IS_SETLKW(cmd))) 4539 return -EINVAL; 4540 4541 if (request->fl_type == F_UNLCK) { 4542 if (state != NULL) 4543 return nfs4_proc_unlck(state, cmd, request); 4544 return 0; 4545 } 4546 4547 if (state == NULL) 4548 return -ENOLCK; 4549 do { 4550 status = nfs4_proc_setlk(state, cmd, request); 4551 if ((status != -EAGAIN) || IS_SETLK(cmd)) 4552 break; 4553 timeout = nfs4_set_lock_task_retry(timeout); 4554 status = -ERESTARTSYS; 4555 if (signalled()) 4556 break; 4557 } while(status < 0); 4558 return status; 4559 } 4560 4561 int nfs4_lock_delegation_recall(struct nfs4_state *state, struct file_lock *fl) 4562 { 4563 struct nfs_server *server = NFS_SERVER(state->inode); 4564 struct nfs4_exception exception = { }; 4565 int err; 4566 4567 err = nfs4_set_lock_state(state, fl); 4568 if (err != 0) 4569 goto out; 4570 do { 4571 err = _nfs4_do_setlk(state, F_SETLK, fl, NFS_LOCK_NEW); 4572 switch (err) { 4573 default: 4574 printk(KERN_ERR "%s: unhandled error %d.\n", 4575 __func__, err); 4576 case 0: 4577 case -ESTALE: 4578 goto out; 4579 case -NFS4ERR_EXPIRED: 4580 nfs4_schedule_stateid_recovery(server, state); 4581 case -NFS4ERR_STALE_CLIENTID: 4582 case -NFS4ERR_STALE_STATEID: 4583 nfs4_schedule_lease_recovery(server->nfs_client); 4584 goto out; 4585 case -NFS4ERR_BADSESSION: 4586 case -NFS4ERR_BADSLOT: 4587 case -NFS4ERR_BAD_HIGH_SLOT: 4588 case -NFS4ERR_CONN_NOT_BOUND_TO_SESSION: 4589 case -NFS4ERR_DEADSESSION: 4590 nfs4_schedule_session_recovery(server->nfs_client->cl_session); 4591 goto out; 4592 case -ERESTARTSYS: 4593 /* 4594 * The show must go on: exit, but mark the 4595 * stateid as needing recovery. 4596 */ 4597 case -NFS4ERR_ADMIN_REVOKED: 4598 case -NFS4ERR_BAD_STATEID: 4599 case -NFS4ERR_OPENMODE: 4600 nfs4_schedule_stateid_recovery(server, state); 4601 err = 0; 4602 goto out; 4603 case -EKEYEXPIRED: 4604 /* 4605 * User RPCSEC_GSS context has expired. 4606 * We cannot recover this stateid now, so 4607 * skip it and allow recovery thread to 4608 * proceed. 4609 */ 4610 err = 0; 4611 goto out; 4612 case -ENOMEM: 4613 case -NFS4ERR_DENIED: 4614 /* kill_proc(fl->fl_pid, SIGLOST, 1); */ 4615 err = 0; 4616 goto out; 4617 case -NFS4ERR_DELAY: 4618 break; 4619 } 4620 err = nfs4_handle_exception(server, err, &exception); 4621 } while (exception.retry); 4622 out: 4623 return err; 4624 } 4625 4626 static void nfs4_release_lockowner_release(void *calldata) 4627 { 4628 kfree(calldata); 4629 } 4630 4631 const struct rpc_call_ops nfs4_release_lockowner_ops = { 4632 .rpc_release = nfs4_release_lockowner_release, 4633 }; 4634 4635 void nfs4_release_lockowner(const struct nfs4_lock_state *lsp) 4636 { 4637 struct nfs_server *server = lsp->ls_state->owner->so_server; 4638 struct nfs_release_lockowner_args *args; 4639 struct rpc_message msg = { 4640 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_RELEASE_LOCKOWNER], 4641 }; 4642 4643 if (server->nfs_client->cl_mvops->minor_version != 0) 4644 return; 4645 args = kmalloc(sizeof(*args), GFP_NOFS); 4646 if (!args) 4647 return; 4648 args->lock_owner.clientid = server->nfs_client->cl_clientid; 4649 args->lock_owner.id = lsp->ls_id.id; 4650 args->lock_owner.s_dev = server->s_dev; 4651 msg.rpc_argp = args; 4652 rpc_call_async(server->client, &msg, 0, &nfs4_release_lockowner_ops, args); 4653 } 4654 4655 #define XATTR_NAME_NFSV4_ACL "system.nfs4_acl" 4656 4657 static int nfs4_xattr_set_nfs4_acl(struct dentry *dentry, const char *key, 4658 const void *buf, size_t buflen, 4659 int flags, int type) 4660 { 4661 if (strcmp(key, "") != 0) 4662 return -EINVAL; 4663 4664 return nfs4_proc_set_acl(dentry->d_inode, buf, buflen); 4665 } 4666 4667 static int nfs4_xattr_get_nfs4_acl(struct dentry *dentry, const char *key, 4668 void *buf, size_t buflen, int type) 4669 { 4670 if (strcmp(key, "") != 0) 4671 return -EINVAL; 4672 4673 return nfs4_proc_get_acl(dentry->d_inode, buf, buflen); 4674 } 4675 4676 static size_t nfs4_xattr_list_nfs4_acl(struct dentry *dentry, char *list, 4677 size_t list_len, const char *name, 4678 size_t name_len, int type) 4679 { 4680 size_t len = sizeof(XATTR_NAME_NFSV4_ACL); 4681 4682 if (!nfs4_server_supports_acls(NFS_SERVER(dentry->d_inode))) 4683 return 0; 4684 4685 if (list && len <= list_len) 4686 memcpy(list, XATTR_NAME_NFSV4_ACL, len); 4687 return len; 4688 } 4689 4690 /* 4691 * nfs_fhget will use either the mounted_on_fileid or the fileid 4692 */ 4693 static void nfs_fixup_referral_attributes(struct nfs_fattr *fattr) 4694 { 4695 if (!(((fattr->valid & NFS_ATTR_FATTR_MOUNTED_ON_FILEID) || 4696 (fattr->valid & NFS_ATTR_FATTR_FILEID)) && 4697 (fattr->valid & NFS_ATTR_FATTR_FSID) && 4698 (fattr->valid & NFS_ATTR_FATTR_V4_REFERRAL))) 4699 return; 4700 4701 fattr->valid |= NFS_ATTR_FATTR_TYPE | NFS_ATTR_FATTR_MODE | 4702 NFS_ATTR_FATTR_NLINK; 4703 fattr->mode = S_IFDIR | S_IRUGO | S_IXUGO; 4704 fattr->nlink = 2; 4705 } 4706 4707 int nfs4_proc_fs_locations(struct inode *dir, const struct qstr *name, 4708 struct nfs4_fs_locations *fs_locations, struct page *page) 4709 { 4710 struct nfs_server *server = NFS_SERVER(dir); 4711 u32 bitmask[2] = { 4712 [0] = FATTR4_WORD0_FSID | FATTR4_WORD0_FS_LOCATIONS, 4713 }; 4714 struct nfs4_fs_locations_arg args = { 4715 .dir_fh = NFS_FH(dir), 4716 .name = name, 4717 .page = page, 4718 .bitmask = bitmask, 4719 }; 4720 struct nfs4_fs_locations_res res = { 4721 .fs_locations = fs_locations, 4722 }; 4723 struct rpc_message msg = { 4724 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_FS_LOCATIONS], 4725 .rpc_argp = &args, 4726 .rpc_resp = &res, 4727 }; 4728 int status; 4729 4730 dprintk("%s: start\n", __func__); 4731 4732 /* Ask for the fileid of the absent filesystem if mounted_on_fileid 4733 * is not supported */ 4734 if (NFS_SERVER(dir)->attr_bitmask[1] & FATTR4_WORD1_MOUNTED_ON_FILEID) 4735 bitmask[1] |= FATTR4_WORD1_MOUNTED_ON_FILEID; 4736 else 4737 bitmask[0] |= FATTR4_WORD0_FILEID; 4738 4739 nfs_fattr_init(&fs_locations->fattr); 4740 fs_locations->server = server; 4741 fs_locations->nlocations = 0; 4742 status = nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 4743 dprintk("%s: returned status = %d\n", __func__, status); 4744 return status; 4745 } 4746 4747 static int _nfs4_proc_secinfo(struct inode *dir, const struct qstr *name, struct nfs4_secinfo_flavors *flavors) 4748 { 4749 int status; 4750 struct nfs4_secinfo_arg args = { 4751 .dir_fh = NFS_FH(dir), 4752 .name = name, 4753 }; 4754 struct nfs4_secinfo_res res = { 4755 .flavors = flavors, 4756 }; 4757 struct rpc_message msg = { 4758 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SECINFO], 4759 .rpc_argp = &args, 4760 .rpc_resp = &res, 4761 }; 4762 4763 dprintk("NFS call secinfo %s\n", name->name); 4764 status = nfs4_call_sync(NFS_SERVER(dir)->client, NFS_SERVER(dir), &msg, &args.seq_args, &res.seq_res, 0); 4765 dprintk("NFS reply secinfo: %d\n", status); 4766 return status; 4767 } 4768 4769 int nfs4_proc_secinfo(struct inode *dir, const struct qstr *name, struct nfs4_secinfo_flavors *flavors) 4770 { 4771 struct nfs4_exception exception = { }; 4772 int err; 4773 do { 4774 err = nfs4_handle_exception(NFS_SERVER(dir), 4775 _nfs4_proc_secinfo(dir, name, flavors), 4776 &exception); 4777 } while (exception.retry); 4778 return err; 4779 } 4780 4781 #ifdef CONFIG_NFS_V4_1 4782 /* 4783 * Check the exchange flags returned by the server for invalid flags, having 4784 * both PNFS and NON_PNFS flags set, and not having one of NON_PNFS, PNFS, or 4785 * DS flags set. 4786 */ 4787 static int nfs4_check_cl_exchange_flags(u32 flags) 4788 { 4789 if (flags & ~EXCHGID4_FLAG_MASK_R) 4790 goto out_inval; 4791 if ((flags & EXCHGID4_FLAG_USE_PNFS_MDS) && 4792 (flags & EXCHGID4_FLAG_USE_NON_PNFS)) 4793 goto out_inval; 4794 if (!(flags & (EXCHGID4_FLAG_MASK_PNFS))) 4795 goto out_inval; 4796 return NFS_OK; 4797 out_inval: 4798 return -NFS4ERR_INVAL; 4799 } 4800 4801 static bool 4802 nfs41_same_server_scope(struct server_scope *a, struct server_scope *b) 4803 { 4804 if (a->server_scope_sz == b->server_scope_sz && 4805 memcmp(a->server_scope, b->server_scope, a->server_scope_sz) == 0) 4806 return true; 4807 4808 return false; 4809 } 4810 4811 /* 4812 * nfs4_proc_exchange_id() 4813 * 4814 * Since the clientid has expired, all compounds using sessions 4815 * associated with the stale clientid will be returning 4816 * NFS4ERR_BADSESSION in the sequence operation, and will therefore 4817 * be in some phase of session reset. 4818 */ 4819 int nfs4_proc_exchange_id(struct nfs_client *clp, struct rpc_cred *cred) 4820 { 4821 nfs4_verifier verifier; 4822 struct nfs41_exchange_id_args args = { 4823 .client = clp, 4824 .flags = EXCHGID4_FLAG_SUPP_MOVED_REFER, 4825 }; 4826 struct nfs41_exchange_id_res res = { 4827 .client = clp, 4828 }; 4829 int status; 4830 struct rpc_message msg = { 4831 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_EXCHANGE_ID], 4832 .rpc_argp = &args, 4833 .rpc_resp = &res, 4834 .rpc_cred = cred, 4835 }; 4836 __be32 *p; 4837 4838 dprintk("--> %s\n", __func__); 4839 BUG_ON(clp == NULL); 4840 4841 p = (u32 *)verifier.data; 4842 *p++ = htonl((u32)clp->cl_boot_time.tv_sec); 4843 *p = htonl((u32)clp->cl_boot_time.tv_nsec); 4844 args.verifier = &verifier; 4845 4846 args.id_len = scnprintf(args.id, sizeof(args.id), 4847 "%s/%s.%s/%u", 4848 clp->cl_ipaddr, 4849 init_utsname()->nodename, 4850 init_utsname()->domainname, 4851 clp->cl_rpcclient->cl_auth->au_flavor); 4852 4853 res.server_scope = kzalloc(sizeof(struct server_scope), GFP_KERNEL); 4854 if (unlikely(!res.server_scope)) 4855 return -ENOMEM; 4856 4857 status = rpc_call_sync(clp->cl_rpcclient, &msg, RPC_TASK_TIMEOUT); 4858 if (!status) 4859 status = nfs4_check_cl_exchange_flags(clp->cl_exchange_flags); 4860 4861 if (!status) { 4862 if (clp->server_scope && 4863 !nfs41_same_server_scope(clp->server_scope, 4864 res.server_scope)) { 4865 dprintk("%s: server_scope mismatch detected\n", 4866 __func__); 4867 set_bit(NFS4CLNT_SERVER_SCOPE_MISMATCH, &clp->cl_state); 4868 kfree(clp->server_scope); 4869 clp->server_scope = NULL; 4870 } 4871 4872 if (!clp->server_scope) 4873 clp->server_scope = res.server_scope; 4874 else 4875 kfree(res.server_scope); 4876 } 4877 4878 dprintk("<-- %s status= %d\n", __func__, status); 4879 return status; 4880 } 4881 4882 struct nfs4_get_lease_time_data { 4883 struct nfs4_get_lease_time_args *args; 4884 struct nfs4_get_lease_time_res *res; 4885 struct nfs_client *clp; 4886 }; 4887 4888 static void nfs4_get_lease_time_prepare(struct rpc_task *task, 4889 void *calldata) 4890 { 4891 int ret; 4892 struct nfs4_get_lease_time_data *data = 4893 (struct nfs4_get_lease_time_data *)calldata; 4894 4895 dprintk("--> %s\n", __func__); 4896 rpc_task_set_priority(task, RPC_PRIORITY_PRIVILEGED); 4897 /* just setup sequence, do not trigger session recovery 4898 since we're invoked within one */ 4899 ret = nfs41_setup_sequence(data->clp->cl_session, 4900 &data->args->la_seq_args, 4901 &data->res->lr_seq_res, 0, task); 4902 4903 BUG_ON(ret == -EAGAIN); 4904 rpc_call_start(task); 4905 dprintk("<-- %s\n", __func__); 4906 } 4907 4908 /* 4909 * Called from nfs4_state_manager thread for session setup, so don't recover 4910 * from sequence operation or clientid errors. 4911 */ 4912 static void nfs4_get_lease_time_done(struct rpc_task *task, void *calldata) 4913 { 4914 struct nfs4_get_lease_time_data *data = 4915 (struct nfs4_get_lease_time_data *)calldata; 4916 4917 dprintk("--> %s\n", __func__); 4918 if (!nfs41_sequence_done(task, &data->res->lr_seq_res)) 4919 return; 4920 switch (task->tk_status) { 4921 case -NFS4ERR_DELAY: 4922 case -NFS4ERR_GRACE: 4923 dprintk("%s Retry: tk_status %d\n", __func__, task->tk_status); 4924 rpc_delay(task, NFS4_POLL_RETRY_MIN); 4925 task->tk_status = 0; 4926 /* fall through */ 4927 case -NFS4ERR_RETRY_UNCACHED_REP: 4928 rpc_restart_call_prepare(task); 4929 return; 4930 } 4931 dprintk("<-- %s\n", __func__); 4932 } 4933 4934 struct rpc_call_ops nfs4_get_lease_time_ops = { 4935 .rpc_call_prepare = nfs4_get_lease_time_prepare, 4936 .rpc_call_done = nfs4_get_lease_time_done, 4937 }; 4938 4939 int nfs4_proc_get_lease_time(struct nfs_client *clp, struct nfs_fsinfo *fsinfo) 4940 { 4941 struct rpc_task *task; 4942 struct nfs4_get_lease_time_args args; 4943 struct nfs4_get_lease_time_res res = { 4944 .lr_fsinfo = fsinfo, 4945 }; 4946 struct nfs4_get_lease_time_data data = { 4947 .args = &args, 4948 .res = &res, 4949 .clp = clp, 4950 }; 4951 struct rpc_message msg = { 4952 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_GET_LEASE_TIME], 4953 .rpc_argp = &args, 4954 .rpc_resp = &res, 4955 }; 4956 struct rpc_task_setup task_setup = { 4957 .rpc_client = clp->cl_rpcclient, 4958 .rpc_message = &msg, 4959 .callback_ops = &nfs4_get_lease_time_ops, 4960 .callback_data = &data, 4961 .flags = RPC_TASK_TIMEOUT, 4962 }; 4963 int status; 4964 4965 dprintk("--> %s\n", __func__); 4966 task = rpc_run_task(&task_setup); 4967 4968 if (IS_ERR(task)) 4969 status = PTR_ERR(task); 4970 else { 4971 status = task->tk_status; 4972 rpc_put_task(task); 4973 } 4974 dprintk("<-- %s return %d\n", __func__, status); 4975 4976 return status; 4977 } 4978 4979 /* 4980 * Reset a slot table 4981 */ 4982 static int nfs4_reset_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, 4983 int ivalue) 4984 { 4985 struct nfs4_slot *new = NULL; 4986 int i; 4987 int ret = 0; 4988 4989 dprintk("--> %s: max_reqs=%u, tbl->max_slots %d\n", __func__, 4990 max_reqs, tbl->max_slots); 4991 4992 /* Does the newly negotiated max_reqs match the existing slot table? */ 4993 if (max_reqs != tbl->max_slots) { 4994 ret = -ENOMEM; 4995 new = kmalloc(max_reqs * sizeof(struct nfs4_slot), 4996 GFP_NOFS); 4997 if (!new) 4998 goto out; 4999 ret = 0; 5000 kfree(tbl->slots); 5001 } 5002 spin_lock(&tbl->slot_tbl_lock); 5003 if (new) { 5004 tbl->slots = new; 5005 tbl->max_slots = max_reqs; 5006 } 5007 for (i = 0; i < tbl->max_slots; ++i) 5008 tbl->slots[i].seq_nr = ivalue; 5009 spin_unlock(&tbl->slot_tbl_lock); 5010 dprintk("%s: tbl=%p slots=%p max_slots=%d\n", __func__, 5011 tbl, tbl->slots, tbl->max_slots); 5012 out: 5013 dprintk("<-- %s: return %d\n", __func__, ret); 5014 return ret; 5015 } 5016 5017 /* 5018 * Reset the forechannel and backchannel slot tables 5019 */ 5020 static int nfs4_reset_slot_tables(struct nfs4_session *session) 5021 { 5022 int status; 5023 5024 status = nfs4_reset_slot_table(&session->fc_slot_table, 5025 session->fc_attrs.max_reqs, 1); 5026 if (status) 5027 return status; 5028 5029 status = nfs4_reset_slot_table(&session->bc_slot_table, 5030 session->bc_attrs.max_reqs, 0); 5031 return status; 5032 } 5033 5034 /* Destroy the slot table */ 5035 static void nfs4_destroy_slot_tables(struct nfs4_session *session) 5036 { 5037 if (session->fc_slot_table.slots != NULL) { 5038 kfree(session->fc_slot_table.slots); 5039 session->fc_slot_table.slots = NULL; 5040 } 5041 if (session->bc_slot_table.slots != NULL) { 5042 kfree(session->bc_slot_table.slots); 5043 session->bc_slot_table.slots = NULL; 5044 } 5045 return; 5046 } 5047 5048 /* 5049 * Initialize slot table 5050 */ 5051 static int nfs4_init_slot_table(struct nfs4_slot_table *tbl, 5052 int max_slots, int ivalue) 5053 { 5054 struct nfs4_slot *slot; 5055 int ret = -ENOMEM; 5056 5057 BUG_ON(max_slots > NFS4_MAX_SLOT_TABLE); 5058 5059 dprintk("--> %s: max_reqs=%u\n", __func__, max_slots); 5060 5061 slot = kcalloc(max_slots, sizeof(struct nfs4_slot), GFP_NOFS); 5062 if (!slot) 5063 goto out; 5064 ret = 0; 5065 5066 spin_lock(&tbl->slot_tbl_lock); 5067 tbl->max_slots = max_slots; 5068 tbl->slots = slot; 5069 tbl->highest_used_slotid = -1; /* no slot is currently used */ 5070 spin_unlock(&tbl->slot_tbl_lock); 5071 dprintk("%s: tbl=%p slots=%p max_slots=%d\n", __func__, 5072 tbl, tbl->slots, tbl->max_slots); 5073 out: 5074 dprintk("<-- %s: return %d\n", __func__, ret); 5075 return ret; 5076 } 5077 5078 /* 5079 * Initialize the forechannel and backchannel tables 5080 */ 5081 static int nfs4_init_slot_tables(struct nfs4_session *session) 5082 { 5083 struct nfs4_slot_table *tbl; 5084 int status = 0; 5085 5086 tbl = &session->fc_slot_table; 5087 if (tbl->slots == NULL) { 5088 status = nfs4_init_slot_table(tbl, 5089 session->fc_attrs.max_reqs, 1); 5090 if (status) 5091 return status; 5092 } 5093 5094 tbl = &session->bc_slot_table; 5095 if (tbl->slots == NULL) { 5096 status = nfs4_init_slot_table(tbl, 5097 session->bc_attrs.max_reqs, 0); 5098 if (status) 5099 nfs4_destroy_slot_tables(session); 5100 } 5101 5102 return status; 5103 } 5104 5105 struct nfs4_session *nfs4_alloc_session(struct nfs_client *clp) 5106 { 5107 struct nfs4_session *session; 5108 struct nfs4_slot_table *tbl; 5109 5110 session = kzalloc(sizeof(struct nfs4_session), GFP_NOFS); 5111 if (!session) 5112 return NULL; 5113 5114 tbl = &session->fc_slot_table; 5115 tbl->highest_used_slotid = -1; 5116 spin_lock_init(&tbl->slot_tbl_lock); 5117 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, "ForeChannel Slot table"); 5118 init_completion(&tbl->complete); 5119 5120 tbl = &session->bc_slot_table; 5121 tbl->highest_used_slotid = -1; 5122 spin_lock_init(&tbl->slot_tbl_lock); 5123 rpc_init_wait_queue(&tbl->slot_tbl_waitq, "BackChannel Slot table"); 5124 init_completion(&tbl->complete); 5125 5126 session->session_state = 1<<NFS4_SESSION_INITING; 5127 5128 session->clp = clp; 5129 return session; 5130 } 5131 5132 void nfs4_destroy_session(struct nfs4_session *session) 5133 { 5134 nfs4_proc_destroy_session(session); 5135 dprintk("%s Destroy backchannel for xprt %p\n", 5136 __func__, session->clp->cl_rpcclient->cl_xprt); 5137 xprt_destroy_backchannel(session->clp->cl_rpcclient->cl_xprt, 5138 NFS41_BC_MIN_CALLBACKS); 5139 nfs4_destroy_slot_tables(session); 5140 kfree(session); 5141 } 5142 5143 /* 5144 * Initialize the values to be used by the client in CREATE_SESSION 5145 * If nfs4_init_session set the fore channel request and response sizes, 5146 * use them. 5147 * 5148 * Set the back channel max_resp_sz_cached to zero to force the client to 5149 * always set csa_cachethis to FALSE because the current implementation 5150 * of the back channel DRC only supports caching the CB_SEQUENCE operation. 5151 */ 5152 static void nfs4_init_channel_attrs(struct nfs41_create_session_args *args) 5153 { 5154 struct nfs4_session *session = args->client->cl_session; 5155 unsigned int mxrqst_sz = session->fc_attrs.max_rqst_sz, 5156 mxresp_sz = session->fc_attrs.max_resp_sz; 5157 5158 if (mxrqst_sz == 0) 5159 mxrqst_sz = NFS_MAX_FILE_IO_SIZE; 5160 if (mxresp_sz == 0) 5161 mxresp_sz = NFS_MAX_FILE_IO_SIZE; 5162 /* Fore channel attributes */ 5163 args->fc_attrs.max_rqst_sz = mxrqst_sz; 5164 args->fc_attrs.max_resp_sz = mxresp_sz; 5165 args->fc_attrs.max_ops = NFS4_MAX_OPS; 5166 args->fc_attrs.max_reqs = session->clp->cl_rpcclient->cl_xprt->max_reqs; 5167 5168 dprintk("%s: Fore Channel : max_rqst_sz=%u max_resp_sz=%u " 5169 "max_ops=%u max_reqs=%u\n", 5170 __func__, 5171 args->fc_attrs.max_rqst_sz, args->fc_attrs.max_resp_sz, 5172 args->fc_attrs.max_ops, args->fc_attrs.max_reqs); 5173 5174 /* Back channel attributes */ 5175 args->bc_attrs.max_rqst_sz = PAGE_SIZE; 5176 args->bc_attrs.max_resp_sz = PAGE_SIZE; 5177 args->bc_attrs.max_resp_sz_cached = 0; 5178 args->bc_attrs.max_ops = NFS4_MAX_BACK_CHANNEL_OPS; 5179 args->bc_attrs.max_reqs = 1; 5180 5181 dprintk("%s: Back Channel : max_rqst_sz=%u max_resp_sz=%u " 5182 "max_resp_sz_cached=%u max_ops=%u max_reqs=%u\n", 5183 __func__, 5184 args->bc_attrs.max_rqst_sz, args->bc_attrs.max_resp_sz, 5185 args->bc_attrs.max_resp_sz_cached, args->bc_attrs.max_ops, 5186 args->bc_attrs.max_reqs); 5187 } 5188 5189 static int nfs4_verify_fore_channel_attrs(struct nfs41_create_session_args *args, struct nfs4_session *session) 5190 { 5191 struct nfs4_channel_attrs *sent = &args->fc_attrs; 5192 struct nfs4_channel_attrs *rcvd = &session->fc_attrs; 5193 5194 if (rcvd->max_resp_sz > sent->max_resp_sz) 5195 return -EINVAL; 5196 /* 5197 * Our requested max_ops is the minimum we need; we're not 5198 * prepared to break up compounds into smaller pieces than that. 5199 * So, no point even trying to continue if the server won't 5200 * cooperate: 5201 */ 5202 if (rcvd->max_ops < sent->max_ops) 5203 return -EINVAL; 5204 if (rcvd->max_reqs == 0) 5205 return -EINVAL; 5206 return 0; 5207 } 5208 5209 static int nfs4_verify_back_channel_attrs(struct nfs41_create_session_args *args, struct nfs4_session *session) 5210 { 5211 struct nfs4_channel_attrs *sent = &args->bc_attrs; 5212 struct nfs4_channel_attrs *rcvd = &session->bc_attrs; 5213 5214 if (rcvd->max_rqst_sz > sent->max_rqst_sz) 5215 return -EINVAL; 5216 if (rcvd->max_resp_sz < sent->max_resp_sz) 5217 return -EINVAL; 5218 if (rcvd->max_resp_sz_cached > sent->max_resp_sz_cached) 5219 return -EINVAL; 5220 /* These would render the backchannel useless: */ 5221 if (rcvd->max_ops == 0) 5222 return -EINVAL; 5223 if (rcvd->max_reqs == 0) 5224 return -EINVAL; 5225 return 0; 5226 } 5227 5228 static int nfs4_verify_channel_attrs(struct nfs41_create_session_args *args, 5229 struct nfs4_session *session) 5230 { 5231 int ret; 5232 5233 ret = nfs4_verify_fore_channel_attrs(args, session); 5234 if (ret) 5235 return ret; 5236 return nfs4_verify_back_channel_attrs(args, session); 5237 } 5238 5239 static int _nfs4_proc_create_session(struct nfs_client *clp) 5240 { 5241 struct nfs4_session *session = clp->cl_session; 5242 struct nfs41_create_session_args args = { 5243 .client = clp, 5244 .cb_program = NFS4_CALLBACK, 5245 }; 5246 struct nfs41_create_session_res res = { 5247 .client = clp, 5248 }; 5249 struct rpc_message msg = { 5250 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_CREATE_SESSION], 5251 .rpc_argp = &args, 5252 .rpc_resp = &res, 5253 }; 5254 int status; 5255 5256 nfs4_init_channel_attrs(&args); 5257 args.flags = (SESSION4_PERSIST | SESSION4_BACK_CHAN); 5258 5259 status = rpc_call_sync(session->clp->cl_rpcclient, &msg, RPC_TASK_TIMEOUT); 5260 5261 if (!status) 5262 /* Verify the session's negotiated channel_attrs values */ 5263 status = nfs4_verify_channel_attrs(&args, session); 5264 if (!status) { 5265 /* Increment the clientid slot sequence id */ 5266 clp->cl_seqid++; 5267 } 5268 5269 return status; 5270 } 5271 5272 /* 5273 * Issues a CREATE_SESSION operation to the server. 5274 * It is the responsibility of the caller to verify the session is 5275 * expired before calling this routine. 5276 */ 5277 int nfs4_proc_create_session(struct nfs_client *clp) 5278 { 5279 int status; 5280 unsigned *ptr; 5281 struct nfs4_session *session = clp->cl_session; 5282 5283 dprintk("--> %s clp=%p session=%p\n", __func__, clp, session); 5284 5285 status = _nfs4_proc_create_session(clp); 5286 if (status) 5287 goto out; 5288 5289 /* Init and reset the fore channel */ 5290 status = nfs4_init_slot_tables(session); 5291 dprintk("slot table initialization returned %d\n", status); 5292 if (status) 5293 goto out; 5294 status = nfs4_reset_slot_tables(session); 5295 dprintk("slot table reset returned %d\n", status); 5296 if (status) 5297 goto out; 5298 5299 ptr = (unsigned *)&session->sess_id.data[0]; 5300 dprintk("%s client>seqid %d sessionid %u:%u:%u:%u\n", __func__, 5301 clp->cl_seqid, ptr[0], ptr[1], ptr[2], ptr[3]); 5302 out: 5303 dprintk("<-- %s\n", __func__); 5304 return status; 5305 } 5306 5307 /* 5308 * Issue the over-the-wire RPC DESTROY_SESSION. 5309 * The caller must serialize access to this routine. 5310 */ 5311 int nfs4_proc_destroy_session(struct nfs4_session *session) 5312 { 5313 int status = 0; 5314 struct rpc_message msg; 5315 5316 dprintk("--> nfs4_proc_destroy_session\n"); 5317 5318 /* session is still being setup */ 5319 if (session->clp->cl_cons_state != NFS_CS_READY) 5320 return status; 5321 5322 msg.rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_DESTROY_SESSION]; 5323 msg.rpc_argp = session; 5324 msg.rpc_resp = NULL; 5325 msg.rpc_cred = NULL; 5326 status = rpc_call_sync(session->clp->cl_rpcclient, &msg, RPC_TASK_TIMEOUT); 5327 5328 if (status) 5329 printk(KERN_WARNING 5330 "Got error %d from the server on DESTROY_SESSION. " 5331 "Session has been destroyed regardless...\n", status); 5332 5333 dprintk("<-- nfs4_proc_destroy_session\n"); 5334 return status; 5335 } 5336 5337 int nfs4_init_session(struct nfs_server *server) 5338 { 5339 struct nfs_client *clp = server->nfs_client; 5340 struct nfs4_session *session; 5341 unsigned int rsize, wsize; 5342 int ret; 5343 5344 if (!nfs4_has_session(clp)) 5345 return 0; 5346 5347 session = clp->cl_session; 5348 if (!test_and_clear_bit(NFS4_SESSION_INITING, &session->session_state)) 5349 return 0; 5350 5351 rsize = server->rsize; 5352 if (rsize == 0) 5353 rsize = NFS_MAX_FILE_IO_SIZE; 5354 wsize = server->wsize; 5355 if (wsize == 0) 5356 wsize = NFS_MAX_FILE_IO_SIZE; 5357 5358 session->fc_attrs.max_rqst_sz = wsize + nfs41_maxwrite_overhead; 5359 session->fc_attrs.max_resp_sz = rsize + nfs41_maxread_overhead; 5360 5361 ret = nfs4_recover_expired_lease(server); 5362 if (!ret) 5363 ret = nfs4_check_client_ready(clp); 5364 return ret; 5365 } 5366 5367 int nfs4_init_ds_session(struct nfs_client *clp) 5368 { 5369 struct nfs4_session *session = clp->cl_session; 5370 int ret; 5371 5372 if (!test_and_clear_bit(NFS4_SESSION_INITING, &session->session_state)) 5373 return 0; 5374 5375 ret = nfs4_client_recover_expired_lease(clp); 5376 if (!ret) 5377 /* Test for the DS role */ 5378 if (!is_ds_client(clp)) 5379 ret = -ENODEV; 5380 if (!ret) 5381 ret = nfs4_check_client_ready(clp); 5382 return ret; 5383 5384 } 5385 EXPORT_SYMBOL_GPL(nfs4_init_ds_session); 5386 5387 5388 /* 5389 * Renew the cl_session lease. 5390 */ 5391 struct nfs4_sequence_data { 5392 struct nfs_client *clp; 5393 struct nfs4_sequence_args args; 5394 struct nfs4_sequence_res res; 5395 }; 5396 5397 static void nfs41_sequence_release(void *data) 5398 { 5399 struct nfs4_sequence_data *calldata = data; 5400 struct nfs_client *clp = calldata->clp; 5401 5402 if (atomic_read(&clp->cl_count) > 1) 5403 nfs4_schedule_state_renewal(clp); 5404 nfs_put_client(clp); 5405 kfree(calldata); 5406 } 5407 5408 static int nfs41_sequence_handle_errors(struct rpc_task *task, struct nfs_client *clp) 5409 { 5410 switch(task->tk_status) { 5411 case -NFS4ERR_DELAY: 5412 rpc_delay(task, NFS4_POLL_RETRY_MAX); 5413 return -EAGAIN; 5414 default: 5415 nfs4_schedule_lease_recovery(clp); 5416 } 5417 return 0; 5418 } 5419 5420 static void nfs41_sequence_call_done(struct rpc_task *task, void *data) 5421 { 5422 struct nfs4_sequence_data *calldata = data; 5423 struct nfs_client *clp = calldata->clp; 5424 5425 if (!nfs41_sequence_done(task, task->tk_msg.rpc_resp)) 5426 return; 5427 5428 if (task->tk_status < 0) { 5429 dprintk("%s ERROR %d\n", __func__, task->tk_status); 5430 if (atomic_read(&clp->cl_count) == 1) 5431 goto out; 5432 5433 if (nfs41_sequence_handle_errors(task, clp) == -EAGAIN) { 5434 rpc_restart_call_prepare(task); 5435 return; 5436 } 5437 } 5438 dprintk("%s rpc_cred %p\n", __func__, task->tk_msg.rpc_cred); 5439 out: 5440 dprintk("<-- %s\n", __func__); 5441 } 5442 5443 static void nfs41_sequence_prepare(struct rpc_task *task, void *data) 5444 { 5445 struct nfs4_sequence_data *calldata = data; 5446 struct nfs_client *clp = calldata->clp; 5447 struct nfs4_sequence_args *args; 5448 struct nfs4_sequence_res *res; 5449 5450 args = task->tk_msg.rpc_argp; 5451 res = task->tk_msg.rpc_resp; 5452 5453 if (nfs41_setup_sequence(clp->cl_session, args, res, 0, task)) 5454 return; 5455 rpc_call_start(task); 5456 } 5457 5458 static const struct rpc_call_ops nfs41_sequence_ops = { 5459 .rpc_call_done = nfs41_sequence_call_done, 5460 .rpc_call_prepare = nfs41_sequence_prepare, 5461 .rpc_release = nfs41_sequence_release, 5462 }; 5463 5464 static struct rpc_task *_nfs41_proc_sequence(struct nfs_client *clp, struct rpc_cred *cred) 5465 { 5466 struct nfs4_sequence_data *calldata; 5467 struct rpc_message msg = { 5468 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SEQUENCE], 5469 .rpc_cred = cred, 5470 }; 5471 struct rpc_task_setup task_setup_data = { 5472 .rpc_client = clp->cl_rpcclient, 5473 .rpc_message = &msg, 5474 .callback_ops = &nfs41_sequence_ops, 5475 .flags = RPC_TASK_ASYNC | RPC_TASK_SOFT, 5476 }; 5477 5478 if (!atomic_inc_not_zero(&clp->cl_count)) 5479 return ERR_PTR(-EIO); 5480 calldata = kzalloc(sizeof(*calldata), GFP_NOFS); 5481 if (calldata == NULL) { 5482 nfs_put_client(clp); 5483 return ERR_PTR(-ENOMEM); 5484 } 5485 msg.rpc_argp = &calldata->args; 5486 msg.rpc_resp = &calldata->res; 5487 calldata->clp = clp; 5488 task_setup_data.callback_data = calldata; 5489 5490 return rpc_run_task(&task_setup_data); 5491 } 5492 5493 static int nfs41_proc_async_sequence(struct nfs_client *clp, struct rpc_cred *cred, unsigned renew_flags) 5494 { 5495 struct rpc_task *task; 5496 int ret = 0; 5497 5498 if ((renew_flags & NFS4_RENEW_TIMEOUT) == 0) 5499 return 0; 5500 task = _nfs41_proc_sequence(clp, cred); 5501 if (IS_ERR(task)) 5502 ret = PTR_ERR(task); 5503 else 5504 rpc_put_task_async(task); 5505 dprintk("<-- %s status=%d\n", __func__, ret); 5506 return ret; 5507 } 5508 5509 static int nfs4_proc_sequence(struct nfs_client *clp, struct rpc_cred *cred) 5510 { 5511 struct rpc_task *task; 5512 int ret; 5513 5514 task = _nfs41_proc_sequence(clp, cred); 5515 if (IS_ERR(task)) { 5516 ret = PTR_ERR(task); 5517 goto out; 5518 } 5519 ret = rpc_wait_for_completion_task(task); 5520 if (!ret) { 5521 struct nfs4_sequence_res *res = task->tk_msg.rpc_resp; 5522 5523 if (task->tk_status == 0) 5524 nfs41_handle_sequence_flag_errors(clp, res->sr_status_flags); 5525 ret = task->tk_status; 5526 } 5527 rpc_put_task(task); 5528 out: 5529 dprintk("<-- %s status=%d\n", __func__, ret); 5530 return ret; 5531 } 5532 5533 struct nfs4_reclaim_complete_data { 5534 struct nfs_client *clp; 5535 struct nfs41_reclaim_complete_args arg; 5536 struct nfs41_reclaim_complete_res res; 5537 }; 5538 5539 static void nfs4_reclaim_complete_prepare(struct rpc_task *task, void *data) 5540 { 5541 struct nfs4_reclaim_complete_data *calldata = data; 5542 5543 rpc_task_set_priority(task, RPC_PRIORITY_PRIVILEGED); 5544 if (nfs41_setup_sequence(calldata->clp->cl_session, 5545 &calldata->arg.seq_args, 5546 &calldata->res.seq_res, 0, task)) 5547 return; 5548 5549 rpc_call_start(task); 5550 } 5551 5552 static int nfs41_reclaim_complete_handle_errors(struct rpc_task *task, struct nfs_client *clp) 5553 { 5554 switch(task->tk_status) { 5555 case 0: 5556 case -NFS4ERR_COMPLETE_ALREADY: 5557 case -NFS4ERR_WRONG_CRED: /* What to do here? */ 5558 break; 5559 case -NFS4ERR_DELAY: 5560 rpc_delay(task, NFS4_POLL_RETRY_MAX); 5561 /* fall through */ 5562 case -NFS4ERR_RETRY_UNCACHED_REP: 5563 return -EAGAIN; 5564 default: 5565 nfs4_schedule_lease_recovery(clp); 5566 } 5567 return 0; 5568 } 5569 5570 static void nfs4_reclaim_complete_done(struct rpc_task *task, void *data) 5571 { 5572 struct nfs4_reclaim_complete_data *calldata = data; 5573 struct nfs_client *clp = calldata->clp; 5574 struct nfs4_sequence_res *res = &calldata->res.seq_res; 5575 5576 dprintk("--> %s\n", __func__); 5577 if (!nfs41_sequence_done(task, res)) 5578 return; 5579 5580 if (nfs41_reclaim_complete_handle_errors(task, clp) == -EAGAIN) { 5581 rpc_restart_call_prepare(task); 5582 return; 5583 } 5584 dprintk("<-- %s\n", __func__); 5585 } 5586 5587 static void nfs4_free_reclaim_complete_data(void *data) 5588 { 5589 struct nfs4_reclaim_complete_data *calldata = data; 5590 5591 kfree(calldata); 5592 } 5593 5594 static const struct rpc_call_ops nfs4_reclaim_complete_call_ops = { 5595 .rpc_call_prepare = nfs4_reclaim_complete_prepare, 5596 .rpc_call_done = nfs4_reclaim_complete_done, 5597 .rpc_release = nfs4_free_reclaim_complete_data, 5598 }; 5599 5600 /* 5601 * Issue a global reclaim complete. 5602 */ 5603 static int nfs41_proc_reclaim_complete(struct nfs_client *clp) 5604 { 5605 struct nfs4_reclaim_complete_data *calldata; 5606 struct rpc_task *task; 5607 struct rpc_message msg = { 5608 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_RECLAIM_COMPLETE], 5609 }; 5610 struct rpc_task_setup task_setup_data = { 5611 .rpc_client = clp->cl_rpcclient, 5612 .rpc_message = &msg, 5613 .callback_ops = &nfs4_reclaim_complete_call_ops, 5614 .flags = RPC_TASK_ASYNC, 5615 }; 5616 int status = -ENOMEM; 5617 5618 dprintk("--> %s\n", __func__); 5619 calldata = kzalloc(sizeof(*calldata), GFP_NOFS); 5620 if (calldata == NULL) 5621 goto out; 5622 calldata->clp = clp; 5623 calldata->arg.one_fs = 0; 5624 5625 msg.rpc_argp = &calldata->arg; 5626 msg.rpc_resp = &calldata->res; 5627 task_setup_data.callback_data = calldata; 5628 task = rpc_run_task(&task_setup_data); 5629 if (IS_ERR(task)) { 5630 status = PTR_ERR(task); 5631 goto out; 5632 } 5633 status = nfs4_wait_for_completion_rpc_task(task); 5634 if (status == 0) 5635 status = task->tk_status; 5636 rpc_put_task(task); 5637 return 0; 5638 out: 5639 dprintk("<-- %s status=%d\n", __func__, status); 5640 return status; 5641 } 5642 5643 static void 5644 nfs4_layoutget_prepare(struct rpc_task *task, void *calldata) 5645 { 5646 struct nfs4_layoutget *lgp = calldata; 5647 struct nfs_server *server = NFS_SERVER(lgp->args.inode); 5648 5649 dprintk("--> %s\n", __func__); 5650 /* Note the is a race here, where a CB_LAYOUTRECALL can come in 5651 * right now covering the LAYOUTGET we are about to send. 5652 * However, that is not so catastrophic, and there seems 5653 * to be no way to prevent it completely. 5654 */ 5655 if (nfs4_setup_sequence(server, &lgp->args.seq_args, 5656 &lgp->res.seq_res, 0, task)) 5657 return; 5658 if (pnfs_choose_layoutget_stateid(&lgp->args.stateid, 5659 NFS_I(lgp->args.inode)->layout, 5660 lgp->args.ctx->state)) { 5661 rpc_exit(task, NFS4_OK); 5662 return; 5663 } 5664 rpc_call_start(task); 5665 } 5666 5667 static void nfs4_layoutget_done(struct rpc_task *task, void *calldata) 5668 { 5669 struct nfs4_layoutget *lgp = calldata; 5670 struct nfs_server *server = NFS_SERVER(lgp->args.inode); 5671 5672 dprintk("--> %s\n", __func__); 5673 5674 if (!nfs4_sequence_done(task, &lgp->res.seq_res)) 5675 return; 5676 5677 switch (task->tk_status) { 5678 case 0: 5679 break; 5680 case -NFS4ERR_LAYOUTTRYLATER: 5681 case -NFS4ERR_RECALLCONFLICT: 5682 task->tk_status = -NFS4ERR_DELAY; 5683 /* Fall through */ 5684 default: 5685 if (nfs4_async_handle_error(task, server, NULL) == -EAGAIN) { 5686 rpc_restart_call_prepare(task); 5687 return; 5688 } 5689 } 5690 dprintk("<-- %s\n", __func__); 5691 } 5692 5693 static void nfs4_layoutget_release(void *calldata) 5694 { 5695 struct nfs4_layoutget *lgp = calldata; 5696 5697 dprintk("--> %s\n", __func__); 5698 put_nfs_open_context(lgp->args.ctx); 5699 kfree(calldata); 5700 dprintk("<-- %s\n", __func__); 5701 } 5702 5703 static const struct rpc_call_ops nfs4_layoutget_call_ops = { 5704 .rpc_call_prepare = nfs4_layoutget_prepare, 5705 .rpc_call_done = nfs4_layoutget_done, 5706 .rpc_release = nfs4_layoutget_release, 5707 }; 5708 5709 int nfs4_proc_layoutget(struct nfs4_layoutget *lgp) 5710 { 5711 struct nfs_server *server = NFS_SERVER(lgp->args.inode); 5712 struct rpc_task *task; 5713 struct rpc_message msg = { 5714 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LAYOUTGET], 5715 .rpc_argp = &lgp->args, 5716 .rpc_resp = &lgp->res, 5717 }; 5718 struct rpc_task_setup task_setup_data = { 5719 .rpc_client = server->client, 5720 .rpc_message = &msg, 5721 .callback_ops = &nfs4_layoutget_call_ops, 5722 .callback_data = lgp, 5723 .flags = RPC_TASK_ASYNC, 5724 }; 5725 int status = 0; 5726 5727 dprintk("--> %s\n", __func__); 5728 5729 lgp->res.layoutp = &lgp->args.layout; 5730 lgp->res.seq_res.sr_slot = NULL; 5731 task = rpc_run_task(&task_setup_data); 5732 if (IS_ERR(task)) 5733 return PTR_ERR(task); 5734 status = nfs4_wait_for_completion_rpc_task(task); 5735 if (status == 0) 5736 status = task->tk_status; 5737 if (status == 0) 5738 status = pnfs_layout_process(lgp); 5739 rpc_put_task(task); 5740 dprintk("<-- %s status=%d\n", __func__, status); 5741 return status; 5742 } 5743 5744 static void 5745 nfs4_layoutreturn_prepare(struct rpc_task *task, void *calldata) 5746 { 5747 struct nfs4_layoutreturn *lrp = calldata; 5748 5749 dprintk("--> %s\n", __func__); 5750 if (nfs41_setup_sequence(lrp->clp->cl_session, &lrp->args.seq_args, 5751 &lrp->res.seq_res, 0, task)) 5752 return; 5753 rpc_call_start(task); 5754 } 5755 5756 static void nfs4_layoutreturn_done(struct rpc_task *task, void *calldata) 5757 { 5758 struct nfs4_layoutreturn *lrp = calldata; 5759 struct nfs_server *server; 5760 struct pnfs_layout_hdr *lo = lrp->args.layout; 5761 5762 dprintk("--> %s\n", __func__); 5763 5764 if (!nfs4_sequence_done(task, &lrp->res.seq_res)) 5765 return; 5766 5767 server = NFS_SERVER(lrp->args.inode); 5768 if (nfs4_async_handle_error(task, server, NULL) == -EAGAIN) { 5769 rpc_restart_call_prepare(task); 5770 return; 5771 } 5772 spin_lock(&lo->plh_inode->i_lock); 5773 if (task->tk_status == 0) { 5774 if (lrp->res.lrs_present) { 5775 pnfs_set_layout_stateid(lo, &lrp->res.stateid, true); 5776 } else 5777 BUG_ON(!list_empty(&lo->plh_segs)); 5778 } 5779 lo->plh_block_lgets--; 5780 spin_unlock(&lo->plh_inode->i_lock); 5781 dprintk("<-- %s\n", __func__); 5782 } 5783 5784 static void nfs4_layoutreturn_release(void *calldata) 5785 { 5786 struct nfs4_layoutreturn *lrp = calldata; 5787 5788 dprintk("--> %s\n", __func__); 5789 put_layout_hdr(lrp->args.layout); 5790 kfree(calldata); 5791 dprintk("<-- %s\n", __func__); 5792 } 5793 5794 static const struct rpc_call_ops nfs4_layoutreturn_call_ops = { 5795 .rpc_call_prepare = nfs4_layoutreturn_prepare, 5796 .rpc_call_done = nfs4_layoutreturn_done, 5797 .rpc_release = nfs4_layoutreturn_release, 5798 }; 5799 5800 int nfs4_proc_layoutreturn(struct nfs4_layoutreturn *lrp) 5801 { 5802 struct rpc_task *task; 5803 struct rpc_message msg = { 5804 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LAYOUTRETURN], 5805 .rpc_argp = &lrp->args, 5806 .rpc_resp = &lrp->res, 5807 }; 5808 struct rpc_task_setup task_setup_data = { 5809 .rpc_client = lrp->clp->cl_rpcclient, 5810 .rpc_message = &msg, 5811 .callback_ops = &nfs4_layoutreturn_call_ops, 5812 .callback_data = lrp, 5813 }; 5814 int status; 5815 5816 dprintk("--> %s\n", __func__); 5817 task = rpc_run_task(&task_setup_data); 5818 if (IS_ERR(task)) 5819 return PTR_ERR(task); 5820 status = task->tk_status; 5821 dprintk("<-- %s status=%d\n", __func__, status); 5822 rpc_put_task(task); 5823 return status; 5824 } 5825 5826 /* 5827 * Retrieve the list of Data Server devices from the MDS. 5828 */ 5829 static int _nfs4_getdevicelist(struct nfs_server *server, 5830 const struct nfs_fh *fh, 5831 struct pnfs_devicelist *devlist) 5832 { 5833 struct nfs4_getdevicelist_args args = { 5834 .fh = fh, 5835 .layoutclass = server->pnfs_curr_ld->id, 5836 }; 5837 struct nfs4_getdevicelist_res res = { 5838 .devlist = devlist, 5839 }; 5840 struct rpc_message msg = { 5841 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_GETDEVICELIST], 5842 .rpc_argp = &args, 5843 .rpc_resp = &res, 5844 }; 5845 int status; 5846 5847 dprintk("--> %s\n", __func__); 5848 status = nfs4_call_sync(server->client, server, &msg, &args.seq_args, 5849 &res.seq_res, 0); 5850 dprintk("<-- %s status=%d\n", __func__, status); 5851 return status; 5852 } 5853 5854 int nfs4_proc_getdevicelist(struct nfs_server *server, 5855 const struct nfs_fh *fh, 5856 struct pnfs_devicelist *devlist) 5857 { 5858 struct nfs4_exception exception = { }; 5859 int err; 5860 5861 do { 5862 err = nfs4_handle_exception(server, 5863 _nfs4_getdevicelist(server, fh, devlist), 5864 &exception); 5865 } while (exception.retry); 5866 5867 dprintk("%s: err=%d, num_devs=%u\n", __func__, 5868 err, devlist->num_devs); 5869 5870 return err; 5871 } 5872 EXPORT_SYMBOL_GPL(nfs4_proc_getdevicelist); 5873 5874 static int 5875 _nfs4_proc_getdeviceinfo(struct nfs_server *server, struct pnfs_device *pdev) 5876 { 5877 struct nfs4_getdeviceinfo_args args = { 5878 .pdev = pdev, 5879 }; 5880 struct nfs4_getdeviceinfo_res res = { 5881 .pdev = pdev, 5882 }; 5883 struct rpc_message msg = { 5884 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_GETDEVICEINFO], 5885 .rpc_argp = &args, 5886 .rpc_resp = &res, 5887 }; 5888 int status; 5889 5890 dprintk("--> %s\n", __func__); 5891 status = nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 5892 dprintk("<-- %s status=%d\n", __func__, status); 5893 5894 return status; 5895 } 5896 5897 int nfs4_proc_getdeviceinfo(struct nfs_server *server, struct pnfs_device *pdev) 5898 { 5899 struct nfs4_exception exception = { }; 5900 int err; 5901 5902 do { 5903 err = nfs4_handle_exception(server, 5904 _nfs4_proc_getdeviceinfo(server, pdev), 5905 &exception); 5906 } while (exception.retry); 5907 return err; 5908 } 5909 EXPORT_SYMBOL_GPL(nfs4_proc_getdeviceinfo); 5910 5911 static void nfs4_layoutcommit_prepare(struct rpc_task *task, void *calldata) 5912 { 5913 struct nfs4_layoutcommit_data *data = calldata; 5914 struct nfs_server *server = NFS_SERVER(data->args.inode); 5915 5916 if (nfs4_setup_sequence(server, &data->args.seq_args, 5917 &data->res.seq_res, 1, task)) 5918 return; 5919 rpc_call_start(task); 5920 } 5921 5922 static void 5923 nfs4_layoutcommit_done(struct rpc_task *task, void *calldata) 5924 { 5925 struct nfs4_layoutcommit_data *data = calldata; 5926 struct nfs_server *server = NFS_SERVER(data->args.inode); 5927 5928 if (!nfs4_sequence_done(task, &data->res.seq_res)) 5929 return; 5930 5931 switch (task->tk_status) { /* Just ignore these failures */ 5932 case NFS4ERR_DELEG_REVOKED: /* layout was recalled */ 5933 case NFS4ERR_BADIOMODE: /* no IOMODE_RW layout for range */ 5934 case NFS4ERR_BADLAYOUT: /* no layout */ 5935 case NFS4ERR_GRACE: /* loca_recalim always false */ 5936 task->tk_status = 0; 5937 } 5938 5939 if (nfs4_async_handle_error(task, server, NULL) == -EAGAIN) { 5940 rpc_restart_call_prepare(task); 5941 return; 5942 } 5943 5944 if (task->tk_status == 0) 5945 nfs_post_op_update_inode_force_wcc(data->args.inode, 5946 data->res.fattr); 5947 } 5948 5949 static void nfs4_layoutcommit_release(void *calldata) 5950 { 5951 struct nfs4_layoutcommit_data *data = calldata; 5952 struct pnfs_layout_segment *lseg, *tmp; 5953 unsigned long *bitlock = &NFS_I(data->args.inode)->flags; 5954 5955 pnfs_cleanup_layoutcommit(data); 5956 /* Matched by references in pnfs_set_layoutcommit */ 5957 list_for_each_entry_safe(lseg, tmp, &data->lseg_list, pls_lc_list) { 5958 list_del_init(&lseg->pls_lc_list); 5959 if (test_and_clear_bit(NFS_LSEG_LAYOUTCOMMIT, 5960 &lseg->pls_flags)) 5961 put_lseg(lseg); 5962 } 5963 5964 clear_bit_unlock(NFS_INO_LAYOUTCOMMITTING, bitlock); 5965 smp_mb__after_clear_bit(); 5966 wake_up_bit(bitlock, NFS_INO_LAYOUTCOMMITTING); 5967 5968 put_rpccred(data->cred); 5969 kfree(data); 5970 } 5971 5972 static const struct rpc_call_ops nfs4_layoutcommit_ops = { 5973 .rpc_call_prepare = nfs4_layoutcommit_prepare, 5974 .rpc_call_done = nfs4_layoutcommit_done, 5975 .rpc_release = nfs4_layoutcommit_release, 5976 }; 5977 5978 int 5979 nfs4_proc_layoutcommit(struct nfs4_layoutcommit_data *data, bool sync) 5980 { 5981 struct rpc_message msg = { 5982 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_LAYOUTCOMMIT], 5983 .rpc_argp = &data->args, 5984 .rpc_resp = &data->res, 5985 .rpc_cred = data->cred, 5986 }; 5987 struct rpc_task_setup task_setup_data = { 5988 .task = &data->task, 5989 .rpc_client = NFS_CLIENT(data->args.inode), 5990 .rpc_message = &msg, 5991 .callback_ops = &nfs4_layoutcommit_ops, 5992 .callback_data = data, 5993 .flags = RPC_TASK_ASYNC, 5994 }; 5995 struct rpc_task *task; 5996 int status = 0; 5997 5998 dprintk("NFS: %4d initiating layoutcommit call. sync %d " 5999 "lbw: %llu inode %lu\n", 6000 data->task.tk_pid, sync, 6001 data->args.lastbytewritten, 6002 data->args.inode->i_ino); 6003 6004 task = rpc_run_task(&task_setup_data); 6005 if (IS_ERR(task)) 6006 return PTR_ERR(task); 6007 if (sync == false) 6008 goto out; 6009 status = nfs4_wait_for_completion_rpc_task(task); 6010 if (status != 0) 6011 goto out; 6012 status = task->tk_status; 6013 out: 6014 dprintk("%s: status %d\n", __func__, status); 6015 rpc_put_task(task); 6016 return status; 6017 } 6018 6019 static int 6020 _nfs41_proc_secinfo_no_name(struct nfs_server *server, struct nfs_fh *fhandle, 6021 struct nfs_fsinfo *info, struct nfs4_secinfo_flavors *flavors) 6022 { 6023 struct nfs41_secinfo_no_name_args args = { 6024 .style = SECINFO_STYLE_CURRENT_FH, 6025 }; 6026 struct nfs4_secinfo_res res = { 6027 .flavors = flavors, 6028 }; 6029 struct rpc_message msg = { 6030 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_SECINFO_NO_NAME], 6031 .rpc_argp = &args, 6032 .rpc_resp = &res, 6033 }; 6034 return nfs4_call_sync(server->client, server, &msg, &args.seq_args, &res.seq_res, 0); 6035 } 6036 6037 static int 6038 nfs41_proc_secinfo_no_name(struct nfs_server *server, struct nfs_fh *fhandle, 6039 struct nfs_fsinfo *info, struct nfs4_secinfo_flavors *flavors) 6040 { 6041 struct nfs4_exception exception = { }; 6042 int err; 6043 do { 6044 err = _nfs41_proc_secinfo_no_name(server, fhandle, info, flavors); 6045 switch (err) { 6046 case 0: 6047 case -NFS4ERR_WRONGSEC: 6048 case -NFS4ERR_NOTSUPP: 6049 break; 6050 default: 6051 err = nfs4_handle_exception(server, err, &exception); 6052 } 6053 } while (exception.retry); 6054 return err; 6055 } 6056 6057 static int 6058 nfs41_find_root_sec(struct nfs_server *server, struct nfs_fh *fhandle, 6059 struct nfs_fsinfo *info) 6060 { 6061 int err; 6062 struct page *page; 6063 rpc_authflavor_t flavor; 6064 struct nfs4_secinfo_flavors *flavors; 6065 6066 page = alloc_page(GFP_KERNEL); 6067 if (!page) { 6068 err = -ENOMEM; 6069 goto out; 6070 } 6071 6072 flavors = page_address(page); 6073 err = nfs41_proc_secinfo_no_name(server, fhandle, info, flavors); 6074 6075 /* 6076 * Fall back on "guess and check" method if 6077 * the server doesn't support SECINFO_NO_NAME 6078 */ 6079 if (err == -NFS4ERR_WRONGSEC || err == -NFS4ERR_NOTSUPP) { 6080 err = nfs4_find_root_sec(server, fhandle, info); 6081 goto out_freepage; 6082 } 6083 if (err) 6084 goto out_freepage; 6085 6086 flavor = nfs_find_best_sec(flavors); 6087 if (err == 0) 6088 err = nfs4_lookup_root_sec(server, fhandle, info, flavor); 6089 6090 out_freepage: 6091 put_page(page); 6092 if (err == -EACCES) 6093 return -EPERM; 6094 out: 6095 return err; 6096 } 6097 static int _nfs41_test_stateid(struct nfs_server *server, struct nfs4_state *state) 6098 { 6099 int status; 6100 struct nfs41_test_stateid_args args = { 6101 .stateid = &state->stateid, 6102 }; 6103 struct nfs41_test_stateid_res res; 6104 struct rpc_message msg = { 6105 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_TEST_STATEID], 6106 .rpc_argp = &args, 6107 .rpc_resp = &res, 6108 }; 6109 args.seq_args.sa_session = res.seq_res.sr_session = NULL; 6110 status = nfs4_call_sync_sequence(server->client, server, &msg, &args.seq_args, &res.seq_res, 0, 1); 6111 return status; 6112 } 6113 6114 static int nfs41_test_stateid(struct nfs_server *server, struct nfs4_state *state) 6115 { 6116 struct nfs4_exception exception = { }; 6117 int err; 6118 do { 6119 err = nfs4_handle_exception(server, 6120 _nfs41_test_stateid(server, state), 6121 &exception); 6122 } while (exception.retry); 6123 return err; 6124 } 6125 6126 static int _nfs4_free_stateid(struct nfs_server *server, struct nfs4_state *state) 6127 { 6128 int status; 6129 struct nfs41_free_stateid_args args = { 6130 .stateid = &state->stateid, 6131 }; 6132 struct nfs41_free_stateid_res res; 6133 struct rpc_message msg = { 6134 .rpc_proc = &nfs4_procedures[NFSPROC4_CLNT_FREE_STATEID], 6135 .rpc_argp = &args, 6136 .rpc_resp = &res, 6137 }; 6138 6139 args.seq_args.sa_session = res.seq_res.sr_session = NULL; 6140 status = nfs4_call_sync_sequence(server->client, server, &msg, &args.seq_args, &res.seq_res, 0, 1); 6141 return status; 6142 } 6143 6144 static int nfs41_free_stateid(struct nfs_server *server, struct nfs4_state *state) 6145 { 6146 struct nfs4_exception exception = { }; 6147 int err; 6148 do { 6149 err = nfs4_handle_exception(server, 6150 _nfs4_free_stateid(server, state), 6151 &exception); 6152 } while (exception.retry); 6153 return err; 6154 } 6155 #endif /* CONFIG_NFS_V4_1 */ 6156 6157 struct nfs4_state_recovery_ops nfs40_reboot_recovery_ops = { 6158 .owner_flag_bit = NFS_OWNER_RECLAIM_REBOOT, 6159 .state_flag_bit = NFS_STATE_RECLAIM_REBOOT, 6160 .recover_open = nfs4_open_reclaim, 6161 .recover_lock = nfs4_lock_reclaim, 6162 .establish_clid = nfs4_init_clientid, 6163 .get_clid_cred = nfs4_get_setclientid_cred, 6164 }; 6165 6166 #if defined(CONFIG_NFS_V4_1) 6167 struct nfs4_state_recovery_ops nfs41_reboot_recovery_ops = { 6168 .owner_flag_bit = NFS_OWNER_RECLAIM_REBOOT, 6169 .state_flag_bit = NFS_STATE_RECLAIM_REBOOT, 6170 .recover_open = nfs4_open_reclaim, 6171 .recover_lock = nfs4_lock_reclaim, 6172 .establish_clid = nfs41_init_clientid, 6173 .get_clid_cred = nfs4_get_exchange_id_cred, 6174 .reclaim_complete = nfs41_proc_reclaim_complete, 6175 }; 6176 #endif /* CONFIG_NFS_V4_1 */ 6177 6178 struct nfs4_state_recovery_ops nfs40_nograce_recovery_ops = { 6179 .owner_flag_bit = NFS_OWNER_RECLAIM_NOGRACE, 6180 .state_flag_bit = NFS_STATE_RECLAIM_NOGRACE, 6181 .recover_open = nfs4_open_expired, 6182 .recover_lock = nfs4_lock_expired, 6183 .establish_clid = nfs4_init_clientid, 6184 .get_clid_cred = nfs4_get_setclientid_cred, 6185 }; 6186 6187 #if defined(CONFIG_NFS_V4_1) 6188 struct nfs4_state_recovery_ops nfs41_nograce_recovery_ops = { 6189 .owner_flag_bit = NFS_OWNER_RECLAIM_NOGRACE, 6190 .state_flag_bit = NFS_STATE_RECLAIM_NOGRACE, 6191 .recover_open = nfs41_open_expired, 6192 .recover_lock = nfs41_lock_expired, 6193 .establish_clid = nfs41_init_clientid, 6194 .get_clid_cred = nfs4_get_exchange_id_cred, 6195 }; 6196 #endif /* CONFIG_NFS_V4_1 */ 6197 6198 struct nfs4_state_maintenance_ops nfs40_state_renewal_ops = { 6199 .sched_state_renewal = nfs4_proc_async_renew, 6200 .get_state_renewal_cred_locked = nfs4_get_renew_cred_locked, 6201 .renew_lease = nfs4_proc_renew, 6202 }; 6203 6204 #if defined(CONFIG_NFS_V4_1) 6205 struct nfs4_state_maintenance_ops nfs41_state_renewal_ops = { 6206 .sched_state_renewal = nfs41_proc_async_sequence, 6207 .get_state_renewal_cred_locked = nfs4_get_machine_cred_locked, 6208 .renew_lease = nfs4_proc_sequence, 6209 }; 6210 #endif 6211 6212 static const struct nfs4_minor_version_ops nfs_v4_0_minor_ops = { 6213 .minor_version = 0, 6214 .call_sync = _nfs4_call_sync, 6215 .validate_stateid = nfs4_validate_delegation_stateid, 6216 .find_root_sec = nfs4_find_root_sec, 6217 .reboot_recovery_ops = &nfs40_reboot_recovery_ops, 6218 .nograce_recovery_ops = &nfs40_nograce_recovery_ops, 6219 .state_renewal_ops = &nfs40_state_renewal_ops, 6220 }; 6221 6222 #if defined(CONFIG_NFS_V4_1) 6223 static const struct nfs4_minor_version_ops nfs_v4_1_minor_ops = { 6224 .minor_version = 1, 6225 .call_sync = _nfs4_call_sync_session, 6226 .validate_stateid = nfs41_validate_delegation_stateid, 6227 .find_root_sec = nfs41_find_root_sec, 6228 .reboot_recovery_ops = &nfs41_reboot_recovery_ops, 6229 .nograce_recovery_ops = &nfs41_nograce_recovery_ops, 6230 .state_renewal_ops = &nfs41_state_renewal_ops, 6231 }; 6232 #endif 6233 6234 const struct nfs4_minor_version_ops *nfs_v4_minor_ops[] = { 6235 [0] = &nfs_v4_0_minor_ops, 6236 #if defined(CONFIG_NFS_V4_1) 6237 [1] = &nfs_v4_1_minor_ops, 6238 #endif 6239 }; 6240 6241 static const struct inode_operations nfs4_file_inode_operations = { 6242 .permission = nfs_permission, 6243 .getattr = nfs_getattr, 6244 .setattr = nfs_setattr, 6245 .getxattr = generic_getxattr, 6246 .setxattr = generic_setxattr, 6247 .listxattr = generic_listxattr, 6248 .removexattr = generic_removexattr, 6249 }; 6250 6251 const struct nfs_rpc_ops nfs_v4_clientops = { 6252 .version = 4, /* protocol version */ 6253 .dentry_ops = &nfs4_dentry_operations, 6254 .dir_inode_ops = &nfs4_dir_inode_operations, 6255 .file_inode_ops = &nfs4_file_inode_operations, 6256 .getroot = nfs4_proc_get_root, 6257 .getattr = nfs4_proc_getattr, 6258 .setattr = nfs4_proc_setattr, 6259 .lookup = nfs4_proc_lookup, 6260 .access = nfs4_proc_access, 6261 .readlink = nfs4_proc_readlink, 6262 .create = nfs4_proc_create, 6263 .remove = nfs4_proc_remove, 6264 .unlink_setup = nfs4_proc_unlink_setup, 6265 .unlink_done = nfs4_proc_unlink_done, 6266 .rename = nfs4_proc_rename, 6267 .rename_setup = nfs4_proc_rename_setup, 6268 .rename_done = nfs4_proc_rename_done, 6269 .link = nfs4_proc_link, 6270 .symlink = nfs4_proc_symlink, 6271 .mkdir = nfs4_proc_mkdir, 6272 .rmdir = nfs4_proc_remove, 6273 .readdir = nfs4_proc_readdir, 6274 .mknod = nfs4_proc_mknod, 6275 .statfs = nfs4_proc_statfs, 6276 .fsinfo = nfs4_proc_fsinfo, 6277 .pathconf = nfs4_proc_pathconf, 6278 .set_capabilities = nfs4_server_capabilities, 6279 .decode_dirent = nfs4_decode_dirent, 6280 .read_setup = nfs4_proc_read_setup, 6281 .read_done = nfs4_read_done, 6282 .write_setup = nfs4_proc_write_setup, 6283 .write_done = nfs4_write_done, 6284 .commit_setup = nfs4_proc_commit_setup, 6285 .commit_done = nfs4_commit_done, 6286 .lock = nfs4_proc_lock, 6287 .clear_acl_cache = nfs4_zap_acl_attr, 6288 .close_context = nfs4_close_context, 6289 .open_context = nfs4_atomic_open, 6290 .init_client = nfs4_init_client, 6291 .secinfo = nfs4_proc_secinfo, 6292 }; 6293 6294 static const struct xattr_handler nfs4_xattr_nfs4_acl_handler = { 6295 .prefix = XATTR_NAME_NFSV4_ACL, 6296 .list = nfs4_xattr_list_nfs4_acl, 6297 .get = nfs4_xattr_get_nfs4_acl, 6298 .set = nfs4_xattr_set_nfs4_acl, 6299 }; 6300 6301 const struct xattr_handler *nfs4_xattr_handlers[] = { 6302 &nfs4_xattr_nfs4_acl_handler, 6303 NULL 6304 }; 6305 6306 /* 6307 * Local variables: 6308 * c-basic-offset: 8 6309 * End: 6310 */ 6311