1 /* 2 * Copyright (c) 2016, Mellanox Technologies inc. All rights reserved. 3 * 4 * This software is available to you under a choice of one of two 5 * licenses. You may choose to be licensed under the terms of the GNU 6 * General Public License (GPL) Version 2, available from the file 7 * COPYING in the main directory of this source tree, or the 8 * OpenIB.org BSD license below: 9 * 10 * Redistribution and use in source and binary forms, with or 11 * without modification, are permitted provided that the following 12 * conditions are met: 13 * 14 * - Redistributions of source code must retain the above 15 * copyright notice, this list of conditions and the following 16 * disclaimer. 17 * 18 * - Redistributions in binary form must reproduce the above 19 * copyright notice, this list of conditions and the following 20 * disclaimer in the documentation and/or other materials 21 * provided with the distribution. 22 * 23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 30 * SOFTWARE. 31 */ 32 33 #include <linux/file.h> 34 #include <linux/anon_inodes.h> 35 #include <linux/sched/mm.h> 36 #include <rdma/ib_verbs.h> 37 #include <rdma/uverbs_types.h> 38 #include <linux/rcupdate.h> 39 #include <rdma/uverbs_ioctl.h> 40 #include <rdma/rdma_user_ioctl.h> 41 #include "uverbs.h" 42 #include "core_priv.h" 43 #include "rdma_core.h" 44 45 static void uverbs_uobject_free(struct kref *ref) 46 { 47 kfree_rcu(container_of(ref, struct ib_uobject, ref), rcu); 48 } 49 50 /* 51 * In order to indicate we no longer needs this uobject, uverbs_uobject_put 52 * is called. When the reference count is decreased, the uobject is freed. 53 * For example, this is used when attaching a completion channel to a CQ. 54 */ 55 void uverbs_uobject_put(struct ib_uobject *uobject) 56 { 57 kref_put(&uobject->ref, uverbs_uobject_free); 58 } 59 EXPORT_SYMBOL(uverbs_uobject_put); 60 61 int uverbs_try_lock_object(struct ib_uobject *uobj, 62 enum rdma_lookup_mode mode) 63 { 64 /* 65 * When a shared access is required, we use a positive counter. Each 66 * shared access request checks that the value != -1 and increment it. 67 * Exclusive access is required for operations like write or destroy. 68 * In exclusive access mode, we check that the counter is zero (nobody 69 * claimed this object) and we set it to -1. Releasing a shared access 70 * lock is done simply by decreasing the counter. As for exclusive 71 * access locks, since only a single one of them is allowed 72 * concurrently, setting the counter to zero is enough for releasing 73 * this lock. 74 */ 75 switch (mode) { 76 case UVERBS_LOOKUP_READ: 77 return atomic_fetch_add_unless(&uobj->usecnt, 1, -1) == -1 ? 78 -EBUSY : 0; 79 case UVERBS_LOOKUP_WRITE: 80 /* lock is exclusive */ 81 return atomic_cmpxchg(&uobj->usecnt, 0, -1) == 0 ? 0 : -EBUSY; 82 case UVERBS_LOOKUP_DESTROY: 83 return 0; 84 } 85 return 0; 86 } 87 EXPORT_SYMBOL(uverbs_try_lock_object); 88 89 static void assert_uverbs_usecnt(struct ib_uobject *uobj, 90 enum rdma_lookup_mode mode) 91 { 92 #ifdef CONFIG_LOCKDEP 93 switch (mode) { 94 case UVERBS_LOOKUP_READ: 95 WARN_ON(atomic_read(&uobj->usecnt) <= 0); 96 break; 97 case UVERBS_LOOKUP_WRITE: 98 WARN_ON(atomic_read(&uobj->usecnt) != -1); 99 break; 100 case UVERBS_LOOKUP_DESTROY: 101 break; 102 } 103 #endif 104 } 105 106 /* 107 * This must be called with the hw_destroy_rwsem locked for read or write, 108 * also the uobject itself must be locked for write. 109 * 110 * Upon return the HW object is guaranteed to be destroyed. 111 * 112 * For RDMA_REMOVE_ABORT, the hw_destroy_rwsem is not required to be held, 113 * however the type's allocat_commit function cannot have been called and the 114 * uobject cannot be on the uobjects_lists 115 * 116 * For RDMA_REMOVE_DESTROY the caller should be holding a kref (eg via 117 * rdma_lookup_get_uobject) and the object is left in a state where the caller 118 * needs to call rdma_lookup_put_uobject. 119 * 120 * For all other destroy modes this function internally unlocks the uobject 121 * and consumes the kref on the uobj. 122 */ 123 static int uverbs_destroy_uobject(struct ib_uobject *uobj, 124 enum rdma_remove_reason reason, 125 struct uverbs_attr_bundle *attrs) 126 { 127 struct ib_uverbs_file *ufile = attrs->ufile; 128 unsigned long flags; 129 int ret; 130 131 lockdep_assert_held(&ufile->hw_destroy_rwsem); 132 assert_uverbs_usecnt(uobj, UVERBS_LOOKUP_WRITE); 133 134 if (reason == RDMA_REMOVE_ABORT) { 135 WARN_ON(!list_empty(&uobj->list)); 136 WARN_ON(!uobj->context); 137 uobj->uapi_object->type_class->alloc_abort(uobj); 138 } else if (uobj->object) { 139 ret = uobj->uapi_object->type_class->destroy_hw(uobj, reason, 140 attrs); 141 if (ret) 142 /* Nothing to be done, wait till ucontext will clean it */ 143 return ret; 144 145 uobj->object = NULL; 146 } 147 148 uobj->context = NULL; 149 150 /* 151 * For DESTROY the usecnt is not changed, the caller is expected to 152 * manage it via uobj_put_destroy(). Only DESTROY can remove the IDR 153 * handle. 154 */ 155 if (reason != RDMA_REMOVE_DESTROY) 156 atomic_set(&uobj->usecnt, 0); 157 else 158 uobj->uapi_object->type_class->remove_handle(uobj); 159 160 if (!list_empty(&uobj->list)) { 161 spin_lock_irqsave(&ufile->uobjects_lock, flags); 162 list_del_init(&uobj->list); 163 spin_unlock_irqrestore(&ufile->uobjects_lock, flags); 164 165 /* 166 * Pairs with the get in rdma_alloc_commit_uobject(), could 167 * destroy uobj. 168 */ 169 uverbs_uobject_put(uobj); 170 } 171 172 /* 173 * When aborting the stack kref remains owned by the core code, and is 174 * not transferred into the type. Pairs with the get in alloc_uobj 175 */ 176 if (reason == RDMA_REMOVE_ABORT) 177 uverbs_uobject_put(uobj); 178 179 return 0; 180 } 181 182 /* 183 * This calls uverbs_destroy_uobject() using the RDMA_REMOVE_DESTROY 184 * sequence. It should only be used from command callbacks. On success the 185 * caller must pair this with uobj_put_destroy(). This 186 * version requires the caller to have already obtained an 187 * LOOKUP_DESTROY uobject kref. 188 */ 189 int uobj_destroy(struct ib_uobject *uobj, struct uverbs_attr_bundle *attrs) 190 { 191 struct ib_uverbs_file *ufile = attrs->ufile; 192 int ret; 193 194 down_read(&ufile->hw_destroy_rwsem); 195 196 /* 197 * Once the uobject is destroyed by RDMA_REMOVE_DESTROY then it is left 198 * write locked as the callers put it back with UVERBS_LOOKUP_DESTROY. 199 * This is because any other concurrent thread can still see the object 200 * in the xarray due to RCU. Leaving it locked ensures nothing else will 201 * touch it. 202 */ 203 ret = uverbs_try_lock_object(uobj, UVERBS_LOOKUP_WRITE); 204 if (ret) 205 goto out_unlock; 206 207 ret = uverbs_destroy_uobject(uobj, RDMA_REMOVE_DESTROY, attrs); 208 if (ret) { 209 atomic_set(&uobj->usecnt, 0); 210 goto out_unlock; 211 } 212 213 out_unlock: 214 up_read(&ufile->hw_destroy_rwsem); 215 return ret; 216 } 217 218 /* 219 * uobj_get_destroy destroys the HW object and returns a handle to the uobj 220 * with a NULL object pointer. The caller must pair this with 221 * uobj_put_destroy(). 222 */ 223 struct ib_uobject *__uobj_get_destroy(const struct uverbs_api_object *obj, 224 u32 id, struct uverbs_attr_bundle *attrs) 225 { 226 struct ib_uobject *uobj; 227 int ret; 228 229 uobj = rdma_lookup_get_uobject(obj, attrs->ufile, id, 230 UVERBS_LOOKUP_DESTROY, attrs); 231 if (IS_ERR(uobj)) 232 return uobj; 233 234 ret = uobj_destroy(uobj, attrs); 235 if (ret) { 236 rdma_lookup_put_uobject(uobj, UVERBS_LOOKUP_DESTROY); 237 return ERR_PTR(ret); 238 } 239 240 return uobj; 241 } 242 243 /* 244 * Does both uobj_get_destroy() and uobj_put_destroy(). Returns 0 on success 245 * (negative errno on failure). For use by callers that do not need the uobj. 246 */ 247 int __uobj_perform_destroy(const struct uverbs_api_object *obj, u32 id, 248 struct uverbs_attr_bundle *attrs) 249 { 250 struct ib_uobject *uobj; 251 252 uobj = __uobj_get_destroy(obj, id, attrs); 253 if (IS_ERR(uobj)) 254 return PTR_ERR(uobj); 255 uobj_put_destroy(uobj); 256 return 0; 257 } 258 259 /* alloc_uobj must be undone by uverbs_destroy_uobject() */ 260 static struct ib_uobject *alloc_uobj(struct uverbs_attr_bundle *attrs, 261 const struct uverbs_api_object *obj) 262 { 263 struct ib_uverbs_file *ufile = attrs->ufile; 264 struct ib_uobject *uobj; 265 266 if (!attrs->context) { 267 struct ib_ucontext *ucontext = 268 ib_uverbs_get_ucontext_file(ufile); 269 270 if (IS_ERR(ucontext)) 271 return ERR_CAST(ucontext); 272 attrs->context = ucontext; 273 } 274 275 uobj = kzalloc(obj->type_attrs->obj_size, GFP_KERNEL); 276 if (!uobj) 277 return ERR_PTR(-ENOMEM); 278 /* 279 * user_handle should be filled by the handler, 280 * The object is added to the list in the commit stage. 281 */ 282 uobj->ufile = ufile; 283 uobj->context = attrs->context; 284 INIT_LIST_HEAD(&uobj->list); 285 uobj->uapi_object = obj; 286 /* 287 * Allocated objects start out as write locked to deny any other 288 * syscalls from accessing them until they are committed. See 289 * rdma_alloc_commit_uobject 290 */ 291 atomic_set(&uobj->usecnt, -1); 292 kref_init(&uobj->ref); 293 294 return uobj; 295 } 296 297 static int idr_add_uobj(struct ib_uobject *uobj) 298 { 299 /* 300 * We start with allocating an idr pointing to NULL. This represents an 301 * object which isn't initialized yet. We'll replace it later on with 302 * the real object once we commit. 303 */ 304 return xa_alloc(&uobj->ufile->idr, &uobj->id, NULL, xa_limit_32b, 305 GFP_KERNEL); 306 } 307 308 /* Returns the ib_uobject or an error. The caller should check for IS_ERR. */ 309 static struct ib_uobject * 310 lookup_get_idr_uobject(const struct uverbs_api_object *obj, 311 struct ib_uverbs_file *ufile, s64 id, 312 enum rdma_lookup_mode mode) 313 { 314 struct ib_uobject *uobj; 315 316 if (id < 0 || id > ULONG_MAX) 317 return ERR_PTR(-EINVAL); 318 319 rcu_read_lock(); 320 /* 321 * The idr_find is guaranteed to return a pointer to something that 322 * isn't freed yet, or NULL, as the free after idr_remove goes through 323 * kfree_rcu(). However the object may still have been released and 324 * kfree() could be called at any time. 325 */ 326 uobj = xa_load(&ufile->idr, id); 327 if (!uobj || !kref_get_unless_zero(&uobj->ref)) 328 uobj = ERR_PTR(-ENOENT); 329 rcu_read_unlock(); 330 return uobj; 331 } 332 333 static struct ib_uobject * 334 lookup_get_fd_uobject(const struct uverbs_api_object *obj, 335 struct ib_uverbs_file *ufile, s64 id, 336 enum rdma_lookup_mode mode) 337 { 338 const struct uverbs_obj_fd_type *fd_type; 339 struct file *f; 340 struct ib_uobject *uobject; 341 int fdno = id; 342 343 if (fdno != id) 344 return ERR_PTR(-EINVAL); 345 346 if (mode != UVERBS_LOOKUP_READ) 347 return ERR_PTR(-EOPNOTSUPP); 348 349 if (!obj->type_attrs) 350 return ERR_PTR(-EIO); 351 fd_type = 352 container_of(obj->type_attrs, struct uverbs_obj_fd_type, type); 353 354 f = fget(fdno); 355 if (!f) 356 return ERR_PTR(-EBADF); 357 358 uobject = f->private_data; 359 /* 360 * fget(id) ensures we are not currently running 361 * uverbs_uobject_fd_release(), and the caller is expected to ensure 362 * that release is never done while a call to lookup is possible. 363 */ 364 if (f->f_op != fd_type->fops || uobject->ufile != ufile) { 365 fput(f); 366 return ERR_PTR(-EBADF); 367 } 368 369 uverbs_uobject_get(uobject); 370 return uobject; 371 } 372 373 struct ib_uobject *rdma_lookup_get_uobject(const struct uverbs_api_object *obj, 374 struct ib_uverbs_file *ufile, s64 id, 375 enum rdma_lookup_mode mode, 376 struct uverbs_attr_bundle *attrs) 377 { 378 struct ib_uobject *uobj; 379 int ret; 380 381 if (obj == ERR_PTR(-ENOMSG)) { 382 /* must be UVERBS_IDR_ANY_OBJECT, see uapi_get_object() */ 383 uobj = lookup_get_idr_uobject(NULL, ufile, id, mode); 384 if (IS_ERR(uobj)) 385 return uobj; 386 } else { 387 if (IS_ERR(obj)) 388 return ERR_PTR(-EINVAL); 389 390 uobj = obj->type_class->lookup_get(obj, ufile, id, mode); 391 if (IS_ERR(uobj)) 392 return uobj; 393 394 if (uobj->uapi_object != obj) { 395 ret = -EINVAL; 396 goto free; 397 } 398 } 399 400 /* 401 * If we have been disassociated block every command except for 402 * DESTROY based commands. 403 */ 404 if (mode != UVERBS_LOOKUP_DESTROY && 405 !srcu_dereference(ufile->device->ib_dev, 406 &ufile->device->disassociate_srcu)) { 407 ret = -EIO; 408 goto free; 409 } 410 411 ret = uverbs_try_lock_object(uobj, mode); 412 if (ret) 413 goto free; 414 if (attrs) 415 attrs->context = uobj->context; 416 417 return uobj; 418 free: 419 uobj->uapi_object->type_class->lookup_put(uobj, mode); 420 uverbs_uobject_put(uobj); 421 return ERR_PTR(ret); 422 } 423 424 static struct ib_uobject * 425 alloc_begin_idr_uobject(const struct uverbs_api_object *obj, 426 struct uverbs_attr_bundle *attrs) 427 { 428 int ret; 429 struct ib_uobject *uobj; 430 431 uobj = alloc_uobj(attrs, obj); 432 if (IS_ERR(uobj)) 433 return uobj; 434 435 ret = idr_add_uobj(uobj); 436 if (ret) 437 goto uobj_put; 438 439 ret = ib_rdmacg_try_charge(&uobj->cg_obj, uobj->context->device, 440 RDMACG_RESOURCE_HCA_OBJECT); 441 if (ret) 442 goto remove; 443 444 return uobj; 445 446 remove: 447 xa_erase(&attrs->ufile->idr, uobj->id); 448 uobj_put: 449 uverbs_uobject_put(uobj); 450 return ERR_PTR(ret); 451 } 452 453 static struct ib_uobject * 454 alloc_begin_fd_uobject(const struct uverbs_api_object *obj, 455 struct uverbs_attr_bundle *attrs) 456 { 457 const struct uverbs_obj_fd_type *fd_type; 458 int new_fd; 459 struct ib_uobject *uobj, *ret; 460 struct file *filp; 461 462 uobj = alloc_uobj(attrs, obj); 463 if (IS_ERR(uobj)) 464 return uobj; 465 466 fd_type = 467 container_of(obj->type_attrs, struct uverbs_obj_fd_type, type); 468 if (WARN_ON(fd_type->fops && fd_type->fops->release != &uverbs_uobject_fd_release && 469 fd_type->fops->release != &uverbs_async_event_release)) { 470 ret = ERR_PTR(-EINVAL); 471 goto err_fd; 472 } 473 474 new_fd = get_unused_fd_flags(O_CLOEXEC); 475 if (new_fd < 0) { 476 ret = ERR_PTR(new_fd); 477 goto err_fd; 478 } 479 480 if (fd_type->fops) { 481 /* Note that uverbs_uobject_fd_release() is called during abort */ 482 filp = anon_inode_getfile(fd_type->name, fd_type->fops, NULL, 483 fd_type->flags); 484 if (IS_ERR(filp)) { 485 ret = ERR_CAST(filp); 486 goto err_getfile; 487 } 488 uobj->object = filp; 489 } 490 491 uobj->id = new_fd; 492 return uobj; 493 494 err_getfile: 495 put_unused_fd(new_fd); 496 err_fd: 497 uverbs_uobject_put(uobj); 498 return ret; 499 } 500 501 struct ib_uobject *rdma_alloc_begin_uobject(const struct uverbs_api_object *obj, 502 struct uverbs_attr_bundle *attrs) 503 { 504 struct ib_uverbs_file *ufile = attrs->ufile; 505 struct ib_uobject *ret; 506 507 if (IS_ERR(obj)) 508 return ERR_PTR(-EINVAL); 509 510 /* 511 * The hw_destroy_rwsem is held across the entire object creation and 512 * released during rdma_alloc_commit_uobject or 513 * rdma_alloc_abort_uobject 514 */ 515 if (!down_read_trylock(&ufile->hw_destroy_rwsem)) 516 return ERR_PTR(-EIO); 517 518 ret = obj->type_class->alloc_begin(obj, attrs); 519 if (IS_ERR(ret)) { 520 up_read(&ufile->hw_destroy_rwsem); 521 return ret; 522 } 523 return ret; 524 } 525 526 static void alloc_abort_idr_uobject(struct ib_uobject *uobj) 527 { 528 ib_rdmacg_uncharge(&uobj->cg_obj, uobj->context->device, 529 RDMACG_RESOURCE_HCA_OBJECT); 530 531 xa_erase(&uobj->ufile->idr, uobj->id); 532 } 533 534 static int __must_check destroy_hw_idr_uobject(struct ib_uobject *uobj, 535 enum rdma_remove_reason why, 536 struct uverbs_attr_bundle *attrs) 537 { 538 const struct uverbs_obj_idr_type *idr_type = 539 container_of(uobj->uapi_object->type_attrs, 540 struct uverbs_obj_idr_type, type); 541 int ret = idr_type->destroy_object(uobj, why, attrs); 542 543 if (ret) 544 return ret; 545 546 if (why == RDMA_REMOVE_ABORT) 547 return 0; 548 549 ib_rdmacg_uncharge(&uobj->cg_obj, uobj->context->device, 550 RDMACG_RESOURCE_HCA_OBJECT); 551 552 return 0; 553 } 554 555 static void remove_handle_idr_uobject(struct ib_uobject *uobj) 556 { 557 xa_erase(&uobj->ufile->idr, uobj->id); 558 /* Matches the kref in alloc_commit_idr_uobject */ 559 uverbs_uobject_put(uobj); 560 } 561 562 static void alloc_abort_fd_uobject(struct ib_uobject *uobj) 563 { 564 struct file *filp = uobj->object; 565 566 if (filp) 567 fput(filp); 568 569 put_unused_fd(uobj->id); 570 } 571 572 static int __must_check destroy_hw_fd_uobject(struct ib_uobject *uobj, 573 enum rdma_remove_reason why, 574 struct uverbs_attr_bundle *attrs) 575 { 576 const struct uverbs_obj_fd_type *fd_type = container_of( 577 uobj->uapi_object->type_attrs, struct uverbs_obj_fd_type, type); 578 579 fd_type->destroy_object(uobj, why); 580 return 0; 581 } 582 583 static void remove_handle_fd_uobject(struct ib_uobject *uobj) 584 { 585 } 586 587 static void alloc_commit_idr_uobject(struct ib_uobject *uobj) 588 { 589 struct ib_uverbs_file *ufile = uobj->ufile; 590 void *old; 591 592 /* 593 * We already allocated this IDR with a NULL object, so 594 * this shouldn't fail. 595 * 596 * NOTE: Storing the uobj transfers our kref on uobj to the XArray. 597 * It will be put by remove_commit_idr_uobject() 598 */ 599 old = xa_store(&ufile->idr, uobj->id, uobj, GFP_KERNEL); 600 WARN_ON(old != NULL); 601 } 602 603 static void swap_idr_uobjects(struct ib_uobject *obj_old, 604 struct ib_uobject *obj_new) 605 { 606 struct ib_uverbs_file *ufile = obj_old->ufile; 607 void *old; 608 609 /* 610 * New must be an object that been allocated but not yet committed, this 611 * moves the pre-committed state to obj_old, new still must be comitted. 612 */ 613 old = xa_cmpxchg(&ufile->idr, obj_old->id, obj_old, XA_ZERO_ENTRY, 614 GFP_KERNEL); 615 if (WARN_ON(old != obj_old)) 616 return; 617 618 swap(obj_old->id, obj_new->id); 619 620 old = xa_cmpxchg(&ufile->idr, obj_old->id, NULL, obj_old, GFP_KERNEL); 621 WARN_ON(old != NULL); 622 } 623 624 static void alloc_commit_fd_uobject(struct ib_uobject *uobj) 625 { 626 int fd = uobj->id; 627 struct file *filp = uobj->object; 628 629 /* Matching put will be done in uverbs_uobject_fd_release() */ 630 kref_get(&uobj->ufile->ref); 631 632 /* This shouldn't be used anymore. Use the file object instead */ 633 uobj->id = 0; 634 635 if (!filp->private_data) { 636 /* 637 * NOTE: Once we install the file we loose ownership of our kref on 638 * uobj. It will be put by uverbs_uobject_fd_release() 639 */ 640 filp->private_data = uobj; 641 } 642 643 fd_install(fd, filp); 644 } 645 646 /* 647 * In all cases rdma_alloc_commit_uobject() consumes the kref to uobj and the 648 * caller can no longer assume uobj is valid. If this function fails it 649 * destroys the uboject, including the attached HW object. 650 */ 651 void rdma_alloc_commit_uobject(struct ib_uobject *uobj, 652 struct uverbs_attr_bundle *attrs) 653 { 654 struct ib_uverbs_file *ufile = attrs->ufile; 655 656 /* kref is held so long as the uobj is on the uobj list. */ 657 uverbs_uobject_get(uobj); 658 spin_lock_irq(&ufile->uobjects_lock); 659 list_add(&uobj->list, &ufile->uobjects); 660 spin_unlock_irq(&ufile->uobjects_lock); 661 662 /* matches atomic_set(-1) in alloc_uobj */ 663 atomic_set(&uobj->usecnt, 0); 664 665 /* alloc_commit consumes the uobj kref */ 666 uobj->uapi_object->type_class->alloc_commit(uobj); 667 668 /* Matches the down_read in rdma_alloc_begin_uobject */ 669 up_read(&ufile->hw_destroy_rwsem); 670 } 671 672 /* 673 * new_uobj will be assigned to the handle currently used by to_uobj, and 674 * to_uobj will be destroyed. 675 * 676 * Upon return the caller must do: 677 * rdma_alloc_commit_uobject(new_uobj) 678 * uobj_put_destroy(to_uobj) 679 * 680 * to_uobj must have a write get but the put mode switches to destroy once 681 * this is called. 682 */ 683 void rdma_assign_uobject(struct ib_uobject *to_uobj, struct ib_uobject *new_uobj, 684 struct uverbs_attr_bundle *attrs) 685 { 686 assert_uverbs_usecnt(new_uobj, UVERBS_LOOKUP_WRITE); 687 688 if (WARN_ON(to_uobj->uapi_object != new_uobj->uapi_object || 689 !to_uobj->uapi_object->type_class->swap_uobjects)) 690 return; 691 692 to_uobj->uapi_object->type_class->swap_uobjects(to_uobj, new_uobj); 693 694 /* 695 * If this fails then the uobject is still completely valid (though with 696 * a new ID) and we leak it until context close. 697 */ 698 uverbs_destroy_uobject(to_uobj, RDMA_REMOVE_DESTROY, attrs); 699 } 700 701 /* 702 * This consumes the kref for uobj. It is up to the caller to unwind the HW 703 * object and anything else connected to uobj before calling this. 704 */ 705 void rdma_alloc_abort_uobject(struct ib_uobject *uobj, 706 struct uverbs_attr_bundle *attrs, 707 bool hw_obj_valid) 708 { 709 struct ib_uverbs_file *ufile = uobj->ufile; 710 int ret; 711 712 if (hw_obj_valid) { 713 ret = uobj->uapi_object->type_class->destroy_hw( 714 uobj, RDMA_REMOVE_ABORT, attrs); 715 /* 716 * If the driver couldn't destroy the object then go ahead and 717 * commit it. Leaking objects that can't be destroyed is only 718 * done during FD close after the driver has a few more tries to 719 * destroy it. 720 */ 721 if (WARN_ON(ret)) 722 return rdma_alloc_commit_uobject(uobj, attrs); 723 } 724 725 uverbs_destroy_uobject(uobj, RDMA_REMOVE_ABORT, attrs); 726 727 /* Matches the down_read in rdma_alloc_begin_uobject */ 728 up_read(&ufile->hw_destroy_rwsem); 729 } 730 731 static void lookup_put_idr_uobject(struct ib_uobject *uobj, 732 enum rdma_lookup_mode mode) 733 { 734 } 735 736 static void lookup_put_fd_uobject(struct ib_uobject *uobj, 737 enum rdma_lookup_mode mode) 738 { 739 struct file *filp = uobj->object; 740 741 WARN_ON(mode != UVERBS_LOOKUP_READ); 742 /* 743 * This indirectly calls uverbs_uobject_fd_release() and free the 744 * object 745 */ 746 fput(filp); 747 } 748 749 void rdma_lookup_put_uobject(struct ib_uobject *uobj, 750 enum rdma_lookup_mode mode) 751 { 752 assert_uverbs_usecnt(uobj, mode); 753 /* 754 * In order to unlock an object, either decrease its usecnt for 755 * read access or zero it in case of exclusive access. See 756 * uverbs_try_lock_object for locking schema information. 757 */ 758 switch (mode) { 759 case UVERBS_LOOKUP_READ: 760 atomic_dec(&uobj->usecnt); 761 break; 762 case UVERBS_LOOKUP_WRITE: 763 atomic_set(&uobj->usecnt, 0); 764 break; 765 case UVERBS_LOOKUP_DESTROY: 766 break; 767 } 768 769 uobj->uapi_object->type_class->lookup_put(uobj, mode); 770 /* Pairs with the kref obtained by type->lookup_get */ 771 uverbs_uobject_put(uobj); 772 } 773 774 void setup_ufile_idr_uobject(struct ib_uverbs_file *ufile) 775 { 776 xa_init_flags(&ufile->idr, XA_FLAGS_ALLOC); 777 } 778 779 void release_ufile_idr_uobject(struct ib_uverbs_file *ufile) 780 { 781 struct ib_uobject *entry; 782 unsigned long id; 783 784 /* 785 * At this point uverbs_cleanup_ufile() is guaranteed to have run, and 786 * there are no HW objects left, however the xarray is still populated 787 * with anything that has not been cleaned up by userspace. Since the 788 * kref on ufile is 0, nothing is allowed to call lookup_get. 789 * 790 * This is an optimized equivalent to remove_handle_idr_uobject 791 */ 792 xa_for_each(&ufile->idr, id, entry) { 793 WARN_ON(entry->object); 794 uverbs_uobject_put(entry); 795 } 796 797 xa_destroy(&ufile->idr); 798 } 799 800 const struct uverbs_obj_type_class uverbs_idr_class = { 801 .alloc_begin = alloc_begin_idr_uobject, 802 .lookup_get = lookup_get_idr_uobject, 803 .alloc_commit = alloc_commit_idr_uobject, 804 .alloc_abort = alloc_abort_idr_uobject, 805 .lookup_put = lookup_put_idr_uobject, 806 .destroy_hw = destroy_hw_idr_uobject, 807 .remove_handle = remove_handle_idr_uobject, 808 .swap_uobjects = swap_idr_uobjects, 809 }; 810 EXPORT_SYMBOL(uverbs_idr_class); 811 812 int uverbs_uobject_release(struct ib_uobject *uobj) 813 { 814 struct ib_uverbs_file *ufile; 815 816 ufile = uobj->ufile; 817 818 if (down_read_trylock(&ufile->hw_destroy_rwsem)) { 819 struct uverbs_attr_bundle attrs = { 820 .context = uobj->context, 821 .ufile = ufile, 822 }; 823 824 /* 825 * lookup_get_fd_uobject holds the kref on the struct file any 826 * time a FD uobj is locked, which prevents this release 827 * method from being invoked. Meaning we can always get the 828 * write lock here, or we have a kernel bug. 829 */ 830 WARN_ON(uverbs_try_lock_object(uobj, UVERBS_LOOKUP_WRITE)); 831 uverbs_destroy_uobject(uobj, RDMA_REMOVE_CLOSE, &attrs); 832 up_read(&ufile->hw_destroy_rwsem); 833 } 834 835 /* Matches the get in alloc_commit_fd_uobject() */ 836 kref_put(&ufile->ref, ib_uverbs_release_file); 837 838 /* Pairs with filp->private_data in alloc_begin_fd_uobject */ 839 uverbs_uobject_put(uobj); 840 return 0; 841 } 842 843 /* 844 * Users of UVERBS_TYPE_ALLOC_FD should set this function as the struct 845 * file_operations release method. 846 */ 847 int uverbs_uobject_fd_release(struct inode *inode, struct file *filp) 848 { 849 /* 850 * This can only happen if the fput came from alloc_abort_fd_uobject() 851 */ 852 if (!filp->private_data) 853 return 0; 854 855 return uverbs_uobject_release(filp->private_data); 856 } 857 EXPORT_SYMBOL(uverbs_uobject_fd_release); 858 859 /* 860 * Drop the ucontext off the ufile and completely disconnect it from the 861 * ib_device 862 */ 863 static void ufile_destroy_ucontext(struct ib_uverbs_file *ufile, 864 enum rdma_remove_reason reason) 865 { 866 struct ib_ucontext *ucontext = ufile->ucontext; 867 struct ib_device *ib_dev = ucontext->device; 868 869 /* 870 * If we are closing the FD then the user mmap VMAs must have 871 * already been destroyed as they hold on to the filep, otherwise 872 * they need to be zap'd. 873 */ 874 if (reason == RDMA_REMOVE_DRIVER_REMOVE) { 875 uverbs_user_mmap_disassociate(ufile); 876 if (ib_dev->ops.disassociate_ucontext) 877 ib_dev->ops.disassociate_ucontext(ucontext); 878 } 879 880 ib_rdmacg_uncharge(&ucontext->cg_obj, ib_dev, 881 RDMACG_RESOURCE_HCA_HANDLE); 882 883 rdma_restrack_del(&ucontext->res); 884 885 ib_dev->ops.dealloc_ucontext(ucontext); 886 WARN_ON(!xa_empty(&ucontext->mmap_xa)); 887 kfree(ucontext); 888 889 ufile->ucontext = NULL; 890 } 891 892 static int __uverbs_cleanup_ufile(struct ib_uverbs_file *ufile, 893 enum rdma_remove_reason reason) 894 { 895 struct uverbs_attr_bundle attrs = { .ufile = ufile }; 896 struct ib_ucontext *ucontext = ufile->ucontext; 897 struct ib_device *ib_dev = ucontext->device; 898 struct ib_uobject *obj, *next_obj; 899 int ret = -EINVAL; 900 901 if (ib_dev->ops.ufile_hw_cleanup) 902 ib_dev->ops.ufile_hw_cleanup(ufile); 903 904 /* 905 * This shouldn't run while executing other commands on this 906 * context. Thus, the only thing we should take care of is 907 * releasing a FD while traversing this list. The FD could be 908 * closed and released from the _release fop of this FD. 909 * In order to mitigate this, we add a lock. 910 * We take and release the lock per traversal in order to let 911 * other threads (which might still use the FDs) chance to run. 912 */ 913 list_for_each_entry_safe(obj, next_obj, &ufile->uobjects, list) { 914 attrs.context = obj->context; 915 /* 916 * if we hit this WARN_ON, that means we are 917 * racing with a lookup_get. 918 */ 919 WARN_ON(uverbs_try_lock_object(obj, UVERBS_LOOKUP_WRITE)); 920 if (reason == RDMA_REMOVE_DRIVER_FAILURE) 921 obj->object = NULL; 922 if (!uverbs_destroy_uobject(obj, reason, &attrs)) 923 ret = 0; 924 else 925 atomic_set(&obj->usecnt, 0); 926 } 927 928 if (reason == RDMA_REMOVE_DRIVER_FAILURE) { 929 WARN_ON(!list_empty(&ufile->uobjects)); 930 return 0; 931 } 932 return ret; 933 } 934 935 /* 936 * Destroy the ucontext and every uobject associated with it. 937 * 938 * This is internally locked and can be called in parallel from multiple 939 * contexts. 940 */ 941 void uverbs_destroy_ufile_hw(struct ib_uverbs_file *ufile, 942 enum rdma_remove_reason reason) 943 { 944 down_write(&ufile->hw_destroy_rwsem); 945 946 /* 947 * If a ucontext was never created then we can't have any uobjects to 948 * cleanup, nothing to do. 949 */ 950 if (!ufile->ucontext) 951 goto done; 952 953 while (!list_empty(&ufile->uobjects) && 954 !__uverbs_cleanup_ufile(ufile, reason)) { 955 } 956 957 if (WARN_ON(!list_empty(&ufile->uobjects))) 958 __uverbs_cleanup_ufile(ufile, RDMA_REMOVE_DRIVER_FAILURE); 959 ufile_destroy_ucontext(ufile, reason); 960 961 done: 962 up_write(&ufile->hw_destroy_rwsem); 963 } 964 965 const struct uverbs_obj_type_class uverbs_fd_class = { 966 .alloc_begin = alloc_begin_fd_uobject, 967 .lookup_get = lookup_get_fd_uobject, 968 .alloc_commit = alloc_commit_fd_uobject, 969 .alloc_abort = alloc_abort_fd_uobject, 970 .lookup_put = lookup_put_fd_uobject, 971 .destroy_hw = destroy_hw_fd_uobject, 972 .remove_handle = remove_handle_fd_uobject, 973 }; 974 EXPORT_SYMBOL(uverbs_fd_class); 975 976 struct ib_uobject * 977 uverbs_get_uobject_from_file(u16 object_id, enum uverbs_obj_access access, 978 s64 id, struct uverbs_attr_bundle *attrs) 979 { 980 const struct uverbs_api_object *obj = 981 uapi_get_object(attrs->ufile->device->uapi, object_id); 982 983 switch (access) { 984 case UVERBS_ACCESS_READ: 985 return rdma_lookup_get_uobject(obj, attrs->ufile, id, 986 UVERBS_LOOKUP_READ, attrs); 987 case UVERBS_ACCESS_DESTROY: 988 /* Actual destruction is done inside uverbs_handle_method */ 989 return rdma_lookup_get_uobject(obj, attrs->ufile, id, 990 UVERBS_LOOKUP_DESTROY, attrs); 991 case UVERBS_ACCESS_WRITE: 992 return rdma_lookup_get_uobject(obj, attrs->ufile, id, 993 UVERBS_LOOKUP_WRITE, attrs); 994 case UVERBS_ACCESS_NEW: 995 return rdma_alloc_begin_uobject(obj, attrs); 996 default: 997 WARN_ON(true); 998 return ERR_PTR(-EOPNOTSUPP); 999 } 1000 } 1001 1002 void uverbs_finalize_object(struct ib_uobject *uobj, 1003 enum uverbs_obj_access access, bool hw_obj_valid, 1004 bool commit, struct uverbs_attr_bundle *attrs) 1005 { 1006 /* 1007 * refcounts should be handled at the object level and not at the 1008 * uobject level. Refcounts of the objects themselves are done in 1009 * handlers. 1010 */ 1011 1012 switch (access) { 1013 case UVERBS_ACCESS_READ: 1014 rdma_lookup_put_uobject(uobj, UVERBS_LOOKUP_READ); 1015 break; 1016 case UVERBS_ACCESS_WRITE: 1017 rdma_lookup_put_uobject(uobj, UVERBS_LOOKUP_WRITE); 1018 break; 1019 case UVERBS_ACCESS_DESTROY: 1020 if (uobj) 1021 rdma_lookup_put_uobject(uobj, UVERBS_LOOKUP_DESTROY); 1022 break; 1023 case UVERBS_ACCESS_NEW: 1024 if (commit) 1025 rdma_alloc_commit_uobject(uobj, attrs); 1026 else 1027 rdma_alloc_abort_uobject(uobj, attrs, hw_obj_valid); 1028 break; 1029 default: 1030 WARN_ON(true); 1031 } 1032 } 1033 1034 /** 1035 * rdma_uattrs_has_raw_cap() - Returns whether a rdma device linked to the 1036 * uverbs attributes file has CAP_NET_RAW 1037 * capability or not. 1038 * 1039 * @attrs: Pointer to uverbs attributes 1040 * 1041 * Returns true if a rdma device's owning user namespace has CAP_NET_RAW 1042 * capability, otherwise false. 1043 */ 1044 bool rdma_uattrs_has_raw_cap(const struct uverbs_attr_bundle *attrs) 1045 { 1046 struct ib_uverbs_file *ufile = attrs->ufile; 1047 struct ib_ucontext *ucontext; 1048 bool has_cap = false; 1049 int srcu_key; 1050 1051 srcu_key = srcu_read_lock(&ufile->device->disassociate_srcu); 1052 ucontext = ib_uverbs_get_ucontext_file(ufile); 1053 if (IS_ERR(ucontext)) 1054 goto out; 1055 has_cap = rdma_dev_has_raw_cap(ucontext->device); 1056 1057 out: 1058 srcu_read_unlock(&ufile->device->disassociate_srcu, srcu_key); 1059 return has_cap; 1060 } 1061 EXPORT_SYMBOL(rdma_uattrs_has_raw_cap); 1062