1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 3 * 4 * Copyright (c) 2005 David Xu <davidxu@freebsd.org> 5 * Copyright (c) 2016-2017 Robert N. M. Watson 6 * All rights reserved. 7 * 8 * Portions of this software were developed by BAE Systems, the University of 9 * Cambridge Computer Laboratory, and Memorial University under DARPA/AFRL 10 * contract FA8650-15-C-7558 ("CADETS"), as part of the DARPA Transparent 11 * Computing (TC) research program. 12 * 13 * Redistribution and use in source and binary forms, with or without 14 * modification, are permitted provided that the following conditions 15 * are met: 16 * 1. Redistributions of source code must retain the above copyright 17 * notice, this list of conditions and the following disclaimer. 18 * 2. Redistributions in binary form must reproduce the above copyright 19 * notice, this list of conditions and the following disclaimer in the 20 * documentation and/or other materials provided with the distribution. 21 * 22 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 25 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 26 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 32 * SUCH DAMAGE. 33 * 34 */ 35 36 /* 37 * POSIX message queue implementation. 38 * 39 * 1) A mqueue filesystem can be mounted, each message queue appears 40 * in mounted directory, user can change queue's permission and 41 * ownership, or remove a queue. Manually creating a file in the 42 * directory causes a message queue to be created in the kernel with 43 * default message queue attributes applied and same name used, this 44 * method is not advocated since mq_open syscall allows user to specify 45 * different attributes. Also the file system can be mounted multiple 46 * times at different mount points but shows same contents. 47 * 48 * 2) Standard POSIX message queue API. The syscalls do not use vfs layer, 49 * but directly operate on internal data structure, this allows user to 50 * use the IPC facility without having to mount mqueue file system. 51 */ 52 53 #include "opt_capsicum.h" 54 55 #include <sys/param.h> 56 #include <sys/kernel.h> 57 #include <sys/systm.h> 58 #include <sys/limits.h> 59 #include <sys/malloc.h> 60 #include <sys/buf.h> 61 #include <sys/capsicum.h> 62 #include <sys/dirent.h> 63 #include <sys/event.h> 64 #include <sys/eventhandler.h> 65 #include <sys/fcntl.h> 66 #include <sys/file.h> 67 #include <sys/filedesc.h> 68 #include <sys/jail.h> 69 #include <sys/lock.h> 70 #include <sys/module.h> 71 #include <sys/mount.h> 72 #include <sys/mqueue.h> 73 #include <sys/mutex.h> 74 #include <sys/namei.h> 75 #include <sys/posix4.h> 76 #include <sys/poll.h> 77 #include <sys/priv.h> 78 #include <sys/proc.h> 79 #include <sys/queue.h> 80 #include <sys/sysproto.h> 81 #include <sys/stat.h> 82 #include <sys/syscall.h> 83 #include <sys/syscallsubr.h> 84 #include <sys/sysent.h> 85 #include <sys/sx.h> 86 #include <sys/sysctl.h> 87 #include <sys/taskqueue.h> 88 #include <sys/unistd.h> 89 #include <sys/user.h> 90 #include <sys/vnode.h> 91 #include <machine/atomic.h> 92 93 #include <security/audit/audit.h> 94 95 FEATURE(p1003_1b_mqueue, "POSIX P1003.1B message queues support"); 96 97 /* 98 * Limits and constants 99 */ 100 #define MQFS_NAMELEN NAME_MAX 101 #define MQFS_DELEN (8 + MQFS_NAMELEN) 102 103 /* node types */ 104 typedef enum { 105 mqfstype_none = 0, 106 mqfstype_root, 107 mqfstype_dir, 108 mqfstype_this, 109 mqfstype_parent, 110 mqfstype_file, 111 mqfstype_symlink, 112 } mqfs_type_t; 113 114 struct mqfs_node; 115 116 /* 117 * mqfs_info: describes a mqfs instance 118 */ 119 struct mqfs_info { 120 struct sx mi_lock; 121 struct mqfs_node *mi_root; 122 struct unrhdr *mi_unrhdr; 123 }; 124 125 struct mqfs_vdata { 126 LIST_ENTRY(mqfs_vdata) mv_link; 127 struct mqfs_node *mv_node; 128 struct vnode *mv_vnode; 129 struct task mv_task; 130 }; 131 132 /* 133 * mqfs_node: describes a node (file or directory) within a mqfs 134 */ 135 struct mqfs_node { 136 char mn_name[MQFS_NAMELEN+1]; 137 struct mqfs_info *mn_info; 138 struct mqfs_node *mn_parent; 139 LIST_HEAD(,mqfs_node) mn_children; 140 LIST_ENTRY(mqfs_node) mn_sibling; 141 LIST_HEAD(,mqfs_vdata) mn_vnodes; 142 const void *mn_pr_root; 143 int mn_refcount; 144 mqfs_type_t mn_type; 145 int mn_deleted; 146 uint32_t mn_fileno; 147 void *mn_data; 148 struct timespec mn_birth; 149 struct timespec mn_ctime; 150 struct timespec mn_atime; 151 struct timespec mn_mtime; 152 uid_t mn_uid; 153 gid_t mn_gid; 154 int mn_mode; 155 }; 156 157 #define VTON(vp) (((struct mqfs_vdata *)((vp)->v_data))->mv_node) 158 #define VTOMQ(vp) ((struct mqueue *)(VTON(vp)->mn_data)) 159 #define VFSTOMQFS(m) ((struct mqfs_info *)((m)->mnt_data)) 160 #define FPTOMQ(fp) ((struct mqueue *)(((struct mqfs_node *) \ 161 (fp)->f_data)->mn_data)) 162 163 TAILQ_HEAD(msgq, mqueue_msg); 164 165 struct mqueue; 166 167 struct mqueue_notifier { 168 LIST_ENTRY(mqueue_notifier) nt_link; 169 struct sigevent nt_sigev; 170 ksiginfo_t nt_ksi; 171 struct proc *nt_proc; 172 }; 173 174 struct mqueue { 175 struct mtx mq_mutex; 176 int mq_flags; 177 long mq_maxmsg; 178 long mq_msgsize; 179 long mq_curmsgs; 180 long mq_totalbytes; 181 struct msgq mq_msgq; 182 int mq_receivers; 183 int mq_senders; 184 struct selinfo mq_rsel; 185 struct selinfo mq_wsel; 186 struct mqueue_notifier *mq_notifier; 187 }; 188 189 #define MQ_RSEL 0x01 190 #define MQ_WSEL 0x02 191 192 struct mqueue_msg { 193 TAILQ_ENTRY(mqueue_msg) msg_link; 194 unsigned int msg_prio; 195 unsigned int msg_size; 196 /* following real data... */ 197 }; 198 199 static SYSCTL_NODE(_kern, OID_AUTO, mqueue, CTLFLAG_RW | CTLFLAG_MPSAFE, 0, 200 "POSIX real time message queue"); 201 202 static int default_maxmsg = 10; 203 SYSCTL_INT(_kern_mqueue, OID_AUTO, default_maxmsg, CTLFLAG_RD, 204 &default_maxmsg, 0, "Default maximum messages in queue"); 205 static int default_msgsize = 1024; 206 SYSCTL_INT(_kern_mqueue, OID_AUTO, default_msgsize, CTLFLAG_RD, 207 &default_msgsize, 0, "Default maximum message size"); 208 209 static int maxmsg = 100; 210 SYSCTL_INT(_kern_mqueue, OID_AUTO, maxmsg, CTLFLAG_RW, 211 &maxmsg, 0, "maximum messages in queue"); 212 static int maxmsgsize = 16384; 213 SYSCTL_INT(_kern_mqueue, OID_AUTO, maxmsgsize, CTLFLAG_RW, 214 &maxmsgsize, 0, "maximum message size"); 215 static int maxmq = 100; 216 SYSCTL_INT(_kern_mqueue, OID_AUTO, maxmq, CTLFLAG_RW, 217 &maxmq, 0, "maximum message queues"); 218 static int curmq = 0; 219 SYSCTL_INT(_kern_mqueue, OID_AUTO, curmq, CTLFLAG_RW, 220 &curmq, 0, "current message queue number"); 221 static int unloadable = 0; 222 static MALLOC_DEFINE(M_MQUEUEDATA, "mqdata", "mqueue data"); 223 224 static eventhandler_tag exit_tag; 225 226 /* Only one instance per-system */ 227 static struct mqfs_info mqfs_data; 228 static uma_zone_t mqnode_zone; 229 static uma_zone_t mqueue_zone; 230 static uma_zone_t mvdata_zone; 231 static uma_zone_t mqnoti_zone; 232 static struct vop_vector mqfs_vnodeops; 233 static struct fileops mqueueops; 234 static unsigned mqfs_osd_jail_slot; 235 236 /* 237 * Directory structure construction and manipulation 238 */ 239 #ifdef notyet 240 static struct mqfs_node *mqfs_create_dir(struct mqfs_node *parent, 241 const char *name, int namelen, struct ucred *cred, int mode); 242 static struct mqfs_node *mqfs_create_link(struct mqfs_node *parent, 243 const char *name, int namelen, struct ucred *cred, int mode); 244 #endif 245 246 static struct mqfs_node *mqfs_create_file(struct mqfs_node *parent, 247 const char *name, int namelen, struct ucred *cred, int mode); 248 static int mqfs_destroy(struct mqfs_node *mn); 249 static void mqfs_fileno_alloc(struct mqfs_info *mi, struct mqfs_node *mn); 250 static void mqfs_fileno_free(struct mqfs_info *mi, struct mqfs_node *mn); 251 static int mqfs_allocv(struct mount *mp, struct vnode **vpp, struct mqfs_node *pn); 252 static int mqfs_prison_remove(void *obj, void *data); 253 254 /* 255 * Message queue construction and maniplation 256 */ 257 static struct mqueue *mqueue_alloc(const struct mq_attr *attr); 258 static void mqueue_free(struct mqueue *mq); 259 static int mqueue_send(struct mqueue *mq, const char *msg_ptr, 260 size_t msg_len, unsigned msg_prio, int waitok, 261 const struct timespec *abs_timeout); 262 static int mqueue_receive(struct mqueue *mq, char *msg_ptr, 263 size_t msg_len, unsigned *msg_prio, int waitok, 264 const struct timespec *abs_timeout); 265 static int _mqueue_send(struct mqueue *mq, struct mqueue_msg *msg, 266 int timo); 267 static int _mqueue_recv(struct mqueue *mq, struct mqueue_msg **msg, 268 int timo); 269 static void mqueue_send_notification(struct mqueue *mq); 270 static void mqueue_fdclose(struct thread *td, int fd, struct file *fp); 271 static void mq_proc_exit(void *arg, struct proc *p); 272 273 /* 274 * kqueue filters 275 */ 276 static void filt_mqdetach(struct knote *kn); 277 static int filt_mqread(struct knote *kn, long hint); 278 static int filt_mqwrite(struct knote *kn, long hint); 279 280 struct filterops mq_rfiltops = { 281 .f_isfd = 1, 282 .f_detach = filt_mqdetach, 283 .f_event = filt_mqread, 284 }; 285 struct filterops mq_wfiltops = { 286 .f_isfd = 1, 287 .f_detach = filt_mqdetach, 288 .f_event = filt_mqwrite, 289 }; 290 291 /* 292 * Initialize fileno bitmap 293 */ 294 static void 295 mqfs_fileno_init(struct mqfs_info *mi) 296 { 297 struct unrhdr *up; 298 299 up = new_unrhdr(1, INT_MAX, NULL); 300 mi->mi_unrhdr = up; 301 } 302 303 /* 304 * Tear down fileno bitmap 305 */ 306 static void 307 mqfs_fileno_uninit(struct mqfs_info *mi) 308 { 309 struct unrhdr *up; 310 311 up = mi->mi_unrhdr; 312 mi->mi_unrhdr = NULL; 313 delete_unrhdr(up); 314 } 315 316 /* 317 * Allocate a file number 318 */ 319 static void 320 mqfs_fileno_alloc(struct mqfs_info *mi, struct mqfs_node *mn) 321 { 322 /* make sure our parent has a file number */ 323 if (mn->mn_parent && !mn->mn_parent->mn_fileno) 324 mqfs_fileno_alloc(mi, mn->mn_parent); 325 326 switch (mn->mn_type) { 327 case mqfstype_root: 328 case mqfstype_dir: 329 case mqfstype_file: 330 case mqfstype_symlink: 331 mn->mn_fileno = alloc_unr(mi->mi_unrhdr); 332 break; 333 case mqfstype_this: 334 KASSERT(mn->mn_parent != NULL, 335 ("mqfstype_this node has no parent")); 336 mn->mn_fileno = mn->mn_parent->mn_fileno; 337 break; 338 case mqfstype_parent: 339 KASSERT(mn->mn_parent != NULL, 340 ("mqfstype_parent node has no parent")); 341 if (mn->mn_parent == mi->mi_root) { 342 mn->mn_fileno = mn->mn_parent->mn_fileno; 343 break; 344 } 345 KASSERT(mn->mn_parent->mn_parent != NULL, 346 ("mqfstype_parent node has no grandparent")); 347 mn->mn_fileno = mn->mn_parent->mn_parent->mn_fileno; 348 break; 349 default: 350 KASSERT(0, 351 ("mqfs_fileno_alloc() called for unknown type node: %d", 352 mn->mn_type)); 353 break; 354 } 355 } 356 357 /* 358 * Release a file number 359 */ 360 static void 361 mqfs_fileno_free(struct mqfs_info *mi, struct mqfs_node *mn) 362 { 363 switch (mn->mn_type) { 364 case mqfstype_root: 365 case mqfstype_dir: 366 case mqfstype_file: 367 case mqfstype_symlink: 368 free_unr(mi->mi_unrhdr, mn->mn_fileno); 369 break; 370 case mqfstype_this: 371 case mqfstype_parent: 372 /* ignore these, as they don't "own" their file number */ 373 break; 374 default: 375 KASSERT(0, 376 ("mqfs_fileno_free() called for unknown type node: %d", 377 mn->mn_type)); 378 break; 379 } 380 } 381 382 static __inline struct mqfs_node * 383 mqnode_alloc(void) 384 { 385 return (uma_zalloc(mqnode_zone, M_WAITOK | M_ZERO)); 386 } 387 388 static __inline void 389 mqnode_free(struct mqfs_node *node) 390 { 391 uma_zfree(mqnode_zone, node); 392 } 393 394 static __inline void 395 mqnode_addref(struct mqfs_node *node) 396 { 397 atomic_add_int(&node->mn_refcount, 1); 398 } 399 400 static __inline void 401 mqnode_release(struct mqfs_node *node) 402 { 403 struct mqfs_info *mqfs; 404 int old, exp; 405 406 mqfs = node->mn_info; 407 old = atomic_fetchadd_int(&node->mn_refcount, -1); 408 if (node->mn_type == mqfstype_dir || 409 node->mn_type == mqfstype_root) 410 exp = 3; /* include . and .. */ 411 else 412 exp = 1; 413 if (old == exp) { 414 int locked = sx_xlocked(&mqfs->mi_lock); 415 if (!locked) 416 sx_xlock(&mqfs->mi_lock); 417 mqfs_destroy(node); 418 if (!locked) 419 sx_xunlock(&mqfs->mi_lock); 420 } 421 } 422 423 /* 424 * Add a node to a directory 425 */ 426 static int 427 mqfs_add_node(struct mqfs_node *parent, struct mqfs_node *node) 428 { 429 KASSERT(parent != NULL, ("%s(): parent is NULL", __func__)); 430 KASSERT(parent->mn_info != NULL, 431 ("%s(): parent has no mn_info", __func__)); 432 KASSERT(parent->mn_type == mqfstype_dir || 433 parent->mn_type == mqfstype_root, 434 ("%s(): parent is not a directory", __func__)); 435 436 node->mn_info = parent->mn_info; 437 node->mn_parent = parent; 438 LIST_INIT(&node->mn_children); 439 LIST_INIT(&node->mn_vnodes); 440 LIST_INSERT_HEAD(&parent->mn_children, node, mn_sibling); 441 mqnode_addref(parent); 442 return (0); 443 } 444 445 static struct mqfs_node * 446 mqfs_create_node(const char *name, int namelen, struct ucred *cred, int mode, 447 int nodetype) 448 { 449 struct mqfs_node *node; 450 451 node = mqnode_alloc(); 452 strncpy(node->mn_name, name, namelen); 453 node->mn_pr_root = cred->cr_prison->pr_root; 454 node->mn_type = nodetype; 455 node->mn_refcount = 1; 456 vfs_timestamp(&node->mn_birth); 457 node->mn_ctime = node->mn_atime = node->mn_mtime = 458 node->mn_birth; 459 node->mn_uid = cred->cr_uid; 460 node->mn_gid = cred->cr_gid; 461 node->mn_mode = mode; 462 return (node); 463 } 464 465 /* 466 * Create a file 467 */ 468 static struct mqfs_node * 469 mqfs_create_file(struct mqfs_node *parent, const char *name, int namelen, 470 struct ucred *cred, int mode) 471 { 472 struct mqfs_node *node; 473 474 node = mqfs_create_node(name, namelen, cred, mode, mqfstype_file); 475 if (mqfs_add_node(parent, node) != 0) { 476 mqnode_free(node); 477 return (NULL); 478 } 479 return (node); 480 } 481 482 /* 483 * Add . and .. to a directory 484 */ 485 static int 486 mqfs_fixup_dir(struct mqfs_node *parent) 487 { 488 struct mqfs_node *dir; 489 490 dir = mqnode_alloc(); 491 dir->mn_name[0] = '.'; 492 dir->mn_type = mqfstype_this; 493 dir->mn_refcount = 1; 494 if (mqfs_add_node(parent, dir) != 0) { 495 mqnode_free(dir); 496 return (-1); 497 } 498 499 dir = mqnode_alloc(); 500 dir->mn_name[0] = dir->mn_name[1] = '.'; 501 dir->mn_type = mqfstype_parent; 502 dir->mn_refcount = 1; 503 504 if (mqfs_add_node(parent, dir) != 0) { 505 mqnode_free(dir); 506 return (-1); 507 } 508 509 return (0); 510 } 511 512 #ifdef notyet 513 514 /* 515 * Create a directory 516 */ 517 static struct mqfs_node * 518 mqfs_create_dir(struct mqfs_node *parent, const char *name, int namelen, 519 struct ucred *cred, int mode) 520 { 521 struct mqfs_node *node; 522 523 node = mqfs_create_node(name, namelen, cred, mode, mqfstype_dir); 524 if (mqfs_add_node(parent, node) != 0) { 525 mqnode_free(node); 526 return (NULL); 527 } 528 529 if (mqfs_fixup_dir(node) != 0) { 530 mqfs_destroy(node); 531 return (NULL); 532 } 533 return (node); 534 } 535 536 /* 537 * Create a symlink 538 */ 539 static struct mqfs_node * 540 mqfs_create_link(struct mqfs_node *parent, const char *name, int namelen, 541 struct ucred *cred, int mode) 542 { 543 struct mqfs_node *node; 544 545 node = mqfs_create_node(name, namelen, cred, mode, mqfstype_symlink); 546 if (mqfs_add_node(parent, node) != 0) { 547 mqnode_free(node); 548 return (NULL); 549 } 550 return (node); 551 } 552 553 #endif 554 555 /* 556 * Destroy a node or a tree of nodes 557 */ 558 static int 559 mqfs_destroy(struct mqfs_node *node) 560 { 561 struct mqfs_node *parent; 562 563 KASSERT(node != NULL, 564 ("%s(): node is NULL", __func__)); 565 KASSERT(node->mn_info != NULL, 566 ("%s(): node has no mn_info", __func__)); 567 568 /* destroy children */ 569 if (node->mn_type == mqfstype_dir || node->mn_type == mqfstype_root) 570 while (! LIST_EMPTY(&node->mn_children)) 571 mqfs_destroy(LIST_FIRST(&node->mn_children)); 572 573 /* unlink from parent */ 574 if ((parent = node->mn_parent) != NULL) { 575 KASSERT(parent->mn_info == node->mn_info, 576 ("%s(): parent has different mn_info", __func__)); 577 LIST_REMOVE(node, mn_sibling); 578 } 579 580 if (node->mn_fileno != 0) 581 mqfs_fileno_free(node->mn_info, node); 582 if (node->mn_data != NULL) 583 mqueue_free(node->mn_data); 584 mqnode_free(node); 585 return (0); 586 } 587 588 /* 589 * Mount a mqfs instance 590 */ 591 static int 592 mqfs_mount(struct mount *mp) 593 { 594 struct statfs *sbp; 595 596 if (mp->mnt_flag & MNT_UPDATE) 597 return (EOPNOTSUPP); 598 599 mp->mnt_data = &mqfs_data; 600 MNT_ILOCK(mp); 601 mp->mnt_flag |= MNT_LOCAL; 602 MNT_IUNLOCK(mp); 603 vfs_getnewfsid(mp); 604 605 sbp = &mp->mnt_stat; 606 vfs_mountedfrom(mp, "mqueue"); 607 sbp->f_bsize = PAGE_SIZE; 608 sbp->f_iosize = PAGE_SIZE; 609 sbp->f_blocks = 1; 610 sbp->f_bfree = 1; 611 sbp->f_bavail = 0; 612 sbp->f_files = 0; 613 sbp->f_ffree = 0; 614 return (0); 615 } 616 617 /* 618 * Unmount a mqfs instance 619 */ 620 static int 621 mqfs_unmount(struct mount *mp, int mntflags) 622 { 623 int error; 624 625 error = vflush(mp, 0, (mntflags & MNT_FORCE) ? FORCECLOSE : 0, 626 curthread); 627 return (error); 628 } 629 630 /* 631 * Return a root vnode 632 */ 633 static int 634 mqfs_root(struct mount *mp, int flags, struct vnode **vpp) 635 { 636 struct mqfs_info *mqfs; 637 int ret; 638 639 mqfs = VFSTOMQFS(mp); 640 ret = mqfs_allocv(mp, vpp, mqfs->mi_root); 641 return (ret); 642 } 643 644 /* 645 * Return filesystem stats 646 */ 647 static int 648 mqfs_statfs(struct mount *mp, struct statfs *sbp) 649 { 650 /* XXX update statistics */ 651 return (0); 652 } 653 654 /* 655 * Initialize a mqfs instance 656 */ 657 static int 658 mqfs_init(struct vfsconf *vfc) 659 { 660 struct mqfs_node *root; 661 struct mqfs_info *mi; 662 osd_method_t methods[PR_MAXMETHOD] = { 663 [PR_METHOD_REMOVE] = mqfs_prison_remove, 664 }; 665 666 mqnode_zone = uma_zcreate("mqnode", sizeof(struct mqfs_node), 667 NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, 0); 668 mqueue_zone = uma_zcreate("mqueue", sizeof(struct mqueue), 669 NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, 0); 670 mvdata_zone = uma_zcreate("mvdata", 671 sizeof(struct mqfs_vdata), NULL, NULL, NULL, 672 NULL, UMA_ALIGN_PTR, 0); 673 mqnoti_zone = uma_zcreate("mqnotifier", sizeof(struct mqueue_notifier), 674 NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, 0); 675 mi = &mqfs_data; 676 sx_init(&mi->mi_lock, "mqfs lock"); 677 /* set up the root diretory */ 678 root = mqfs_create_node("/", 1, curthread->td_ucred, 01777, 679 mqfstype_root); 680 root->mn_info = mi; 681 LIST_INIT(&root->mn_children); 682 LIST_INIT(&root->mn_vnodes); 683 mi->mi_root = root; 684 mqfs_fileno_init(mi); 685 mqfs_fileno_alloc(mi, root); 686 mqfs_fixup_dir(root); 687 exit_tag = EVENTHANDLER_REGISTER(process_exit, mq_proc_exit, NULL, 688 EVENTHANDLER_PRI_ANY); 689 mq_fdclose = mqueue_fdclose; 690 p31b_setcfg(CTL_P1003_1B_MESSAGE_PASSING, _POSIX_MESSAGE_PASSING); 691 mqfs_osd_jail_slot = osd_jail_register(NULL, methods); 692 return (0); 693 } 694 695 /* 696 * Destroy a mqfs instance 697 */ 698 static int 699 mqfs_uninit(struct vfsconf *vfc) 700 { 701 struct mqfs_info *mi; 702 703 if (!unloadable) 704 return (EOPNOTSUPP); 705 osd_jail_deregister(mqfs_osd_jail_slot); 706 EVENTHANDLER_DEREGISTER(process_exit, exit_tag); 707 mi = &mqfs_data; 708 mqfs_destroy(mi->mi_root); 709 mi->mi_root = NULL; 710 mqfs_fileno_uninit(mi); 711 sx_destroy(&mi->mi_lock); 712 uma_zdestroy(mqnode_zone); 713 uma_zdestroy(mqueue_zone); 714 uma_zdestroy(mvdata_zone); 715 uma_zdestroy(mqnoti_zone); 716 return (0); 717 } 718 719 /* 720 * task routine 721 */ 722 static void 723 do_recycle(void *context, int pending __unused) 724 { 725 struct vnode *vp = (struct vnode *)context; 726 727 vn_lock(vp, LK_EXCLUSIVE | LK_RETRY); 728 vrecycle(vp); 729 VOP_UNLOCK(vp); 730 vdrop(vp); 731 } 732 733 /* 734 * Allocate a vnode 735 */ 736 static int 737 mqfs_allocv(struct mount *mp, struct vnode **vpp, struct mqfs_node *pn) 738 { 739 struct mqfs_vdata *vd; 740 struct mqfs_info *mqfs; 741 struct vnode *newvpp; 742 int error; 743 744 mqfs = pn->mn_info; 745 *vpp = NULL; 746 sx_xlock(&mqfs->mi_lock); 747 LIST_FOREACH(vd, &pn->mn_vnodes, mv_link) { 748 if (vd->mv_vnode->v_mount == mp) { 749 vhold(vd->mv_vnode); 750 break; 751 } 752 } 753 754 if (vd != NULL) { 755 found: 756 *vpp = vd->mv_vnode; 757 sx_xunlock(&mqfs->mi_lock); 758 error = vget(*vpp, LK_RETRY | LK_EXCLUSIVE); 759 vdrop(*vpp); 760 return (error); 761 } 762 sx_xunlock(&mqfs->mi_lock); 763 764 error = getnewvnode("mqueue", mp, &mqfs_vnodeops, &newvpp); 765 if (error) 766 return (error); 767 vn_lock(newvpp, LK_EXCLUSIVE | LK_RETRY); 768 error = insmntque(newvpp, mp); 769 if (error != 0) 770 return (error); 771 772 sx_xlock(&mqfs->mi_lock); 773 /* 774 * Check if it has already been allocated 775 * while we were blocked. 776 */ 777 LIST_FOREACH(vd, &pn->mn_vnodes, mv_link) { 778 if (vd->mv_vnode->v_mount == mp) { 779 vhold(vd->mv_vnode); 780 sx_xunlock(&mqfs->mi_lock); 781 782 vgone(newvpp); 783 vput(newvpp); 784 goto found; 785 } 786 } 787 788 *vpp = newvpp; 789 790 vd = uma_zalloc(mvdata_zone, M_WAITOK); 791 (*vpp)->v_data = vd; 792 vd->mv_vnode = *vpp; 793 vd->mv_node = pn; 794 TASK_INIT(&vd->mv_task, 0, do_recycle, *vpp); 795 LIST_INSERT_HEAD(&pn->mn_vnodes, vd, mv_link); 796 mqnode_addref(pn); 797 switch (pn->mn_type) { 798 case mqfstype_root: 799 (*vpp)->v_vflag = VV_ROOT; 800 /* fall through */ 801 case mqfstype_dir: 802 case mqfstype_this: 803 case mqfstype_parent: 804 (*vpp)->v_type = VDIR; 805 break; 806 case mqfstype_file: 807 (*vpp)->v_type = VREG; 808 break; 809 case mqfstype_symlink: 810 (*vpp)->v_type = VLNK; 811 break; 812 case mqfstype_none: 813 KASSERT(0, ("mqfs_allocf called for null node\n")); 814 default: 815 panic("%s has unexpected type: %d", pn->mn_name, pn->mn_type); 816 } 817 sx_xunlock(&mqfs->mi_lock); 818 vn_set_state(*vpp, VSTATE_CONSTRUCTED); 819 return (0); 820 } 821 822 /* 823 * Search a directory entry 824 */ 825 static struct mqfs_node * 826 mqfs_search(struct mqfs_node *pd, const char *name, int len, struct ucred *cred) 827 { 828 struct mqfs_node *pn; 829 const void *pr_root; 830 831 sx_assert(&pd->mn_info->mi_lock, SX_LOCKED); 832 pr_root = cred->cr_prison->pr_root; 833 LIST_FOREACH(pn, &pd->mn_children, mn_sibling) { 834 /* Only match names within the same prison root directory */ 835 if ((pn->mn_pr_root == NULL || pn->mn_pr_root == pr_root) && 836 strncmp(pn->mn_name, name, len) == 0 && 837 pn->mn_name[len] == '\0') 838 return (pn); 839 } 840 return (NULL); 841 } 842 843 /* 844 * Look up a file or directory. 845 */ 846 static int 847 mqfs_lookupx(struct vop_cachedlookup_args *ap) 848 { 849 struct componentname *cnp; 850 struct vnode *dvp, **vpp; 851 struct mqfs_node *pd; 852 struct mqfs_node *pn; 853 struct mqfs_info *mqfs; 854 int nameiop, flags, error, namelen; 855 char *pname; 856 struct thread *td; 857 858 td = curthread; 859 cnp = ap->a_cnp; 860 vpp = ap->a_vpp; 861 dvp = ap->a_dvp; 862 pname = cnp->cn_nameptr; 863 namelen = cnp->cn_namelen; 864 flags = cnp->cn_flags; 865 nameiop = cnp->cn_nameiop; 866 pd = VTON(dvp); 867 pn = NULL; 868 mqfs = pd->mn_info; 869 *vpp = NULLVP; 870 871 if (dvp->v_type != VDIR) 872 return (ENOTDIR); 873 874 error = VOP_ACCESS(dvp, VEXEC, cnp->cn_cred, td); 875 if (error) 876 return (error); 877 878 /* shortcut: check if the name is too long */ 879 if (cnp->cn_namelen >= MQFS_NAMELEN) 880 return (ENOENT); 881 882 /* self */ 883 if (namelen == 1 && pname[0] == '.') { 884 if ((flags & ISLASTCN) && nameiop != LOOKUP) 885 return (EINVAL); 886 pn = pd; 887 *vpp = dvp; 888 VREF(dvp); 889 return (0); 890 } 891 892 /* parent */ 893 if (cnp->cn_flags & ISDOTDOT) { 894 if (dvp->v_vflag & VV_ROOT) 895 return (EIO); 896 if ((flags & ISLASTCN) && nameiop != LOOKUP) 897 return (EINVAL); 898 VOP_UNLOCK(dvp); 899 KASSERT(pd->mn_parent, ("non-root directory has no parent")); 900 pn = pd->mn_parent; 901 error = mqfs_allocv(dvp->v_mount, vpp, pn); 902 vn_lock(dvp, LK_EXCLUSIVE | LK_RETRY); 903 return (error); 904 } 905 906 /* named node */ 907 sx_xlock(&mqfs->mi_lock); 908 pn = mqfs_search(pd, pname, namelen, cnp->cn_cred); 909 if (pn != NULL) 910 mqnode_addref(pn); 911 sx_xunlock(&mqfs->mi_lock); 912 913 /* found */ 914 if (pn != NULL) { 915 /* DELETE */ 916 if (nameiop == DELETE && (flags & ISLASTCN)) { 917 error = VOP_ACCESS(dvp, VWRITE, cnp->cn_cred, td); 918 if (error) { 919 mqnode_release(pn); 920 return (error); 921 } 922 if (*vpp == dvp) { 923 VREF(dvp); 924 *vpp = dvp; 925 mqnode_release(pn); 926 return (0); 927 } 928 } 929 930 /* allocate vnode */ 931 error = mqfs_allocv(dvp->v_mount, vpp, pn); 932 mqnode_release(pn); 933 if (error == 0 && cnp->cn_flags & MAKEENTRY) 934 cache_enter(dvp, *vpp, cnp); 935 return (error); 936 } 937 938 /* not found */ 939 940 /* will create a new entry in the directory ? */ 941 if ((nameiop == CREATE || nameiop == RENAME) && (flags & LOCKPARENT) 942 && (flags & ISLASTCN)) { 943 error = VOP_ACCESS(dvp, VWRITE, cnp->cn_cred, td); 944 if (error) 945 return (error); 946 return (EJUSTRETURN); 947 } 948 return (ENOENT); 949 } 950 951 #if 0 952 struct vop_lookup_args { 953 struct vop_generic_args a_gen; 954 struct vnode *a_dvp; 955 struct vnode **a_vpp; 956 struct componentname *a_cnp; 957 }; 958 #endif 959 960 /* 961 * vnode lookup operation 962 */ 963 static int 964 mqfs_lookup(struct vop_cachedlookup_args *ap) 965 { 966 int rc; 967 968 rc = mqfs_lookupx(ap); 969 return (rc); 970 } 971 972 #if 0 973 struct vop_create_args { 974 struct vnode *a_dvp; 975 struct vnode **a_vpp; 976 struct componentname *a_cnp; 977 struct vattr *a_vap; 978 }; 979 #endif 980 981 /* 982 * vnode creation operation 983 */ 984 static int 985 mqfs_create(struct vop_create_args *ap) 986 { 987 struct mqfs_info *mqfs = VFSTOMQFS(ap->a_dvp->v_mount); 988 struct componentname *cnp = ap->a_cnp; 989 struct mqfs_node *pd; 990 struct mqfs_node *pn; 991 struct mqueue *mq; 992 int error; 993 994 pd = VTON(ap->a_dvp); 995 if (pd->mn_type != mqfstype_root && pd->mn_type != mqfstype_dir) 996 return (ENOTDIR); 997 mq = mqueue_alloc(NULL); 998 if (mq == NULL) 999 return (EAGAIN); 1000 sx_xlock(&mqfs->mi_lock); 1001 pn = mqfs_create_file(pd, cnp->cn_nameptr, cnp->cn_namelen, 1002 cnp->cn_cred, ap->a_vap->va_mode); 1003 if (pn == NULL) { 1004 sx_xunlock(&mqfs->mi_lock); 1005 error = ENOSPC; 1006 } else { 1007 mqnode_addref(pn); 1008 sx_xunlock(&mqfs->mi_lock); 1009 error = mqfs_allocv(ap->a_dvp->v_mount, ap->a_vpp, pn); 1010 mqnode_release(pn); 1011 if (error) 1012 mqfs_destroy(pn); 1013 else 1014 pn->mn_data = mq; 1015 } 1016 if (error) 1017 mqueue_free(mq); 1018 return (error); 1019 } 1020 1021 /* 1022 * Remove an entry 1023 */ 1024 static int 1025 do_unlink(struct mqfs_node *pn, struct ucred *ucred) 1026 { 1027 struct mqfs_node *parent; 1028 struct mqfs_vdata *vd; 1029 int error = 0; 1030 1031 sx_assert(&pn->mn_info->mi_lock, SX_LOCKED); 1032 1033 if (ucred->cr_uid != pn->mn_uid && 1034 (error = priv_check_cred(ucred, PRIV_MQ_ADMIN)) != 0) 1035 error = EACCES; 1036 else if (!pn->mn_deleted) { 1037 parent = pn->mn_parent; 1038 pn->mn_parent = NULL; 1039 pn->mn_deleted = 1; 1040 LIST_REMOVE(pn, mn_sibling); 1041 LIST_FOREACH(vd, &pn->mn_vnodes, mv_link) { 1042 cache_purge(vd->mv_vnode); 1043 vhold(vd->mv_vnode); 1044 taskqueue_enqueue(taskqueue_thread, &vd->mv_task); 1045 } 1046 mqnode_release(pn); 1047 mqnode_release(parent); 1048 } else 1049 error = ENOENT; 1050 return (error); 1051 } 1052 1053 #if 0 1054 struct vop_remove_args { 1055 struct vnode *a_dvp; 1056 struct vnode *a_vp; 1057 struct componentname *a_cnp; 1058 }; 1059 #endif 1060 1061 /* 1062 * vnode removal operation 1063 */ 1064 static int 1065 mqfs_remove(struct vop_remove_args *ap) 1066 { 1067 struct mqfs_info *mqfs = VFSTOMQFS(ap->a_dvp->v_mount); 1068 struct mqfs_node *pn; 1069 int error; 1070 1071 if (ap->a_vp->v_type == VDIR) 1072 return (EPERM); 1073 pn = VTON(ap->a_vp); 1074 sx_xlock(&mqfs->mi_lock); 1075 error = do_unlink(pn, ap->a_cnp->cn_cred); 1076 sx_xunlock(&mqfs->mi_lock); 1077 return (error); 1078 } 1079 1080 #if 0 1081 struct vop_inactive_args { 1082 struct vnode *a_vp; 1083 struct thread *a_td; 1084 }; 1085 #endif 1086 1087 static int 1088 mqfs_inactive(struct vop_inactive_args *ap) 1089 { 1090 struct mqfs_node *pn = VTON(ap->a_vp); 1091 1092 if (pn->mn_deleted) 1093 vrecycle(ap->a_vp); 1094 return (0); 1095 } 1096 1097 #if 0 1098 struct vop_reclaim_args { 1099 struct vop_generic_args a_gen; 1100 struct vnode *a_vp; 1101 }; 1102 #endif 1103 1104 static int 1105 mqfs_reclaim(struct vop_reclaim_args *ap) 1106 { 1107 struct mqfs_info *mqfs = VFSTOMQFS(ap->a_vp->v_mount); 1108 struct vnode *vp = ap->a_vp; 1109 struct mqfs_node *pn; 1110 struct mqfs_vdata *vd; 1111 1112 vd = vp->v_data; 1113 pn = vd->mv_node; 1114 sx_xlock(&mqfs->mi_lock); 1115 vp->v_data = NULL; 1116 LIST_REMOVE(vd, mv_link); 1117 mqnode_release(pn); 1118 sx_xunlock(&mqfs->mi_lock); 1119 uma_zfree(mvdata_zone, vd); 1120 return (0); 1121 } 1122 1123 #if 0 1124 struct vop_open_args { 1125 struct vop_generic_args a_gen; 1126 struct vnode *a_vp; 1127 int a_mode; 1128 struct ucred *a_cred; 1129 struct thread *a_td; 1130 struct file *a_fp; 1131 }; 1132 #endif 1133 1134 static int 1135 mqfs_open(struct vop_open_args *ap) 1136 { 1137 return (0); 1138 } 1139 1140 #if 0 1141 struct vop_close_args { 1142 struct vop_generic_args a_gen; 1143 struct vnode *a_vp; 1144 int a_fflag; 1145 struct ucred *a_cred; 1146 struct thread *a_td; 1147 }; 1148 #endif 1149 1150 static int 1151 mqfs_close(struct vop_close_args *ap) 1152 { 1153 return (0); 1154 } 1155 1156 #if 0 1157 struct vop_access_args { 1158 struct vop_generic_args a_gen; 1159 struct vnode *a_vp; 1160 accmode_t a_accmode; 1161 struct ucred *a_cred; 1162 struct thread *a_td; 1163 }; 1164 #endif 1165 1166 /* 1167 * Verify permissions 1168 */ 1169 static int 1170 mqfs_access(struct vop_access_args *ap) 1171 { 1172 struct vnode *vp = ap->a_vp; 1173 struct vattr vattr; 1174 int error; 1175 1176 error = VOP_GETATTR(vp, &vattr, ap->a_cred); 1177 if (error) 1178 return (error); 1179 error = vaccess(vp->v_type, vattr.va_mode, vattr.va_uid, vattr.va_gid, 1180 ap->a_accmode, ap->a_cred); 1181 return (error); 1182 } 1183 1184 #if 0 1185 struct vop_getattr_args { 1186 struct vop_generic_args a_gen; 1187 struct vnode *a_vp; 1188 struct vattr *a_vap; 1189 struct ucred *a_cred; 1190 }; 1191 #endif 1192 1193 /* 1194 * Get file attributes 1195 */ 1196 static int 1197 mqfs_getattr(struct vop_getattr_args *ap) 1198 { 1199 struct vnode *vp = ap->a_vp; 1200 struct mqfs_node *pn = VTON(vp); 1201 struct vattr *vap = ap->a_vap; 1202 int error = 0; 1203 1204 vap->va_type = vp->v_type; 1205 vap->va_mode = pn->mn_mode; 1206 vap->va_nlink = 1; 1207 vap->va_uid = pn->mn_uid; 1208 vap->va_gid = pn->mn_gid; 1209 vap->va_fsid = vp->v_mount->mnt_stat.f_fsid.val[0]; 1210 vap->va_fileid = pn->mn_fileno; 1211 vap->va_size = 0; 1212 vap->va_blocksize = PAGE_SIZE; 1213 vap->va_bytes = vap->va_size = 0; 1214 vap->va_atime = pn->mn_atime; 1215 vap->va_mtime = pn->mn_mtime; 1216 vap->va_ctime = pn->mn_ctime; 1217 vap->va_birthtime = pn->mn_birth; 1218 vap->va_gen = 0; 1219 vap->va_flags = 0; 1220 vap->va_rdev = NODEV; 1221 vap->va_bytes = 0; 1222 vap->va_filerev = 0; 1223 return (error); 1224 } 1225 1226 #if 0 1227 struct vop_setattr_args { 1228 struct vop_generic_args a_gen; 1229 struct vnode *a_vp; 1230 struct vattr *a_vap; 1231 struct ucred *a_cred; 1232 }; 1233 #endif 1234 /* 1235 * Set attributes 1236 */ 1237 static int 1238 mqfs_setattr(struct vop_setattr_args *ap) 1239 { 1240 struct mqfs_node *pn; 1241 struct vattr *vap; 1242 struct vnode *vp; 1243 struct thread *td; 1244 int c, error; 1245 uid_t uid; 1246 gid_t gid; 1247 1248 td = curthread; 1249 vap = ap->a_vap; 1250 vp = ap->a_vp; 1251 if (vap->va_type != VNON || 1252 vap->va_nlink != VNOVAL || 1253 vap->va_fsid != VNOVAL || 1254 vap->va_fileid != VNOVAL || 1255 vap->va_blocksize != VNOVAL || 1256 (vap->va_flags != VNOVAL && vap->va_flags != 0) || 1257 vap->va_rdev != VNOVAL || 1258 (int)vap->va_bytes != VNOVAL || 1259 vap->va_gen != VNOVAL) { 1260 return (EINVAL); 1261 } 1262 1263 pn = VTON(vp); 1264 1265 error = c = 0; 1266 if (vap->va_uid == (uid_t)VNOVAL) 1267 uid = pn->mn_uid; 1268 else 1269 uid = vap->va_uid; 1270 if (vap->va_gid == (gid_t)VNOVAL) 1271 gid = pn->mn_gid; 1272 else 1273 gid = vap->va_gid; 1274 1275 if (uid != pn->mn_uid || gid != pn->mn_gid) { 1276 /* 1277 * To modify the ownership of a file, must possess VADMIN 1278 * for that file. 1279 */ 1280 if ((error = VOP_ACCESS(vp, VADMIN, ap->a_cred, td))) 1281 return (error); 1282 1283 /* 1284 * XXXRW: Why is there a privilege check here: shouldn't the 1285 * check in VOP_ACCESS() be enough? Also, are the group bits 1286 * below definitely right? 1287 */ 1288 if ((ap->a_cred->cr_uid != pn->mn_uid || uid != pn->mn_uid || 1289 (gid != pn->mn_gid && !groupmember(gid, ap->a_cred))) && 1290 (error = priv_check(td, PRIV_MQ_ADMIN)) != 0) 1291 return (error); 1292 pn->mn_uid = uid; 1293 pn->mn_gid = gid; 1294 c = 1; 1295 } 1296 1297 if (vap->va_mode != (mode_t)VNOVAL) { 1298 if (ap->a_cred->cr_uid != pn->mn_uid && 1299 (error = priv_check(td, PRIV_MQ_ADMIN))) 1300 return (error); 1301 pn->mn_mode = vap->va_mode; 1302 c = 1; 1303 } 1304 1305 if (vap->va_atime.tv_sec != VNOVAL || vap->va_mtime.tv_sec != VNOVAL) { 1306 /* See the comment in ufs_vnops::ufs_setattr(). */ 1307 if ((error = VOP_ACCESS(vp, VADMIN, ap->a_cred, td)) && 1308 ((vap->va_vaflags & VA_UTIMES_NULL) == 0 || 1309 (error = VOP_ACCESS(vp, VWRITE, ap->a_cred, td)))) 1310 return (error); 1311 if (vap->va_atime.tv_sec != VNOVAL) { 1312 pn->mn_atime = vap->va_atime; 1313 } 1314 if (vap->va_mtime.tv_sec != VNOVAL) { 1315 pn->mn_mtime = vap->va_mtime; 1316 } 1317 c = 1; 1318 } 1319 if (c) { 1320 vfs_timestamp(&pn->mn_ctime); 1321 } 1322 return (0); 1323 } 1324 1325 #if 0 1326 struct vop_read_args { 1327 struct vop_generic_args a_gen; 1328 struct vnode *a_vp; 1329 struct uio *a_uio; 1330 int a_ioflag; 1331 struct ucred *a_cred; 1332 }; 1333 #endif 1334 1335 /* 1336 * Read from a file 1337 */ 1338 static int 1339 mqfs_read(struct vop_read_args *ap) 1340 { 1341 char buf[80]; 1342 struct vnode *vp = ap->a_vp; 1343 struct uio *uio = ap->a_uio; 1344 struct mqueue *mq; 1345 int len, error; 1346 1347 if (vp->v_type != VREG) 1348 return (EINVAL); 1349 1350 mq = VTOMQ(vp); 1351 snprintf(buf, sizeof(buf), 1352 "QSIZE:%-10ld MAXMSG:%-10ld CURMSG:%-10ld MSGSIZE:%-10ld\n", 1353 mq->mq_totalbytes, 1354 mq->mq_maxmsg, 1355 mq->mq_curmsgs, 1356 mq->mq_msgsize); 1357 buf[sizeof(buf)-1] = '\0'; 1358 len = strlen(buf); 1359 error = uiomove_frombuf(buf, len, uio); 1360 return (error); 1361 } 1362 1363 #if 0 1364 struct vop_readdir_args { 1365 struct vop_generic_args a_gen; 1366 struct vnode *a_vp; 1367 struct uio *a_uio; 1368 struct ucred *a_cred; 1369 int *a_eofflag; 1370 int *a_ncookies; 1371 uint64_t **a_cookies; 1372 }; 1373 #endif 1374 1375 /* 1376 * Return directory entries. 1377 */ 1378 static int 1379 mqfs_readdir(struct vop_readdir_args *ap) 1380 { 1381 struct vnode *vp; 1382 struct mqfs_info *mi; 1383 struct mqfs_node *pd; 1384 struct mqfs_node *pn; 1385 struct dirent entry; 1386 struct uio *uio; 1387 const void *pr_root; 1388 int *tmp_ncookies = NULL; 1389 off_t offset; 1390 int error, i; 1391 1392 vp = ap->a_vp; 1393 mi = VFSTOMQFS(vp->v_mount); 1394 pd = VTON(vp); 1395 uio = ap->a_uio; 1396 1397 if (vp->v_type != VDIR) 1398 return (ENOTDIR); 1399 1400 if (uio->uio_offset < 0) 1401 return (EINVAL); 1402 1403 if (ap->a_ncookies != NULL) { 1404 tmp_ncookies = ap->a_ncookies; 1405 *ap->a_ncookies = 0; 1406 ap->a_ncookies = NULL; 1407 } 1408 1409 error = 0; 1410 offset = 0; 1411 1412 pr_root = ap->a_cred->cr_prison->pr_root; 1413 sx_xlock(&mi->mi_lock); 1414 1415 LIST_FOREACH(pn, &pd->mn_children, mn_sibling) { 1416 entry.d_reclen = sizeof(entry); 1417 1418 /* 1419 * Only show names within the same prison root directory 1420 * (or not associated with a prison, e.g. "." and ".."). 1421 */ 1422 if (pn->mn_pr_root != NULL && pn->mn_pr_root != pr_root) 1423 continue; 1424 if (!pn->mn_fileno) 1425 mqfs_fileno_alloc(mi, pn); 1426 entry.d_fileno = pn->mn_fileno; 1427 entry.d_off = offset + entry.d_reclen; 1428 for (i = 0; i < MQFS_NAMELEN - 1 && pn->mn_name[i] != '\0'; ++i) 1429 entry.d_name[i] = pn->mn_name[i]; 1430 entry.d_namlen = i; 1431 switch (pn->mn_type) { 1432 case mqfstype_root: 1433 case mqfstype_dir: 1434 case mqfstype_this: 1435 case mqfstype_parent: 1436 entry.d_type = DT_DIR; 1437 break; 1438 case mqfstype_file: 1439 entry.d_type = DT_REG; 1440 break; 1441 case mqfstype_symlink: 1442 entry.d_type = DT_LNK; 1443 break; 1444 default: 1445 panic("%s has unexpected node type: %d", pn->mn_name, 1446 pn->mn_type); 1447 } 1448 dirent_terminate(&entry); 1449 if (entry.d_reclen > uio->uio_resid) 1450 break; 1451 if (offset >= uio->uio_offset) { 1452 error = vfs_read_dirent(ap, &entry, offset); 1453 if (error) 1454 break; 1455 } 1456 offset += entry.d_reclen; 1457 } 1458 sx_xunlock(&mi->mi_lock); 1459 1460 uio->uio_offset = offset; 1461 1462 if (tmp_ncookies != NULL) 1463 ap->a_ncookies = tmp_ncookies; 1464 1465 return (error); 1466 } 1467 1468 #ifdef notyet 1469 1470 #if 0 1471 struct vop_mkdir_args { 1472 struct vnode *a_dvp; 1473 struvt vnode **a_vpp; 1474 struvt componentname *a_cnp; 1475 struct vattr *a_vap; 1476 }; 1477 #endif 1478 1479 /* 1480 * Create a directory. 1481 */ 1482 static int 1483 mqfs_mkdir(struct vop_mkdir_args *ap) 1484 { 1485 struct mqfs_info *mqfs = VFSTOMQFS(ap->a_dvp->v_mount); 1486 struct componentname *cnp = ap->a_cnp; 1487 struct mqfs_node *pd = VTON(ap->a_dvp); 1488 struct mqfs_node *pn; 1489 int error; 1490 1491 if (pd->mn_type != mqfstype_root && pd->mn_type != mqfstype_dir) 1492 return (ENOTDIR); 1493 sx_xlock(&mqfs->mi_lock); 1494 pn = mqfs_create_dir(pd, cnp->cn_nameptr, cnp->cn_namelen, 1495 ap->a_vap->cn_cred, ap->a_vap->va_mode); 1496 if (pn != NULL) 1497 mqnode_addref(pn); 1498 sx_xunlock(&mqfs->mi_lock); 1499 if (pn == NULL) { 1500 error = ENOSPC; 1501 } else { 1502 error = mqfs_allocv(ap->a_dvp->v_mount, ap->a_vpp, pn); 1503 mqnode_release(pn); 1504 } 1505 return (error); 1506 } 1507 1508 #if 0 1509 struct vop_rmdir_args { 1510 struct vnode *a_dvp; 1511 struct vnode *a_vp; 1512 struct componentname *a_cnp; 1513 }; 1514 #endif 1515 1516 /* 1517 * Remove a directory. 1518 */ 1519 static int 1520 mqfs_rmdir(struct vop_rmdir_args *ap) 1521 { 1522 struct mqfs_info *mqfs = VFSTOMQFS(ap->a_dvp->v_mount); 1523 struct mqfs_node *pn = VTON(ap->a_vp); 1524 struct mqfs_node *pt; 1525 1526 if (pn->mn_type != mqfstype_dir) 1527 return (ENOTDIR); 1528 1529 sx_xlock(&mqfs->mi_lock); 1530 if (pn->mn_deleted) { 1531 sx_xunlock(&mqfs->mi_lock); 1532 return (ENOENT); 1533 } 1534 1535 pt = LIST_FIRST(&pn->mn_children); 1536 pt = LIST_NEXT(pt, mn_sibling); 1537 pt = LIST_NEXT(pt, mn_sibling); 1538 if (pt != NULL) { 1539 sx_xunlock(&mqfs->mi_lock); 1540 return (ENOTEMPTY); 1541 } 1542 pt = pn->mn_parent; 1543 pn->mn_parent = NULL; 1544 pn->mn_deleted = 1; 1545 LIST_REMOVE(pn, mn_sibling); 1546 mqnode_release(pn); 1547 mqnode_release(pt); 1548 sx_xunlock(&mqfs->mi_lock); 1549 cache_purge(ap->a_vp); 1550 return (0); 1551 } 1552 1553 #endif /* notyet */ 1554 1555 /* 1556 * See if this prison root is obsolete, and clean up associated queues if it is. 1557 */ 1558 static int 1559 mqfs_prison_remove(void *obj, void *data __unused) 1560 { 1561 const struct prison *pr = obj; 1562 struct prison *tpr; 1563 struct mqfs_node *pn, *tpn; 1564 struct vnode *pr_root; 1565 1566 pr_root = pr->pr_root; 1567 if (pr->pr_parent->pr_root == pr_root) 1568 return (0); 1569 TAILQ_FOREACH(tpr, &allprison, pr_list) { 1570 if (tpr != pr && tpr->pr_root == pr_root) 1571 return (0); 1572 } 1573 /* 1574 * No jails are rooted in this directory anymore, 1575 * so no queues should be either. 1576 */ 1577 sx_xlock(&mqfs_data.mi_lock); 1578 LIST_FOREACH_SAFE(pn, &mqfs_data.mi_root->mn_children, 1579 mn_sibling, tpn) { 1580 if (pn->mn_pr_root == pr_root) 1581 (void)do_unlink(pn, curthread->td_ucred); 1582 } 1583 sx_xunlock(&mqfs_data.mi_lock); 1584 return (0); 1585 } 1586 1587 /* 1588 * Allocate a message queue 1589 */ 1590 static struct mqueue * 1591 mqueue_alloc(const struct mq_attr *attr) 1592 { 1593 struct mqueue *mq; 1594 1595 if (curmq >= maxmq) 1596 return (NULL); 1597 mq = uma_zalloc(mqueue_zone, M_WAITOK | M_ZERO); 1598 TAILQ_INIT(&mq->mq_msgq); 1599 if (attr != NULL) { 1600 mq->mq_maxmsg = attr->mq_maxmsg; 1601 mq->mq_msgsize = attr->mq_msgsize; 1602 } else { 1603 mq->mq_maxmsg = default_maxmsg; 1604 mq->mq_msgsize = default_msgsize; 1605 } 1606 mtx_init(&mq->mq_mutex, "mqueue lock", NULL, MTX_DEF); 1607 knlist_init_mtx(&mq->mq_rsel.si_note, &mq->mq_mutex); 1608 knlist_init_mtx(&mq->mq_wsel.si_note, &mq->mq_mutex); 1609 atomic_add_int(&curmq, 1); 1610 return (mq); 1611 } 1612 1613 /* 1614 * Destroy a message queue 1615 */ 1616 static void 1617 mqueue_free(struct mqueue *mq) 1618 { 1619 struct mqueue_msg *msg; 1620 1621 while ((msg = TAILQ_FIRST(&mq->mq_msgq)) != NULL) { 1622 TAILQ_REMOVE(&mq->mq_msgq, msg, msg_link); 1623 free(msg, M_MQUEUEDATA); 1624 } 1625 1626 mtx_destroy(&mq->mq_mutex); 1627 seldrain(&mq->mq_rsel); 1628 seldrain(&mq->mq_wsel); 1629 knlist_destroy(&mq->mq_rsel.si_note); 1630 knlist_destroy(&mq->mq_wsel.si_note); 1631 uma_zfree(mqueue_zone, mq); 1632 atomic_add_int(&curmq, -1); 1633 } 1634 1635 /* 1636 * Load a message from user space 1637 */ 1638 static struct mqueue_msg * 1639 mqueue_loadmsg(const char *msg_ptr, size_t msg_size, int msg_prio) 1640 { 1641 struct mqueue_msg *msg; 1642 size_t len; 1643 int error; 1644 1645 len = sizeof(struct mqueue_msg) + msg_size; 1646 msg = malloc(len, M_MQUEUEDATA, M_WAITOK); 1647 error = copyin(msg_ptr, ((char *)msg) + sizeof(struct mqueue_msg), 1648 msg_size); 1649 if (error) { 1650 free(msg, M_MQUEUEDATA); 1651 msg = NULL; 1652 } else { 1653 msg->msg_size = msg_size; 1654 msg->msg_prio = msg_prio; 1655 } 1656 return (msg); 1657 } 1658 1659 /* 1660 * Save a message to user space 1661 */ 1662 static int 1663 mqueue_savemsg(struct mqueue_msg *msg, char *msg_ptr, int *msg_prio) 1664 { 1665 int error; 1666 1667 error = copyout(((char *)msg) + sizeof(*msg), msg_ptr, 1668 msg->msg_size); 1669 if (error == 0 && msg_prio != NULL) 1670 error = copyout(&msg->msg_prio, msg_prio, sizeof(int)); 1671 return (error); 1672 } 1673 1674 /* 1675 * Free a message's memory 1676 */ 1677 static __inline void 1678 mqueue_freemsg(struct mqueue_msg *msg) 1679 { 1680 free(msg, M_MQUEUEDATA); 1681 } 1682 1683 /* 1684 * Send a message. if waitok is false, thread will not be 1685 * blocked if there is no data in queue, otherwise, absolute 1686 * time will be checked. 1687 */ 1688 int 1689 mqueue_send(struct mqueue *mq, const char *msg_ptr, 1690 size_t msg_len, unsigned msg_prio, int waitok, 1691 const struct timespec *abs_timeout) 1692 { 1693 struct mqueue_msg *msg; 1694 struct timespec ts, ts2; 1695 struct timeval tv; 1696 int error; 1697 1698 if (msg_prio >= MQ_PRIO_MAX) 1699 return (EINVAL); 1700 if (msg_len > mq->mq_msgsize) 1701 return (EMSGSIZE); 1702 msg = mqueue_loadmsg(msg_ptr, msg_len, msg_prio); 1703 if (msg == NULL) 1704 return (EFAULT); 1705 1706 /* O_NONBLOCK case */ 1707 if (!waitok) { 1708 error = _mqueue_send(mq, msg, -1); 1709 if (error) 1710 goto bad; 1711 return (0); 1712 } 1713 1714 /* we allow a null timeout (wait forever) */ 1715 if (abs_timeout == NULL) { 1716 error = _mqueue_send(mq, msg, 0); 1717 if (error) 1718 goto bad; 1719 return (0); 1720 } 1721 1722 /* send it before checking time */ 1723 error = _mqueue_send(mq, msg, -1); 1724 if (error == 0) 1725 return (0); 1726 1727 if (error != EAGAIN) 1728 goto bad; 1729 1730 if (abs_timeout->tv_nsec >= 1000000000 || abs_timeout->tv_nsec < 0) { 1731 error = EINVAL; 1732 goto bad; 1733 } 1734 for (;;) { 1735 getnanotime(&ts); 1736 timespecsub(abs_timeout, &ts, &ts2); 1737 if (ts2.tv_sec < 0 || (ts2.tv_sec == 0 && ts2.tv_nsec <= 0)) { 1738 error = ETIMEDOUT; 1739 break; 1740 } 1741 TIMESPEC_TO_TIMEVAL(&tv, &ts2); 1742 error = _mqueue_send(mq, msg, tvtohz(&tv)); 1743 if (error != ETIMEDOUT) 1744 break; 1745 } 1746 if (error == 0) 1747 return (0); 1748 bad: 1749 mqueue_freemsg(msg); 1750 return (error); 1751 } 1752 1753 /* 1754 * Common routine to send a message 1755 */ 1756 static int 1757 _mqueue_send(struct mqueue *mq, struct mqueue_msg *msg, int timo) 1758 { 1759 struct mqueue_msg *msg2; 1760 int error = 0; 1761 1762 mtx_lock(&mq->mq_mutex); 1763 while (mq->mq_curmsgs >= mq->mq_maxmsg && error == 0) { 1764 if (timo < 0) { 1765 mtx_unlock(&mq->mq_mutex); 1766 return (EAGAIN); 1767 } 1768 mq->mq_senders++; 1769 error = msleep(&mq->mq_senders, &mq->mq_mutex, 1770 PCATCH, "mqsend", timo); 1771 mq->mq_senders--; 1772 if (error == EAGAIN) 1773 error = ETIMEDOUT; 1774 } 1775 if (mq->mq_curmsgs >= mq->mq_maxmsg) { 1776 mtx_unlock(&mq->mq_mutex); 1777 return (error); 1778 } 1779 error = 0; 1780 if (TAILQ_EMPTY(&mq->mq_msgq)) { 1781 TAILQ_INSERT_HEAD(&mq->mq_msgq, msg, msg_link); 1782 } else { 1783 if (msg->msg_prio <= TAILQ_LAST(&mq->mq_msgq, msgq)->msg_prio) { 1784 TAILQ_INSERT_TAIL(&mq->mq_msgq, msg, msg_link); 1785 } else { 1786 TAILQ_FOREACH(msg2, &mq->mq_msgq, msg_link) { 1787 if (msg2->msg_prio < msg->msg_prio) 1788 break; 1789 } 1790 TAILQ_INSERT_BEFORE(msg2, msg, msg_link); 1791 } 1792 } 1793 mq->mq_curmsgs++; 1794 mq->mq_totalbytes += msg->msg_size; 1795 if (mq->mq_receivers) 1796 wakeup_one(&mq->mq_receivers); 1797 else if (mq->mq_notifier != NULL) 1798 mqueue_send_notification(mq); 1799 if (mq->mq_flags & MQ_RSEL) { 1800 mq->mq_flags &= ~MQ_RSEL; 1801 selwakeup(&mq->mq_rsel); 1802 } 1803 KNOTE_LOCKED(&mq->mq_rsel.si_note, 0); 1804 mtx_unlock(&mq->mq_mutex); 1805 return (0); 1806 } 1807 1808 /* 1809 * Send realtime a signal to process which registered itself 1810 * successfully by mq_notify. 1811 */ 1812 static void 1813 mqueue_send_notification(struct mqueue *mq) 1814 { 1815 struct mqueue_notifier *nt; 1816 struct thread *td; 1817 struct proc *p; 1818 int error; 1819 1820 mtx_assert(&mq->mq_mutex, MA_OWNED); 1821 nt = mq->mq_notifier; 1822 if (nt->nt_sigev.sigev_notify != SIGEV_NONE) { 1823 p = nt->nt_proc; 1824 error = sigev_findtd(p, &nt->nt_sigev, &td); 1825 if (error) { 1826 mq->mq_notifier = NULL; 1827 return; 1828 } 1829 if (!KSI_ONQ(&nt->nt_ksi)) { 1830 ksiginfo_set_sigev(&nt->nt_ksi, &nt->nt_sigev); 1831 tdsendsignal(p, td, nt->nt_ksi.ksi_signo, &nt->nt_ksi); 1832 } 1833 PROC_UNLOCK(p); 1834 } 1835 mq->mq_notifier = NULL; 1836 } 1837 1838 /* 1839 * Get a message. if waitok is false, thread will not be 1840 * blocked if there is no data in queue, otherwise, absolute 1841 * time will be checked. 1842 */ 1843 int 1844 mqueue_receive(struct mqueue *mq, char *msg_ptr, 1845 size_t msg_len, unsigned *msg_prio, int waitok, 1846 const struct timespec *abs_timeout) 1847 { 1848 struct mqueue_msg *msg; 1849 struct timespec ts, ts2; 1850 struct timeval tv; 1851 int error; 1852 1853 if (msg_len < mq->mq_msgsize) 1854 return (EMSGSIZE); 1855 1856 /* O_NONBLOCK case */ 1857 if (!waitok) { 1858 error = _mqueue_recv(mq, &msg, -1); 1859 if (error) 1860 return (error); 1861 goto received; 1862 } 1863 1864 /* we allow a null timeout (wait forever). */ 1865 if (abs_timeout == NULL) { 1866 error = _mqueue_recv(mq, &msg, 0); 1867 if (error) 1868 return (error); 1869 goto received; 1870 } 1871 1872 /* try to get a message before checking time */ 1873 error = _mqueue_recv(mq, &msg, -1); 1874 if (error == 0) 1875 goto received; 1876 1877 if (error != EAGAIN) 1878 return (error); 1879 1880 if (abs_timeout->tv_nsec >= 1000000000 || abs_timeout->tv_nsec < 0) { 1881 error = EINVAL; 1882 return (error); 1883 } 1884 1885 for (;;) { 1886 getnanotime(&ts); 1887 timespecsub(abs_timeout, &ts, &ts2); 1888 if (ts2.tv_sec < 0 || (ts2.tv_sec == 0 && ts2.tv_nsec <= 0)) { 1889 error = ETIMEDOUT; 1890 return (error); 1891 } 1892 TIMESPEC_TO_TIMEVAL(&tv, &ts2); 1893 error = _mqueue_recv(mq, &msg, tvtohz(&tv)); 1894 if (error == 0) 1895 break; 1896 if (error != ETIMEDOUT) 1897 return (error); 1898 } 1899 1900 received: 1901 error = mqueue_savemsg(msg, msg_ptr, msg_prio); 1902 if (error == 0) { 1903 curthread->td_retval[0] = msg->msg_size; 1904 curthread->td_retval[1] = 0; 1905 } 1906 mqueue_freemsg(msg); 1907 return (error); 1908 } 1909 1910 /* 1911 * Common routine to receive a message 1912 */ 1913 static int 1914 _mqueue_recv(struct mqueue *mq, struct mqueue_msg **msg, int timo) 1915 { 1916 int error = 0; 1917 1918 mtx_lock(&mq->mq_mutex); 1919 while ((*msg = TAILQ_FIRST(&mq->mq_msgq)) == NULL && error == 0) { 1920 if (timo < 0) { 1921 mtx_unlock(&mq->mq_mutex); 1922 return (EAGAIN); 1923 } 1924 mq->mq_receivers++; 1925 error = msleep(&mq->mq_receivers, &mq->mq_mutex, 1926 PCATCH, "mqrecv", timo); 1927 mq->mq_receivers--; 1928 if (error == EAGAIN) 1929 error = ETIMEDOUT; 1930 } 1931 if (*msg != NULL) { 1932 error = 0; 1933 TAILQ_REMOVE(&mq->mq_msgq, *msg, msg_link); 1934 mq->mq_curmsgs--; 1935 mq->mq_totalbytes -= (*msg)->msg_size; 1936 if (mq->mq_senders) 1937 wakeup_one(&mq->mq_senders); 1938 if (mq->mq_flags & MQ_WSEL) { 1939 mq->mq_flags &= ~MQ_WSEL; 1940 selwakeup(&mq->mq_wsel); 1941 } 1942 KNOTE_LOCKED(&mq->mq_wsel.si_note, 0); 1943 } 1944 if (mq->mq_notifier != NULL && mq->mq_receivers == 0 && 1945 !TAILQ_EMPTY(&mq->mq_msgq)) { 1946 mqueue_send_notification(mq); 1947 } 1948 mtx_unlock(&mq->mq_mutex); 1949 return (error); 1950 } 1951 1952 static __inline struct mqueue_notifier * 1953 notifier_alloc(void) 1954 { 1955 return (uma_zalloc(mqnoti_zone, M_WAITOK | M_ZERO)); 1956 } 1957 1958 static __inline void 1959 notifier_free(struct mqueue_notifier *p) 1960 { 1961 uma_zfree(mqnoti_zone, p); 1962 } 1963 1964 static struct mqueue_notifier * 1965 notifier_search(struct proc *p, int fd) 1966 { 1967 struct mqueue_notifier *nt; 1968 1969 LIST_FOREACH(nt, &p->p_mqnotifier, nt_link) { 1970 if (nt->nt_ksi.ksi_mqd == fd) 1971 break; 1972 } 1973 return (nt); 1974 } 1975 1976 static __inline void 1977 notifier_insert(struct proc *p, struct mqueue_notifier *nt) 1978 { 1979 LIST_INSERT_HEAD(&p->p_mqnotifier, nt, nt_link); 1980 } 1981 1982 static __inline void 1983 notifier_delete(struct proc *p, struct mqueue_notifier *nt) 1984 { 1985 LIST_REMOVE(nt, nt_link); 1986 notifier_free(nt); 1987 } 1988 1989 static void 1990 notifier_remove(struct proc *p, struct mqueue *mq, int fd) 1991 { 1992 struct mqueue_notifier *nt; 1993 1994 mtx_assert(&mq->mq_mutex, MA_OWNED); 1995 PROC_LOCK(p); 1996 nt = notifier_search(p, fd); 1997 if (nt != NULL) { 1998 if (mq->mq_notifier == nt) 1999 mq->mq_notifier = NULL; 2000 sigqueue_take(&nt->nt_ksi); 2001 notifier_delete(p, nt); 2002 } 2003 PROC_UNLOCK(p); 2004 } 2005 2006 int 2007 kern_kmq_open(struct thread *td, const char *upath, int flags, mode_t mode, 2008 const struct mq_attr *attr) 2009 { 2010 char *path, pathbuf[MQFS_NAMELEN + 1]; 2011 struct mqfs_node *pn; 2012 struct pwddesc *pdp; 2013 struct file *fp; 2014 struct mqueue *mq; 2015 int fd, error, len, cmode; 2016 2017 AUDIT_ARG_FFLAGS(flags); 2018 AUDIT_ARG_MODE(mode); 2019 2020 pdp = td->td_proc->p_pd; 2021 cmode = ((mode & ~pdp->pd_cmask) & ALLPERMS) & ~S_ISTXT; 2022 mq = NULL; 2023 if ((flags & O_CREAT) != 0 && attr != NULL) { 2024 if (attr->mq_maxmsg <= 0 || attr->mq_maxmsg > maxmsg) 2025 return (EINVAL); 2026 if (attr->mq_msgsize <= 0 || attr->mq_msgsize > maxmsgsize) 2027 return (EINVAL); 2028 } 2029 2030 path = pathbuf; 2031 error = copyinstr(upath, path, MQFS_NAMELEN + 1, NULL); 2032 if (error) 2033 return (error); 2034 2035 /* 2036 * The first character of name may be a slash (/) character 2037 * and the remaining characters of name cannot include any slash 2038 * characters. 2039 */ 2040 len = strlen(path); 2041 if (len < 2 || strchr(path + 1, '/') != NULL) 2042 return (EINVAL); 2043 if (path[0] == '/') { 2044 path++; 2045 len--; 2046 } 2047 /* 2048 * "." and ".." are magic directories, populated on the fly, and cannot 2049 * be opened as queues. 2050 */ 2051 if (strcmp(path, ".") == 0 || strcmp(path, "..") == 0) 2052 return (EINVAL); 2053 AUDIT_ARG_UPATH1_CANON(pathbuf); 2054 2055 error = falloc(td, &fp, &fd, O_CLOEXEC); 2056 if (error) 2057 return (error); 2058 2059 sx_xlock(&mqfs_data.mi_lock); 2060 pn = mqfs_search(mqfs_data.mi_root, path, len, td->td_ucred); 2061 if (pn == NULL) { 2062 if (!(flags & O_CREAT)) { 2063 error = ENOENT; 2064 } else { 2065 mq = mqueue_alloc(attr); 2066 if (mq == NULL) { 2067 error = ENFILE; 2068 } else { 2069 pn = mqfs_create_file(mqfs_data.mi_root, 2070 path, len, td->td_ucred, 2071 cmode); 2072 if (pn == NULL) { 2073 error = ENOSPC; 2074 mqueue_free(mq); 2075 } 2076 } 2077 } 2078 2079 if (error == 0) { 2080 pn->mn_data = mq; 2081 } 2082 } else { 2083 if ((flags & (O_CREAT | O_EXCL)) == (O_CREAT | O_EXCL)) { 2084 error = EEXIST; 2085 } else { 2086 accmode_t accmode = 0; 2087 2088 if (flags & FREAD) 2089 accmode |= VREAD; 2090 if (flags & FWRITE) 2091 accmode |= VWRITE; 2092 error = vaccess(VREG, pn->mn_mode, pn->mn_uid, 2093 pn->mn_gid, accmode, td->td_ucred); 2094 } 2095 } 2096 2097 if (error) { 2098 sx_xunlock(&mqfs_data.mi_lock); 2099 fdclose(td, fp, fd); 2100 fdrop(fp, td); 2101 return (error); 2102 } 2103 2104 mqnode_addref(pn); 2105 sx_xunlock(&mqfs_data.mi_lock); 2106 2107 finit(fp, flags & (FREAD | FWRITE | O_NONBLOCK), DTYPE_MQUEUE, pn, 2108 &mqueueops); 2109 2110 td->td_retval[0] = fd; 2111 fdrop(fp, td); 2112 return (0); 2113 } 2114 2115 /* 2116 * Syscall to open a message queue. 2117 */ 2118 int 2119 sys_kmq_open(struct thread *td, struct kmq_open_args *uap) 2120 { 2121 struct mq_attr attr; 2122 int flags, error; 2123 2124 if ((uap->flags & O_ACCMODE) == O_ACCMODE || uap->flags & O_EXEC) 2125 return (EINVAL); 2126 flags = FFLAGS(uap->flags); 2127 if ((flags & O_CREAT) != 0 && uap->attr != NULL) { 2128 error = copyin(uap->attr, &attr, sizeof(attr)); 2129 if (error) 2130 return (error); 2131 } 2132 return (kern_kmq_open(td, uap->path, flags, uap->mode, 2133 uap->attr != NULL ? &attr : NULL)); 2134 } 2135 2136 /* 2137 * Syscall to unlink a message queue. 2138 */ 2139 int 2140 sys_kmq_unlink(struct thread *td, struct kmq_unlink_args *uap) 2141 { 2142 char *path, pathbuf[MQFS_NAMELEN + 1]; 2143 struct mqfs_node *pn; 2144 int error, len; 2145 2146 path = pathbuf; 2147 error = copyinstr(uap->path, path, MQFS_NAMELEN + 1, NULL); 2148 if (error) 2149 return (error); 2150 2151 len = strlen(path); 2152 if (len < 2 || strchr(path + 1, '/') != NULL) 2153 return (EINVAL); 2154 if (path[0] == '/') { 2155 path++; 2156 len--; 2157 } 2158 if (strcmp(path, ".") == 0 || strcmp(path, "..") == 0) 2159 return (EINVAL); 2160 AUDIT_ARG_UPATH1_CANON(pathbuf); 2161 2162 sx_xlock(&mqfs_data.mi_lock); 2163 pn = mqfs_search(mqfs_data.mi_root, path, len, td->td_ucred); 2164 if (pn != NULL) 2165 error = do_unlink(pn, td->td_ucred); 2166 else 2167 error = ENOENT; 2168 sx_xunlock(&mqfs_data.mi_lock); 2169 return (error); 2170 } 2171 2172 typedef int (*_fgetf)(struct thread *, int, cap_rights_t *, struct file **); 2173 2174 /* 2175 * Get message queue by giving file slot 2176 */ 2177 static int 2178 _getmq(struct thread *td, int fd, cap_rights_t *rightsp, _fgetf func, 2179 struct file **fpp, struct mqfs_node **ppn, struct mqueue **pmq) 2180 { 2181 struct mqfs_node *pn; 2182 int error; 2183 2184 error = func(td, fd, rightsp, fpp); 2185 if (error) 2186 return (error); 2187 if (&mqueueops != (*fpp)->f_ops) { 2188 fdrop(*fpp, td); 2189 return (EBADF); 2190 } 2191 pn = (*fpp)->f_data; 2192 if (ppn) 2193 *ppn = pn; 2194 if (pmq) 2195 *pmq = pn->mn_data; 2196 return (0); 2197 } 2198 2199 static __inline int 2200 getmq(struct thread *td, int fd, struct file **fpp, struct mqfs_node **ppn, 2201 struct mqueue **pmq) 2202 { 2203 2204 return _getmq(td, fd, &cap_event_rights, fget, 2205 fpp, ppn, pmq); 2206 } 2207 2208 static __inline int 2209 getmq_read(struct thread *td, int fd, struct file **fpp, 2210 struct mqfs_node **ppn, struct mqueue **pmq) 2211 { 2212 2213 return _getmq(td, fd, &cap_read_rights, fget_read, 2214 fpp, ppn, pmq); 2215 } 2216 2217 static __inline int 2218 getmq_write(struct thread *td, int fd, struct file **fpp, 2219 struct mqfs_node **ppn, struct mqueue **pmq) 2220 { 2221 2222 return _getmq(td, fd, &cap_write_rights, fget_write, 2223 fpp, ppn, pmq); 2224 } 2225 2226 int 2227 kern_kmq_setattr(struct thread *td, int mqd, const struct mq_attr *attr, 2228 struct mq_attr *oattr) 2229 { 2230 struct mqueue *mq; 2231 struct file *fp; 2232 u_int oflag, flag; 2233 int error; 2234 2235 AUDIT_ARG_FD(mqd); 2236 if (attr != NULL && (attr->mq_flags & ~O_NONBLOCK) != 0) 2237 return (EINVAL); 2238 error = getmq(td, mqd, &fp, NULL, &mq); 2239 if (error) 2240 return (error); 2241 oattr->mq_maxmsg = mq->mq_maxmsg; 2242 oattr->mq_msgsize = mq->mq_msgsize; 2243 oattr->mq_curmsgs = mq->mq_curmsgs; 2244 if (attr != NULL) { 2245 do { 2246 oflag = flag = fp->f_flag; 2247 flag &= ~O_NONBLOCK; 2248 flag |= (attr->mq_flags & O_NONBLOCK); 2249 } while (atomic_cmpset_int(&fp->f_flag, oflag, flag) == 0); 2250 } else 2251 oflag = fp->f_flag; 2252 oattr->mq_flags = (O_NONBLOCK & oflag); 2253 fdrop(fp, td); 2254 return (error); 2255 } 2256 2257 int 2258 sys_kmq_setattr(struct thread *td, struct kmq_setattr_args *uap) 2259 { 2260 struct mq_attr attr, oattr; 2261 int error; 2262 2263 if (uap->attr != NULL) { 2264 error = copyin(uap->attr, &attr, sizeof(attr)); 2265 if (error != 0) 2266 return (error); 2267 } 2268 error = kern_kmq_setattr(td, uap->mqd, uap->attr != NULL ? &attr : NULL, 2269 &oattr); 2270 if (error == 0 && uap->oattr != NULL) { 2271 bzero(oattr.__reserved, sizeof(oattr.__reserved)); 2272 error = copyout(&oattr, uap->oattr, sizeof(oattr)); 2273 } 2274 return (error); 2275 } 2276 2277 int 2278 kern_kmq_timedreceive(struct thread *td, int mqd, char *msg_ptr, 2279 size_t msg_len, unsigned int *msg_prio, const struct timespec *abs_timeout) 2280 { 2281 struct mqueue *mq; 2282 struct file *fp; 2283 int error, waitok; 2284 2285 AUDIT_ARG_FD(mqd); 2286 error = getmq_read(td, mqd, &fp, NULL, &mq); 2287 if (error != 0) 2288 return (error); 2289 waitok = (fp->f_flag & O_NONBLOCK) == 0; 2290 error = mqueue_receive(mq, msg_ptr, msg_len, msg_prio, waitok, 2291 abs_timeout); 2292 fdrop(fp, td); 2293 return (error); 2294 } 2295 2296 int 2297 sys_kmq_timedreceive(struct thread *td, struct kmq_timedreceive_args *uap) 2298 { 2299 struct timespec *abs_timeout, ets; 2300 int error; 2301 2302 if (uap->abs_timeout != NULL) { 2303 error = copyin(uap->abs_timeout, &ets, sizeof(ets)); 2304 if (error != 0) 2305 return (error); 2306 abs_timeout = &ets; 2307 } else 2308 abs_timeout = NULL; 2309 2310 return (kern_kmq_timedreceive(td, uap->mqd, uap->msg_ptr, uap->msg_len, 2311 uap->msg_prio, abs_timeout)); 2312 } 2313 2314 int 2315 kern_kmq_timedsend(struct thread *td, int mqd, const char *msg_ptr, 2316 size_t msg_len, unsigned int msg_prio, const struct timespec *abs_timeout) 2317 { 2318 struct mqueue *mq; 2319 struct file *fp; 2320 int error, waitok; 2321 2322 AUDIT_ARG_FD(mqd); 2323 error = getmq_write(td, mqd, &fp, NULL, &mq); 2324 if (error != 0) 2325 return (error); 2326 waitok = (fp->f_flag & O_NONBLOCK) == 0; 2327 error = mqueue_send(mq, msg_ptr, msg_len, msg_prio, waitok, 2328 abs_timeout); 2329 fdrop(fp, td); 2330 return (error); 2331 } 2332 2333 int 2334 sys_kmq_timedsend(struct thread *td, struct kmq_timedsend_args *uap) 2335 { 2336 struct timespec *abs_timeout, ets; 2337 int error; 2338 2339 if (uap->abs_timeout != NULL) { 2340 error = copyin(uap->abs_timeout, &ets, sizeof(ets)); 2341 if (error != 0) 2342 return (error); 2343 abs_timeout = &ets; 2344 } else 2345 abs_timeout = NULL; 2346 2347 return (kern_kmq_timedsend(td, uap->mqd, uap->msg_ptr, uap->msg_len, 2348 uap->msg_prio, abs_timeout)); 2349 } 2350 2351 int 2352 kern_kmq_notify(struct thread *td, int mqd, struct sigevent *sigev) 2353 { 2354 struct filedesc *fdp; 2355 struct proc *p; 2356 struct mqueue *mq; 2357 struct file *fp, *fp2; 2358 struct mqueue_notifier *nt, *newnt = NULL; 2359 int error; 2360 2361 AUDIT_ARG_FD(mqd); 2362 if (sigev != NULL) { 2363 if (sigev->sigev_notify != SIGEV_SIGNAL && 2364 sigev->sigev_notify != SIGEV_THREAD_ID && 2365 sigev->sigev_notify != SIGEV_NONE) 2366 return (EINVAL); 2367 if ((sigev->sigev_notify == SIGEV_SIGNAL || 2368 sigev->sigev_notify == SIGEV_THREAD_ID) && 2369 !_SIG_VALID(sigev->sigev_signo)) 2370 return (EINVAL); 2371 } 2372 p = td->td_proc; 2373 fdp = td->td_proc->p_fd; 2374 error = getmq(td, mqd, &fp, NULL, &mq); 2375 if (error) 2376 return (error); 2377 again: 2378 FILEDESC_SLOCK(fdp); 2379 fp2 = fget_noref(fdp, mqd); 2380 if (fp2 == NULL) { 2381 FILEDESC_SUNLOCK(fdp); 2382 error = EBADF; 2383 goto out; 2384 } 2385 #ifdef CAPABILITIES 2386 error = cap_check(cap_rights(fdp, mqd), &cap_event_rights); 2387 if (error) { 2388 FILEDESC_SUNLOCK(fdp); 2389 goto out; 2390 } 2391 #endif 2392 if (fp2 != fp) { 2393 FILEDESC_SUNLOCK(fdp); 2394 error = EBADF; 2395 goto out; 2396 } 2397 mtx_lock(&mq->mq_mutex); 2398 FILEDESC_SUNLOCK(fdp); 2399 if (sigev != NULL) { 2400 if (mq->mq_notifier != NULL) { 2401 error = EBUSY; 2402 } else { 2403 PROC_LOCK(p); 2404 nt = notifier_search(p, mqd); 2405 if (nt == NULL) { 2406 if (newnt == NULL) { 2407 PROC_UNLOCK(p); 2408 mtx_unlock(&mq->mq_mutex); 2409 newnt = notifier_alloc(); 2410 goto again; 2411 } 2412 } 2413 2414 if (nt != NULL) { 2415 sigqueue_take(&nt->nt_ksi); 2416 if (newnt != NULL) { 2417 notifier_free(newnt); 2418 newnt = NULL; 2419 } 2420 } else { 2421 nt = newnt; 2422 newnt = NULL; 2423 ksiginfo_init(&nt->nt_ksi); 2424 nt->nt_ksi.ksi_flags |= KSI_INS | KSI_EXT; 2425 nt->nt_ksi.ksi_code = SI_MESGQ; 2426 nt->nt_proc = p; 2427 nt->nt_ksi.ksi_mqd = mqd; 2428 notifier_insert(p, nt); 2429 } 2430 nt->nt_sigev = *sigev; 2431 mq->mq_notifier = nt; 2432 PROC_UNLOCK(p); 2433 /* 2434 * if there is no receivers and message queue 2435 * is not empty, we should send notification 2436 * as soon as possible. 2437 */ 2438 if (mq->mq_receivers == 0 && 2439 !TAILQ_EMPTY(&mq->mq_msgq)) 2440 mqueue_send_notification(mq); 2441 } 2442 } else { 2443 notifier_remove(p, mq, mqd); 2444 } 2445 mtx_unlock(&mq->mq_mutex); 2446 2447 out: 2448 fdrop(fp, td); 2449 if (newnt != NULL) 2450 notifier_free(newnt); 2451 return (error); 2452 } 2453 2454 int 2455 sys_kmq_notify(struct thread *td, struct kmq_notify_args *uap) 2456 { 2457 struct sigevent ev, *evp; 2458 int error; 2459 2460 if (uap->sigev == NULL) { 2461 evp = NULL; 2462 } else { 2463 error = copyin(uap->sigev, &ev, sizeof(ev)); 2464 if (error != 0) 2465 return (error); 2466 evp = &ev; 2467 } 2468 return (kern_kmq_notify(td, uap->mqd, evp)); 2469 } 2470 2471 static void 2472 mqueue_fdclose(struct thread *td, int fd, struct file *fp) 2473 { 2474 struct mqueue *mq; 2475 #ifdef INVARIANTS 2476 struct filedesc *fdp; 2477 2478 fdp = td->td_proc->p_fd; 2479 FILEDESC_LOCK_ASSERT(fdp); 2480 #endif 2481 2482 if (fp->f_ops == &mqueueops) { 2483 mq = FPTOMQ(fp); 2484 mtx_lock(&mq->mq_mutex); 2485 notifier_remove(td->td_proc, mq, fd); 2486 2487 /* have to wakeup thread in same process */ 2488 if (mq->mq_flags & MQ_RSEL) { 2489 mq->mq_flags &= ~MQ_RSEL; 2490 selwakeup(&mq->mq_rsel); 2491 } 2492 if (mq->mq_flags & MQ_WSEL) { 2493 mq->mq_flags &= ~MQ_WSEL; 2494 selwakeup(&mq->mq_wsel); 2495 } 2496 mtx_unlock(&mq->mq_mutex); 2497 } 2498 } 2499 2500 static void 2501 mq_proc_exit(void *arg __unused, struct proc *p) 2502 { 2503 struct filedesc *fdp; 2504 struct file *fp; 2505 struct mqueue *mq; 2506 int i; 2507 2508 fdp = p->p_fd; 2509 FILEDESC_SLOCK(fdp); 2510 for (i = 0; i < fdp->fd_nfiles; ++i) { 2511 fp = fget_noref(fdp, i); 2512 if (fp != NULL && fp->f_ops == &mqueueops) { 2513 mq = FPTOMQ(fp); 2514 mtx_lock(&mq->mq_mutex); 2515 notifier_remove(p, FPTOMQ(fp), i); 2516 mtx_unlock(&mq->mq_mutex); 2517 } 2518 } 2519 FILEDESC_SUNLOCK(fdp); 2520 KASSERT(LIST_EMPTY(&p->p_mqnotifier), ("mq notifiers left")); 2521 } 2522 2523 static int 2524 mqf_poll(struct file *fp, int events, struct ucred *active_cred, 2525 struct thread *td) 2526 { 2527 struct mqueue *mq = FPTOMQ(fp); 2528 int revents = 0; 2529 2530 mtx_lock(&mq->mq_mutex); 2531 if (events & (POLLIN | POLLRDNORM)) { 2532 if (mq->mq_curmsgs) { 2533 revents |= events & (POLLIN | POLLRDNORM); 2534 } else { 2535 mq->mq_flags |= MQ_RSEL; 2536 selrecord(td, &mq->mq_rsel); 2537 } 2538 } 2539 if (events & POLLOUT) { 2540 if (mq->mq_curmsgs < mq->mq_maxmsg) 2541 revents |= POLLOUT; 2542 else { 2543 mq->mq_flags |= MQ_WSEL; 2544 selrecord(td, &mq->mq_wsel); 2545 } 2546 } 2547 mtx_unlock(&mq->mq_mutex); 2548 return (revents); 2549 } 2550 2551 static int 2552 mqf_close(struct file *fp, struct thread *td) 2553 { 2554 struct mqfs_node *pn; 2555 2556 fp->f_ops = &badfileops; 2557 pn = fp->f_data; 2558 fp->f_data = NULL; 2559 sx_xlock(&mqfs_data.mi_lock); 2560 mqnode_release(pn); 2561 sx_xunlock(&mqfs_data.mi_lock); 2562 return (0); 2563 } 2564 2565 static int 2566 mqf_stat(struct file *fp, struct stat *st, struct ucred *active_cred) 2567 { 2568 struct mqfs_node *pn = fp->f_data; 2569 2570 bzero(st, sizeof *st); 2571 sx_xlock(&mqfs_data.mi_lock); 2572 st->st_atim = pn->mn_atime; 2573 st->st_mtim = pn->mn_mtime; 2574 st->st_ctim = pn->mn_ctime; 2575 st->st_birthtim = pn->mn_birth; 2576 st->st_uid = pn->mn_uid; 2577 st->st_gid = pn->mn_gid; 2578 st->st_mode = S_IFIFO | pn->mn_mode; 2579 sx_xunlock(&mqfs_data.mi_lock); 2580 return (0); 2581 } 2582 2583 static int 2584 mqf_chmod(struct file *fp, mode_t mode, struct ucred *active_cred, 2585 struct thread *td) 2586 { 2587 struct mqfs_node *pn; 2588 int error; 2589 2590 error = 0; 2591 pn = fp->f_data; 2592 sx_xlock(&mqfs_data.mi_lock); 2593 error = vaccess(VREG, pn->mn_mode, pn->mn_uid, pn->mn_gid, VADMIN, 2594 active_cred); 2595 if (error != 0) 2596 goto out; 2597 pn->mn_mode = mode & ACCESSPERMS; 2598 out: 2599 sx_xunlock(&mqfs_data.mi_lock); 2600 return (error); 2601 } 2602 2603 static int 2604 mqf_chown(struct file *fp, uid_t uid, gid_t gid, struct ucred *active_cred, 2605 struct thread *td) 2606 { 2607 struct mqfs_node *pn; 2608 int error; 2609 2610 error = 0; 2611 pn = fp->f_data; 2612 sx_xlock(&mqfs_data.mi_lock); 2613 if (uid == (uid_t)-1) 2614 uid = pn->mn_uid; 2615 if (gid == (gid_t)-1) 2616 gid = pn->mn_gid; 2617 if (((uid != pn->mn_uid && uid != active_cred->cr_uid) || 2618 (gid != pn->mn_gid && !groupmember(gid, active_cred))) && 2619 (error = priv_check_cred(active_cred, PRIV_VFS_CHOWN))) 2620 goto out; 2621 pn->mn_uid = uid; 2622 pn->mn_gid = gid; 2623 out: 2624 sx_xunlock(&mqfs_data.mi_lock); 2625 return (error); 2626 } 2627 2628 static int 2629 mqf_kqfilter(struct file *fp, struct knote *kn) 2630 { 2631 struct mqueue *mq = FPTOMQ(fp); 2632 int error = 0; 2633 2634 if (kn->kn_filter == EVFILT_READ) { 2635 kn->kn_fop = &mq_rfiltops; 2636 knlist_add(&mq->mq_rsel.si_note, kn, 0); 2637 } else if (kn->kn_filter == EVFILT_WRITE) { 2638 kn->kn_fop = &mq_wfiltops; 2639 knlist_add(&mq->mq_wsel.si_note, kn, 0); 2640 } else 2641 error = EINVAL; 2642 return (error); 2643 } 2644 2645 static void 2646 filt_mqdetach(struct knote *kn) 2647 { 2648 struct mqueue *mq = FPTOMQ(kn->kn_fp); 2649 2650 if (kn->kn_filter == EVFILT_READ) 2651 knlist_remove(&mq->mq_rsel.si_note, kn, 0); 2652 else if (kn->kn_filter == EVFILT_WRITE) 2653 knlist_remove(&mq->mq_wsel.si_note, kn, 0); 2654 else 2655 panic("filt_mqdetach"); 2656 } 2657 2658 static int 2659 filt_mqread(struct knote *kn, long hint) 2660 { 2661 struct mqueue *mq = FPTOMQ(kn->kn_fp); 2662 2663 mtx_assert(&mq->mq_mutex, MA_OWNED); 2664 return (mq->mq_curmsgs != 0); 2665 } 2666 2667 static int 2668 filt_mqwrite(struct knote *kn, long hint) 2669 { 2670 struct mqueue *mq = FPTOMQ(kn->kn_fp); 2671 2672 mtx_assert(&mq->mq_mutex, MA_OWNED); 2673 return (mq->mq_curmsgs < mq->mq_maxmsg); 2674 } 2675 2676 static int 2677 mqf_fill_kinfo(struct file *fp, struct kinfo_file *kif, struct filedesc *fdp) 2678 { 2679 2680 kif->kf_type = KF_TYPE_MQUEUE; 2681 return (0); 2682 } 2683 2684 static struct fileops mqueueops = { 2685 .fo_read = invfo_rdwr, 2686 .fo_write = invfo_rdwr, 2687 .fo_truncate = invfo_truncate, 2688 .fo_ioctl = invfo_ioctl, 2689 .fo_poll = mqf_poll, 2690 .fo_kqfilter = mqf_kqfilter, 2691 .fo_stat = mqf_stat, 2692 .fo_close = mqf_close, 2693 .fo_chmod = mqf_chmod, 2694 .fo_chown = mqf_chown, 2695 .fo_sendfile = invfo_sendfile, 2696 .fo_fill_kinfo = mqf_fill_kinfo, 2697 .fo_cmp = file_kcmp_generic, 2698 .fo_flags = DFLAG_PASSABLE, 2699 }; 2700 2701 static struct vop_vector mqfs_vnodeops = { 2702 .vop_default = &default_vnodeops, 2703 .vop_access = mqfs_access, 2704 .vop_cachedlookup = mqfs_lookup, 2705 .vop_lookup = vfs_cache_lookup, 2706 .vop_reclaim = mqfs_reclaim, 2707 .vop_create = mqfs_create, 2708 .vop_remove = mqfs_remove, 2709 .vop_inactive = mqfs_inactive, 2710 .vop_open = mqfs_open, 2711 .vop_close = mqfs_close, 2712 .vop_getattr = mqfs_getattr, 2713 .vop_setattr = mqfs_setattr, 2714 .vop_read = mqfs_read, 2715 .vop_write = VOP_EOPNOTSUPP, 2716 .vop_readdir = mqfs_readdir, 2717 .vop_mkdir = VOP_EOPNOTSUPP, 2718 .vop_rmdir = VOP_EOPNOTSUPP 2719 }; 2720 VFS_VOP_VECTOR_REGISTER(mqfs_vnodeops); 2721 2722 static struct vfsops mqfs_vfsops = { 2723 .vfs_init = mqfs_init, 2724 .vfs_uninit = mqfs_uninit, 2725 .vfs_mount = mqfs_mount, 2726 .vfs_unmount = mqfs_unmount, 2727 .vfs_root = mqfs_root, 2728 .vfs_statfs = mqfs_statfs, 2729 }; 2730 2731 static struct vfsconf mqueuefs_vfsconf = { 2732 .vfc_version = VFS_VERSION, 2733 .vfc_name = "mqueuefs", 2734 .vfc_vfsops = &mqfs_vfsops, 2735 .vfc_typenum = -1, 2736 .vfc_flags = VFCF_SYNTHETIC 2737 }; 2738 2739 static struct syscall_helper_data mq_syscalls[] = { 2740 SYSCALL_INIT_HELPER(kmq_open), 2741 SYSCALL_INIT_HELPER_F(kmq_setattr, SYF_CAPENABLED), 2742 SYSCALL_INIT_HELPER_F(kmq_timedsend, SYF_CAPENABLED), 2743 SYSCALL_INIT_HELPER_F(kmq_timedreceive, SYF_CAPENABLED), 2744 SYSCALL_INIT_HELPER_F(kmq_notify, SYF_CAPENABLED), 2745 SYSCALL_INIT_HELPER(kmq_unlink), 2746 SYSCALL_INIT_LAST 2747 }; 2748 2749 #ifdef COMPAT_FREEBSD32 2750 #include <compat/freebsd32/freebsd32.h> 2751 #include <compat/freebsd32/freebsd32_proto.h> 2752 #include <compat/freebsd32/freebsd32_signal.h> 2753 #include <compat/freebsd32/freebsd32_syscall.h> 2754 #include <compat/freebsd32/freebsd32_util.h> 2755 2756 static void 2757 mq_attr_from32(const struct mq_attr32 *from, struct mq_attr *to) 2758 { 2759 2760 to->mq_flags = from->mq_flags; 2761 to->mq_maxmsg = from->mq_maxmsg; 2762 to->mq_msgsize = from->mq_msgsize; 2763 to->mq_curmsgs = from->mq_curmsgs; 2764 } 2765 2766 static void 2767 mq_attr_to32(const struct mq_attr *from, struct mq_attr32 *to) 2768 { 2769 2770 to->mq_flags = from->mq_flags; 2771 to->mq_maxmsg = from->mq_maxmsg; 2772 to->mq_msgsize = from->mq_msgsize; 2773 to->mq_curmsgs = from->mq_curmsgs; 2774 } 2775 2776 int 2777 freebsd32_kmq_open(struct thread *td, struct freebsd32_kmq_open_args *uap) 2778 { 2779 struct mq_attr attr; 2780 struct mq_attr32 attr32; 2781 int flags, error; 2782 2783 if ((uap->flags & O_ACCMODE) == O_ACCMODE || uap->flags & O_EXEC) 2784 return (EINVAL); 2785 flags = FFLAGS(uap->flags); 2786 if ((flags & O_CREAT) != 0 && uap->attr != NULL) { 2787 error = copyin(uap->attr, &attr32, sizeof(attr32)); 2788 if (error) 2789 return (error); 2790 mq_attr_from32(&attr32, &attr); 2791 } 2792 return (kern_kmq_open(td, uap->path, flags, uap->mode, 2793 uap->attr != NULL ? &attr : NULL)); 2794 } 2795 2796 int 2797 freebsd32_kmq_setattr(struct thread *td, struct freebsd32_kmq_setattr_args *uap) 2798 { 2799 struct mq_attr attr, oattr; 2800 struct mq_attr32 attr32, oattr32; 2801 int error; 2802 2803 if (uap->attr != NULL) { 2804 error = copyin(uap->attr, &attr32, sizeof(attr32)); 2805 if (error != 0) 2806 return (error); 2807 mq_attr_from32(&attr32, &attr); 2808 } 2809 error = kern_kmq_setattr(td, uap->mqd, uap->attr != NULL ? &attr : NULL, 2810 &oattr); 2811 if (error == 0 && uap->oattr != NULL) { 2812 mq_attr_to32(&oattr, &oattr32); 2813 bzero(oattr32.__reserved, sizeof(oattr32.__reserved)); 2814 error = copyout(&oattr32, uap->oattr, sizeof(oattr32)); 2815 } 2816 return (error); 2817 } 2818 2819 int 2820 freebsd32_kmq_timedsend(struct thread *td, 2821 struct freebsd32_kmq_timedsend_args *uap) 2822 { 2823 struct timespec32 ets32; 2824 struct timespec *abs_timeout, ets; 2825 int error; 2826 2827 if (uap->abs_timeout != NULL) { 2828 error = copyin(uap->abs_timeout, &ets32, sizeof(ets32)); 2829 if (error != 0) 2830 return (error); 2831 CP(ets32, ets, tv_sec); 2832 CP(ets32, ets, tv_nsec); 2833 abs_timeout = &ets; 2834 } else 2835 abs_timeout = NULL; 2836 2837 return (kern_kmq_timedsend(td, uap->mqd, uap->msg_ptr, uap->msg_len, 2838 uap->msg_prio, abs_timeout)); 2839 } 2840 2841 int 2842 freebsd32_kmq_timedreceive(struct thread *td, 2843 struct freebsd32_kmq_timedreceive_args *uap) 2844 { 2845 struct timespec32 ets32; 2846 struct timespec *abs_timeout, ets; 2847 int error; 2848 2849 if (uap->abs_timeout != NULL) { 2850 error = copyin(uap->abs_timeout, &ets32, sizeof(ets32)); 2851 if (error != 0) 2852 return (error); 2853 CP(ets32, ets, tv_sec); 2854 CP(ets32, ets, tv_nsec); 2855 abs_timeout = &ets; 2856 } else 2857 abs_timeout = NULL; 2858 2859 return (kern_kmq_timedreceive(td, uap->mqd, uap->msg_ptr, uap->msg_len, 2860 uap->msg_prio, abs_timeout)); 2861 } 2862 2863 int 2864 freebsd32_kmq_notify(struct thread *td, struct freebsd32_kmq_notify_args *uap) 2865 { 2866 struct sigevent ev, *evp; 2867 struct sigevent32 ev32; 2868 int error; 2869 2870 if (uap->sigev == NULL) { 2871 evp = NULL; 2872 } else { 2873 error = copyin(uap->sigev, &ev32, sizeof(ev32)); 2874 if (error != 0) 2875 return (error); 2876 error = convert_sigevent32(&ev32, &ev); 2877 if (error != 0) 2878 return (error); 2879 evp = &ev; 2880 } 2881 return (kern_kmq_notify(td, uap->mqd, evp)); 2882 } 2883 2884 static struct syscall_helper_data mq32_syscalls[] = { 2885 SYSCALL32_INIT_HELPER(freebsd32_kmq_open), 2886 SYSCALL32_INIT_HELPER_F(freebsd32_kmq_setattr, SYF_CAPENABLED), 2887 SYSCALL32_INIT_HELPER_F(freebsd32_kmq_timedsend, SYF_CAPENABLED), 2888 SYSCALL32_INIT_HELPER_F(freebsd32_kmq_timedreceive, SYF_CAPENABLED), 2889 SYSCALL32_INIT_HELPER_F(freebsd32_kmq_notify, SYF_CAPENABLED), 2890 SYSCALL32_INIT_HELPER_COMPAT(kmq_unlink), 2891 SYSCALL_INIT_LAST 2892 }; 2893 #endif 2894 2895 static int 2896 mqinit(void) 2897 { 2898 int error; 2899 2900 error = syscall_helper_register(mq_syscalls, SY_THR_STATIC_KLD); 2901 if (error != 0) 2902 return (error); 2903 #ifdef COMPAT_FREEBSD32 2904 error = syscall32_helper_register(mq32_syscalls, SY_THR_STATIC_KLD); 2905 if (error != 0) 2906 return (error); 2907 #endif 2908 return (0); 2909 } 2910 2911 static int 2912 mqunload(void) 2913 { 2914 2915 #ifdef COMPAT_FREEBSD32 2916 syscall32_helper_unregister(mq32_syscalls); 2917 #endif 2918 syscall_helper_unregister(mq_syscalls); 2919 return (0); 2920 } 2921 2922 static int 2923 mq_modload(struct module *module, int cmd, void *arg) 2924 { 2925 int error = 0; 2926 2927 error = vfs_modevent(module, cmd, arg); 2928 if (error != 0) 2929 return (error); 2930 2931 switch (cmd) { 2932 case MOD_LOAD: 2933 error = mqinit(); 2934 if (error != 0) 2935 mqunload(); 2936 break; 2937 case MOD_UNLOAD: 2938 error = mqunload(); 2939 break; 2940 default: 2941 break; 2942 } 2943 return (error); 2944 } 2945 2946 static moduledata_t mqueuefs_mod = { 2947 "mqueuefs", 2948 mq_modload, 2949 &mqueuefs_vfsconf 2950 }; 2951 DECLARE_MODULE(mqueuefs, mqueuefs_mod, SI_SUB_VFS, SI_ORDER_MIDDLE); 2952 MODULE_VERSION(mqueuefs, 1); 2953