1 /*- 2 * Copyright (c) 1998 Berkeley Software Design, Inc. All rights reserved. 3 * 4 * Redistribution and use in source and binary forms, with or without 5 * modification, are permitted provided that the following conditions 6 * are met: 7 * 1. Redistributions of source code must retain the above copyright 8 * notice, this list of conditions and the following disclaimer. 9 * 2. Redistributions in binary form must reproduce the above copyright 10 * notice, this list of conditions and the following disclaimer in the 11 * documentation and/or other materials provided with the distribution. 12 * 3. Berkeley Software Design Inc's name may not be used to endorse or 13 * promote products derived from this software without specific prior 14 * written permission. 15 * 16 * THIS SOFTWARE IS PROVIDED BY BERKELEY SOFTWARE DESIGN INC ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL BERKELEY SOFTWARE DESIGN INC BE LIABLE 20 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 * 28 * from BSDI $Id: mutex_witness.c,v 1.1.2.20 2000/04/27 03:10:27 cp Exp $ 29 * and BSDI $Id: synch_machdep.c,v 2.3.2.39 2000/04/27 03:10:25 cp Exp $ 30 * $FreeBSD$ 31 */ 32 33 /* 34 * Machine independent bits of mutex implementation and implementation of 35 * `witness' structure & related debugging routines. 36 */ 37 38 /* 39 * Main Entry: witness 40 * Pronunciation: 'wit-n&s 41 * Function: noun 42 * Etymology: Middle English witnesse, from Old English witnes knowledge, 43 * testimony, witness, from 2wit 44 * Date: before 12th century 45 * 1 : attestation of a fact or event : TESTIMONY 46 * 2 : one that gives evidence; specifically : one who testifies in 47 * a cause or before a judicial tribunal 48 * 3 : one asked to be present at a transaction so as to be able to 49 * testify to its having taken place 50 * 4 : one who has personal knowledge of something 51 * 5 a : something serving as evidence or proof : SIGN 52 * b : public affirmation by word or example of usually 53 * religious faith or conviction <the heroic witness to divine 54 * life -- Pilot> 55 * 6 capitalized : a member of the Jehovah's Witnesses 56 */ 57 58 #include "opt_ddb.h" 59 60 #include <sys/param.h> 61 #include <sys/bus.h> 62 #include <sys/kernel.h> 63 #include <sys/lock.h> 64 #include <sys/malloc.h> 65 #include <sys/mutex.h> 66 #include <sys/proc.h> 67 #include <sys/resourcevar.h> 68 #include <sys/sysctl.h> 69 #include <sys/systm.h> 70 #include <sys/vmmeter.h> 71 #include <sys/ktr.h> 72 73 #include <machine/atomic.h> 74 #include <machine/bus.h> 75 #include <machine/clock.h> 76 #include <machine/cpu.h> 77 78 #include <ddb/ddb.h> 79 80 #include <vm/vm.h> 81 #include <vm/vm_extern.h> 82 83 /* 84 * Internal utility macros. 85 */ 86 #define mtx_unowned(m) ((m)->mtx_lock == MTX_UNOWNED) 87 88 #define mtx_owner(m) (mtx_unowned((m)) ? NULL \ 89 : (struct proc *)((m)->mtx_lock & MTX_FLAGMASK)) 90 91 #define SET_PRIO(p, pri) (p)->p_pri.pri_level = (pri) 92 93 /* 94 * Lock classes for sleep and spin mutexes. 95 */ 96 struct lock_class lock_class_mtx_sleep = { 97 "sleep mutex", 98 LC_SLEEPLOCK | LC_RECURSABLE 99 }; 100 struct lock_class lock_class_mtx_spin = { 101 "spin mutex", 102 LC_SPINLOCK | LC_RECURSABLE 103 }; 104 105 /* 106 * Prototypes for non-exported routines. 107 */ 108 static void propagate_priority(struct proc *); 109 110 static void 111 propagate_priority(struct proc *p) 112 { 113 int pri = p->p_pri.pri_level; 114 struct mtx *m = p->p_blocked; 115 116 mtx_assert(&sched_lock, MA_OWNED); 117 for (;;) { 118 struct proc *p1; 119 120 p = mtx_owner(m); 121 122 if (p == NULL) { 123 /* 124 * This really isn't quite right. Really 125 * ought to bump priority of process that 126 * next acquires the mutex. 127 */ 128 MPASS(m->mtx_lock == MTX_CONTESTED); 129 return; 130 } 131 132 MPASS(p->p_magic == P_MAGIC); 133 KASSERT(p->p_stat != SSLEEP, ("sleeping process owns a mutex")); 134 if (p->p_pri.pri_level <= pri) 135 return; 136 137 /* 138 * Bump this process' priority. 139 */ 140 SET_PRIO(p, pri); 141 142 /* 143 * If lock holder is actually running, just bump priority. 144 */ 145 if (p->p_oncpu != NOCPU) { 146 MPASS(p->p_stat == SRUN || p->p_stat == SZOMB || p->p_stat == SSTOP); 147 return; 148 } 149 150 #ifndef SMP 151 /* 152 * For UP, we check to see if p is curproc (this shouldn't 153 * ever happen however as it would mean we are in a deadlock.) 154 */ 155 KASSERT(p != curproc, ("Deadlock detected")); 156 #endif 157 158 /* 159 * If on run queue move to new run queue, and 160 * quit. 161 */ 162 if (p->p_stat == SRUN) { 163 MPASS(p->p_blocked == NULL); 164 remrunqueue(p); 165 setrunqueue(p); 166 return; 167 } 168 169 /* 170 * If we aren't blocked on a mutex, we should be. 171 */ 172 KASSERT(p->p_stat == SMTX, ( 173 "process %d(%s):%d holds %s but isn't blocked on a mutex\n", 174 p->p_pid, p->p_comm, p->p_stat, 175 m->mtx_object.lo_name)); 176 177 /* 178 * Pick up the mutex that p is blocked on. 179 */ 180 m = p->p_blocked; 181 MPASS(m != NULL); 182 183 /* 184 * Check if the proc needs to be moved up on 185 * the blocked chain 186 */ 187 if (p == TAILQ_FIRST(&m->mtx_blocked)) { 188 continue; 189 } 190 191 p1 = TAILQ_PREV(p, procqueue, p_procq); 192 if (p1->p_pri.pri_level <= pri) { 193 continue; 194 } 195 196 /* 197 * Remove proc from blocked chain and determine where 198 * it should be moved up to. Since we know that p1 has 199 * a lower priority than p, we know that at least one 200 * process in the chain has a lower priority and that 201 * p1 will thus not be NULL after the loop. 202 */ 203 TAILQ_REMOVE(&m->mtx_blocked, p, p_procq); 204 TAILQ_FOREACH(p1, &m->mtx_blocked, p_procq) { 205 MPASS(p1->p_magic == P_MAGIC); 206 if (p1->p_pri.pri_level > pri) 207 break; 208 } 209 210 MPASS(p1 != NULL); 211 TAILQ_INSERT_BEFORE(p1, p, p_procq); 212 CTR4(KTR_LOCK, 213 "propagate_priority: p %p moved before %p on [%p] %s", 214 p, p1, m, m->mtx_object.lo_name); 215 } 216 } 217 218 /* 219 * Function versions of the inlined __mtx_* macros. These are used by 220 * modules and can also be called from assembly language if needed. 221 */ 222 void 223 _mtx_lock_flags(struct mtx *m, int opts, const char *file, int line) 224 { 225 226 __mtx_lock_flags(m, opts, file, line); 227 } 228 229 void 230 _mtx_unlock_flags(struct mtx *m, int opts, const char *file, int line) 231 { 232 233 __mtx_unlock_flags(m, opts, file, line); 234 } 235 236 void 237 _mtx_lock_spin_flags(struct mtx *m, int opts, const char *file, int line) 238 { 239 240 __mtx_lock_spin_flags(m, opts, file, line); 241 } 242 243 void 244 _mtx_unlock_spin_flags(struct mtx *m, int opts, const char *file, int line) 245 { 246 247 __mtx_unlock_spin_flags(m, opts, file, line); 248 } 249 250 /* 251 * The important part of mtx_trylock{,_flags}() 252 * Tries to acquire lock `m.' We do NOT handle recursion here; we assume that 253 * if we're called, it's because we know we don't already own this lock. 254 */ 255 int 256 _mtx_trylock(struct mtx *m, int opts, const char *file, int line) 257 { 258 int rval; 259 260 MPASS(curproc != NULL); 261 262 /* 263 * _mtx_trylock does not accept MTX_NOSWITCH option. 264 */ 265 KASSERT((opts & MTX_NOSWITCH) == 0, 266 ("mtx_trylock() called with invalid option flag(s) %d", opts)); 267 268 rval = _obtain_lock(m, curproc); 269 270 LOCK_LOG_TRY("LOCK", &m->mtx_object, opts, rval, file, line); 271 if (rval) { 272 /* 273 * We do not handle recursion in _mtx_trylock; see the 274 * note at the top of the routine. 275 */ 276 KASSERT(!mtx_recursed(m), 277 ("mtx_trylock() called on a recursed mutex")); 278 WITNESS_LOCK(&m->mtx_object, opts | LOP_EXCLUSIVE | LOP_TRYLOCK, 279 file, line); 280 } 281 282 return (rval); 283 } 284 285 /* 286 * _mtx_lock_sleep: the tougher part of acquiring an MTX_DEF lock. 287 * 288 * We call this if the lock is either contested (i.e. we need to go to 289 * sleep waiting for it), or if we need to recurse on it. 290 */ 291 void 292 _mtx_lock_sleep(struct mtx *m, int opts, const char *file, int line) 293 { 294 struct proc *p = curproc; 295 296 if ((m->mtx_lock & MTX_FLAGMASK) == (uintptr_t)p) { 297 m->mtx_recurse++; 298 atomic_set_ptr(&m->mtx_lock, MTX_RECURSED); 299 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 300 CTR1(KTR_LOCK, "_mtx_lock_sleep: %p recursing", m); 301 return; 302 } 303 304 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 305 CTR4(KTR_LOCK, 306 "_mtx_lock_sleep: %s contested (lock=%p) at %s:%d", 307 m->mtx_object.lo_name, (void *)m->mtx_lock, file, line); 308 309 while (!_obtain_lock(m, p)) { 310 uintptr_t v; 311 struct proc *p1; 312 313 mtx_lock_spin(&sched_lock); 314 /* 315 * Check if the lock has been released while spinning for 316 * the sched_lock. 317 */ 318 if ((v = m->mtx_lock) == MTX_UNOWNED) { 319 mtx_unlock_spin(&sched_lock); 320 continue; 321 } 322 323 /* 324 * The mutex was marked contested on release. This means that 325 * there are processes blocked on it. 326 */ 327 if (v == MTX_CONTESTED) { 328 p1 = TAILQ_FIRST(&m->mtx_blocked); 329 MPASS(p1 != NULL); 330 m->mtx_lock = (uintptr_t)p | MTX_CONTESTED; 331 332 if (p1->p_pri.pri_level < p->p_pri.pri_level) 333 SET_PRIO(p, p1->p_pri.pri_level); 334 mtx_unlock_spin(&sched_lock); 335 return; 336 } 337 338 /* 339 * If the mutex isn't already contested and a failure occurs 340 * setting the contested bit, the mutex was either released 341 * or the state of the MTX_RECURSED bit changed. 342 */ 343 if ((v & MTX_CONTESTED) == 0 && 344 !atomic_cmpset_ptr(&m->mtx_lock, (void *)v, 345 (void *)(v | MTX_CONTESTED))) { 346 mtx_unlock_spin(&sched_lock); 347 continue; 348 } 349 350 /* 351 * We deffinately must sleep for this lock. 352 */ 353 mtx_assert(m, MA_NOTOWNED); 354 355 #ifdef notyet 356 /* 357 * If we're borrowing an interrupted thread's VM context, we 358 * must clean up before going to sleep. 359 */ 360 if (p->p_ithd != NULL) { 361 struct ithd *it = p->p_ithd; 362 363 if (it->it_interrupted) { 364 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 365 CTR2(KTR_LOCK, 366 "_mtx_lock_sleep: %p interrupted %p", 367 it, it->it_interrupted); 368 intr_thd_fixup(it); 369 } 370 } 371 #endif 372 373 /* 374 * Put us on the list of threads blocked on this mutex. 375 */ 376 if (TAILQ_EMPTY(&m->mtx_blocked)) { 377 p1 = (struct proc *)(m->mtx_lock & MTX_FLAGMASK); 378 LIST_INSERT_HEAD(&p1->p_contested, m, mtx_contested); 379 TAILQ_INSERT_TAIL(&m->mtx_blocked, p, p_procq); 380 } else { 381 TAILQ_FOREACH(p1, &m->mtx_blocked, p_procq) 382 if (p1->p_pri.pri_level > p->p_pri.pri_level) 383 break; 384 if (p1) 385 TAILQ_INSERT_BEFORE(p1, p, p_procq); 386 else 387 TAILQ_INSERT_TAIL(&m->mtx_blocked, p, p_procq); 388 } 389 390 /* 391 * Save who we're blocked on. 392 */ 393 p->p_blocked = m; 394 p->p_mtxname = m->mtx_object.lo_name; 395 p->p_stat = SMTX; 396 propagate_priority(p); 397 398 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 399 CTR3(KTR_LOCK, 400 "_mtx_lock_sleep: p %p blocked on [%p] %s", p, m, 401 m->mtx_object.lo_name); 402 403 p->p_stats->p_ru.ru_nvcsw++; 404 mi_switch(); 405 406 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 407 CTR3(KTR_LOCK, 408 "_mtx_lock_sleep: p %p free from blocked on [%p] %s", 409 p, m, m->mtx_object.lo_name); 410 411 mtx_unlock_spin(&sched_lock); 412 } 413 414 return; 415 } 416 417 /* 418 * _mtx_lock_spin: the tougher part of acquiring an MTX_SPIN lock. 419 * 420 * This is only called if we need to actually spin for the lock. Recursion 421 * is handled inline. 422 */ 423 void 424 _mtx_lock_spin(struct mtx *m, int opts, critical_t mtx_crit, const char *file, 425 int line) 426 { 427 int i = 0; 428 429 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 430 CTR1(KTR_LOCK, "_mtx_lock_spin: %p spinning", m); 431 432 for (;;) { 433 if (_obtain_lock(m, curproc)) 434 break; 435 436 /* Give interrupts a chance while we spin. */ 437 critical_exit(mtx_crit); 438 while (m->mtx_lock != MTX_UNOWNED) { 439 if (i++ < 1000000) 440 continue; 441 if (i++ < 6000000) 442 DELAY(1); 443 #ifdef DDB 444 else if (!db_active) 445 #else 446 else 447 #endif 448 panic("spin lock %s held by %p for > 5 seconds", 449 m->mtx_object.lo_name, (void *)m->mtx_lock); 450 } 451 mtx_crit = critical_enter(); 452 } 453 454 m->mtx_savecrit = mtx_crit; 455 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 456 CTR1(KTR_LOCK, "_mtx_lock_spin: %p spin done", m); 457 458 return; 459 } 460 461 /* 462 * _mtx_unlock_sleep: the tougher part of releasing an MTX_DEF lock. 463 * 464 * We are only called here if the lock is recursed or contested (i.e. we 465 * need to wake up a blocked thread). 466 */ 467 void 468 _mtx_unlock_sleep(struct mtx *m, int opts, const char *file, int line) 469 { 470 struct proc *p, *p1; 471 struct mtx *m1; 472 int pri; 473 474 p = curproc; 475 476 if (mtx_recursed(m)) { 477 if (--(m->mtx_recurse) == 0) 478 atomic_clear_ptr(&m->mtx_lock, MTX_RECURSED); 479 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 480 CTR1(KTR_LOCK, "_mtx_unlock_sleep: %p unrecurse", m); 481 return; 482 } 483 484 mtx_lock_spin(&sched_lock); 485 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 486 CTR1(KTR_LOCK, "_mtx_unlock_sleep: %p contested", m); 487 488 p1 = TAILQ_FIRST(&m->mtx_blocked); 489 MPASS(p->p_magic == P_MAGIC); 490 MPASS(p1->p_magic == P_MAGIC); 491 492 TAILQ_REMOVE(&m->mtx_blocked, p1, p_procq); 493 494 if (TAILQ_EMPTY(&m->mtx_blocked)) { 495 LIST_REMOVE(m, mtx_contested); 496 _release_lock_quick(m); 497 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 498 CTR1(KTR_LOCK, "_mtx_unlock_sleep: %p not held", m); 499 } else 500 atomic_store_rel_ptr(&m->mtx_lock, (void *)MTX_CONTESTED); 501 502 pri = PRI_MAX; 503 LIST_FOREACH(m1, &p->p_contested, mtx_contested) { 504 int cp = TAILQ_FIRST(&m1->mtx_blocked)->p_pri.pri_level; 505 if (cp < pri) 506 pri = cp; 507 } 508 509 if (pri > p->p_pri.pri_native) 510 pri = p->p_pri.pri_native; 511 SET_PRIO(p, pri); 512 513 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 514 CTR2(KTR_LOCK, "_mtx_unlock_sleep: %p contested setrunqueue %p", 515 m, p1); 516 517 p1->p_blocked = NULL; 518 p1->p_stat = SRUN; 519 setrunqueue(p1); 520 521 if ((opts & MTX_NOSWITCH) == 0 && p1->p_pri.pri_level < pri) { 522 #ifdef notyet 523 if (p->p_ithd != NULL) { 524 struct ithd *it = p->p_ithd; 525 526 if (it->it_interrupted) { 527 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 528 CTR2(KTR_LOCK, 529 "_mtx_unlock_sleep: %p interrupted %p", 530 it, it->it_interrupted); 531 intr_thd_fixup(it); 532 } 533 } 534 #endif 535 setrunqueue(p); 536 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 537 CTR2(KTR_LOCK, 538 "_mtx_unlock_sleep: %p switching out lock=%p", m, 539 (void *)m->mtx_lock); 540 541 p->p_stats->p_ru.ru_nivcsw++; 542 mi_switch(); 543 if (LOCK_LOG_TEST(&m->mtx_object, opts)) 544 CTR2(KTR_LOCK, "_mtx_unlock_sleep: %p resuming lock=%p", 545 m, (void *)m->mtx_lock); 546 } 547 548 mtx_unlock_spin(&sched_lock); 549 550 return; 551 } 552 553 /* 554 * All the unlocking of MTX_SPIN locks is done inline. 555 * See the _rel_spin_lock() macro for the details. 556 */ 557 558 /* 559 * The backing function for the INVARIANTS-enabled mtx_assert() 560 */ 561 #ifdef INVARIANT_SUPPORT 562 void 563 _mtx_assert(struct mtx *m, int what, const char *file, int line) 564 { 565 switch (what) { 566 case MA_OWNED: 567 case MA_OWNED | MA_RECURSED: 568 case MA_OWNED | MA_NOTRECURSED: 569 if (!mtx_owned(m)) 570 panic("mutex %s not owned at %s:%d", 571 m->mtx_object.lo_name, file, line); 572 if (mtx_recursed(m)) { 573 if ((what & MA_NOTRECURSED) != 0) 574 panic("mutex %s recursed at %s:%d", 575 m->mtx_object.lo_name, file, line); 576 } else if ((what & MA_RECURSED) != 0) { 577 panic("mutex %s unrecursed at %s:%d", 578 m->mtx_object.lo_name, file, line); 579 } 580 break; 581 case MA_NOTOWNED: 582 if (mtx_owned(m)) 583 panic("mutex %s owned at %s:%d", 584 m->mtx_object.lo_name, file, line); 585 break; 586 default: 587 panic("unknown mtx_assert at %s:%d", file, line); 588 } 589 } 590 #endif 591 592 /* 593 * The MUTEX_DEBUG-enabled mtx_validate() 594 * 595 * Most of these checks have been moved off into the LO_INITIALIZED flag 596 * maintained by the witness code. 597 */ 598 #ifdef MUTEX_DEBUG 599 600 void mtx_validate __P((struct mtx *)); 601 602 void 603 mtx_validate(struct mtx *m) 604 { 605 606 /* 607 * XXX - When kernacc() is fixed on the alpha to handle K0_SEG memory properly 608 * we can re-enable the kernacc() checks. 609 */ 610 #ifndef __alpha__ 611 if (!kernacc((caddr_t)m, sizeof(m), VM_PROT_READ | VM_PROT_WRITE)) 612 panic("Can't read and write to mutex %p", m); 613 #endif 614 } 615 #endif 616 617 /* 618 * Mutex initialization routine; initialize lock `m' of type contained in 619 * `opts' with options contained in `opts' and description `description.' 620 */ 621 void 622 mtx_init(struct mtx *m, const char *description, int opts) 623 { 624 struct lock_object *lock; 625 626 MPASS((opts & ~(MTX_SPIN | MTX_QUIET | MTX_RECURSE | 627 MTX_SLEEPABLE | MTX_NOWITNESS)) == 0); 628 629 #ifdef MUTEX_DEBUG 630 /* Diagnostic and error correction */ 631 mtx_validate(m); 632 #endif 633 634 bzero(m, sizeof(*m)); 635 lock = &m->mtx_object; 636 if (opts & MTX_SPIN) 637 lock->lo_class = &lock_class_mtx_spin; 638 else 639 lock->lo_class = &lock_class_mtx_sleep; 640 lock->lo_name = description; 641 if (opts & MTX_QUIET) 642 lock->lo_flags = LO_QUIET; 643 if (opts & MTX_RECURSE) 644 lock->lo_flags |= LO_RECURSABLE; 645 if (opts & MTX_SLEEPABLE) 646 lock->lo_flags |= LO_SLEEPABLE; 647 if ((opts & MTX_NOWITNESS) == 0) 648 lock->lo_flags |= LO_WITNESS; 649 650 m->mtx_lock = MTX_UNOWNED; 651 TAILQ_INIT(&m->mtx_blocked); 652 653 LOCK_LOG_INIT(lock, opts); 654 655 WITNESS_INIT(lock); 656 } 657 658 /* 659 * Remove lock `m' from all_mtx queue. We don't allow MTX_QUIET to be 660 * passed in as a flag here because if the corresponding mtx_init() was 661 * called with MTX_QUIET set, then it will already be set in the mutex's 662 * flags. 663 */ 664 void 665 mtx_destroy(struct mtx *m) 666 { 667 668 LOCK_LOG_DESTROY(&m->mtx_object, 0); 669 670 if (!mtx_owned(m)) 671 MPASS(mtx_unowned(m)); 672 else { 673 MPASS((m->mtx_lock & (MTX_RECURSED|MTX_CONTESTED)) == 0); 674 675 /* Tell witness this isn't locked to make it happy. */ 676 WITNESS_UNLOCK(&m->mtx_object, LOP_EXCLUSIVE | LOP_NOSWITCH, 677 __FILE__, __LINE__); 678 } 679 680 WITNESS_DESTROY(&m->mtx_object); 681 } 682