1 /* 2 * Copyright (c) 2001 Jake Burkholder <jake@FreeBSD.org> 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 * 26 * $FreeBSD$ 27 */ 28 29 /*** 30 31 Here is the logic.. 32 33 If there are N processors, then there are at most N KSEs (kernel 34 schedulable entities) working to process threads that belong to a 35 KSEGOUP (kg). If there are X of these KSEs actually running at the 36 moment in question, then there are at most M (N-X) of these KSEs on 37 the run queue, as running KSEs are not on the queue. 38 39 Runnable threads are queued off the KSEGROUP in priority order. 40 If there are M or more threads runnable, the top M threads 41 (by priority) are 'preassigned' to the M KSEs not running. The KSEs take 42 their priority from those threads and are put on the run queue. 43 44 The last thread that had a priority high enough to have a KSE associated 45 with it, AND IS ON THE RUN QUEUE is pointed to by 46 kg->kg_last_assigned. If no threads queued off the KSEGROUP have KSEs 47 assigned as all the available KSEs are activly running, or because there 48 are no threads queued, that pointer is NULL. 49 50 When a KSE is removed from the run queue to become runnable, we know 51 it was associated with the highest priority thread in the queue (at the head 52 of the queue). If it is also the last assigned we know M was 1 and must 53 now be 0. Since the thread is no longer queued that pointer must be 54 removed from it. Since we know there were no more KSEs available, 55 (M was 1 and is now 0) and since we are not FREEING our KSE 56 but using it, we know there are STILL no more KSEs available, we can prove 57 that the next thread in the ksegrp list will not have a KSE to assign to 58 it, so we can show that the pointer must be made 'invalid' (NULL). 59 60 The pointer exists so that when a new thread is made runnable, it can 61 have its priority compared with the last assigned thread to see if 62 it should 'steal' its KSE or not.. i.e. is it 'earlier' 63 on the list than that thread or later.. If it's earlier, then the KSE is 64 removed from the last assigned (which is now not assigned a KSE) 65 and reassigned to the new thread, which is placed earlier in the list. 66 The pointer is then backed up to the previous thread (which may or may not 67 be the new thread). 68 69 When a thread sleeps or is removed, the KSE becomes available and if there 70 are queued threads that are not assigned KSEs, the highest priority one of 71 them is assigned the KSE, which is then placed back on the run queue at 72 the approipriate place, and the kg->kg_last_assigned pointer is adjusted down 73 to point to it. 74 75 The following diagram shows 2 KSEs and 3 threads from a single process. 76 77 RUNQ: --->KSE---KSE--... (KSEs queued at priorities from threads) 78 \ \____ 79 \ \ 80 KSEGROUP---thread--thread--thread (queued in priority order) 81 \ / 82 \_______________/ 83 (last_assigned) 84 85 The result of this scheme is that the M available KSEs are always 86 queued at the priorities they have inherrited from the M highest priority 87 threads for that KSEGROUP. If this situation changes, the KSEs are 88 reassigned to keep this true. 89 90 */ 91 92 #include <sys/param.h> 93 #include <sys/systm.h> 94 #include <sys/kernel.h> 95 #include <sys/ktr.h> 96 #include <sys/lock.h> 97 #include <sys/mutex.h> 98 #include <sys/proc.h> 99 #include <sys/queue.h> 100 #include <machine/critical.h> 101 102 CTASSERT((RQB_BPW * RQB_LEN) == RQ_NQS); 103 104 /* 105 * Global run queue. 106 */ 107 static struct runq runq; 108 SYSINIT(runq, SI_SUB_RUN_QUEUE, SI_ORDER_FIRST, runq_init, &runq) 109 110 static void runq_readjust(struct runq *rq, struct kse *ke); 111 /************************************************************************ 112 * Functions that manipulate runnability from a thread perspective. * 113 ************************************************************************/ 114 115 /* 116 * Select the KSE that will be run next. From that find the thread, and x 117 * remove it from the KSEGRP's run queue. If there is thread clustering, 118 * this will be what does it. 119 */ 120 struct thread * 121 choosethread(void) 122 { 123 struct kse *ke; 124 struct thread *td; 125 struct ksegrp *kg; 126 127 retry: 128 if ((ke = runq_choose(&runq))) { 129 td = ke->ke_thread; 130 KASSERT((td->td_kse == ke), ("kse/thread mismatch")); 131 kg = ke->ke_ksegrp; 132 if (td->td_flags & TDF_UNBOUND) { 133 TAILQ_REMOVE(&kg->kg_runq, td, td_runq); 134 if (kg->kg_last_assigned == td) 135 if (TAILQ_PREV(td, threadqueue, td_runq) 136 != NULL) 137 printf("Yo MAMA!\n"); 138 kg->kg_last_assigned = TAILQ_PREV(td, 139 threadqueue, td_runq); 140 /* 141 * If we have started running an upcall, 142 * Then TDF_UNBOUND WAS set because the thread was 143 * created without a KSE. Now that we have one, 144 * and it is our time to run, we make sure 145 * that BOUND semantics apply for the rest of 146 * the journey to userland, and into the UTS. 147 */ 148 #ifdef NOTYET 149 if (td->td_flags & TDF_UPCALLING) 150 tdf->td_flags &= ~TDF_UNBOUND; 151 #endif 152 } 153 kg->kg_runnable--; 154 CTR2(KTR_RUNQ, "choosethread: td=%p pri=%d", 155 td, td->td_priority); 156 } else { 157 /* Simulate runq_choose() having returned the idle thread */ 158 td = PCPU_GET(idlethread); 159 CTR1(KTR_RUNQ, "choosethread: td=%p (idle)", td); 160 } 161 if (panicstr && ((td->td_proc->p_flag & P_SYSTEM) == 0 && 162 (td->td_flags & TDF_INPANIC) == 0)) 163 goto retry; 164 td->td_state = TDS_RUNNING; 165 return (td); 166 } 167 168 /* 169 * Given a KSE (now surplus), either assign a new runable thread to it 170 * (and put it in the run queue) or put it in the ksegrp's idle KSE list. 171 * Assumes the kse is not linked to any threads any more. (has been cleaned). 172 */ 173 void 174 kse_reassign(struct kse *ke) 175 { 176 struct ksegrp *kg; 177 struct thread *td; 178 179 kg = ke->ke_ksegrp; 180 181 /* 182 * Find the first unassigned thread 183 * If there is a 'last assigned' then see what's next. 184 * otherwise look at what is first. 185 */ 186 if ((td = kg->kg_last_assigned)) { 187 td = TAILQ_NEXT(td, td_runq); 188 } else { 189 td = TAILQ_FIRST(&kg->kg_runq); 190 } 191 192 /* 193 * If we found one assign it the kse, otherwise idle the kse. 194 */ 195 if (td) { 196 kg->kg_last_assigned = td; 197 td->td_kse = ke; 198 ke->ke_thread = td; 199 runq_add(&runq, ke); 200 CTR2(KTR_RUNQ, "kse_reassign: ke%p -> td%p", ke, td); 201 } else { 202 ke->ke_state = KES_IDLE; 203 ke->ke_thread = NULL; 204 TAILQ_INSERT_HEAD(&kg->kg_iq, ke, ke_kgrlist); 205 kg->kg_idle_kses++; 206 CTR1(KTR_RUNQ, "kse_reassign: ke%p idled", ke); 207 } 208 } 209 210 int 211 kserunnable(void) 212 { 213 return runq_check(&runq); 214 } 215 216 /* 217 * Remove a thread from its KSEGRP's run queue. 218 * This in turn may remove it from a KSE if it was already assigned 219 * to one, possibly causing a new thread to be assigned to the KSE 220 * and the KSE getting a new priority (unless it's a BOUND thread/KSE pair). 221 */ 222 void 223 remrunqueue(struct thread *td) 224 { 225 struct thread *td2, *td3; 226 struct ksegrp *kg; 227 struct kse *ke; 228 229 mtx_assert(&sched_lock, MA_OWNED); 230 KASSERT ((td->td_state == TDS_RUNQ), 231 ("remrunqueue: Bad state on run queue")); 232 kg = td->td_ksegrp; 233 ke = td->td_kse; 234 /* 235 * If it's a bound thread/KSE pair, take the shortcut. All non-KSE 236 * threads are BOUND. 237 */ 238 CTR1(KTR_RUNQ, "remrunqueue: td%p", td); 239 td->td_state = TDS_UNQUEUED; 240 kg->kg_runnable--; 241 if ((td->td_flags & TDF_UNBOUND) == 0) { 242 /* Bring its kse with it, leave the thread attached */ 243 runq_remove(&runq, ke); 244 ke->ke_state = KES_THREAD; 245 return; 246 } 247 if (ke) { 248 /* 249 * This thread has been assigned to a KSE. 250 * We need to dissociate it and try assign the 251 * KSE to the next available thread. Then, we should 252 * see if we need to move the KSE in the run queues. 253 */ 254 td2 = kg->kg_last_assigned; 255 KASSERT((td2 != NULL), ("last assigned has wrong value ")); 256 td->td_kse = NULL; 257 if ((td3 = TAILQ_NEXT(td2, td_runq))) { 258 KASSERT(td3 != td, ("td3 somehow matched td")); 259 /* 260 * Give the next unassigned thread to the KSE 261 * so the number of runnable KSEs remains 262 * constant. 263 */ 264 td3->td_kse = ke; 265 ke->ke_thread = td3; 266 kg->kg_last_assigned = td3; 267 runq_readjust(&runq, ke); 268 } else { 269 /* 270 * There is no unassigned thread. 271 * If we were the last assigned one, 272 * adjust the last assigned pointer back 273 * one, which may result in NULL. 274 */ 275 if (td == td2) { 276 kg->kg_last_assigned = 277 TAILQ_PREV(td, threadqueue, td_runq); 278 } 279 runq_remove(&runq, ke); 280 KASSERT((ke->ke_state != KES_IDLE), 281 ("kse already idle")); 282 ke->ke_state = KES_IDLE; 283 ke->ke_thread = NULL; 284 TAILQ_INSERT_HEAD(&kg->kg_iq, ke, ke_kgrlist); 285 kg->kg_idle_kses++; 286 } 287 } 288 TAILQ_REMOVE(&kg->kg_runq, td, td_runq); 289 } 290 291 void 292 setrunqueue(struct thread *td) 293 { 294 struct kse *ke; 295 struct ksegrp *kg; 296 struct thread *td2; 297 struct thread *tda; 298 299 CTR1(KTR_RUNQ, "setrunqueue: td%p", td); 300 mtx_assert(&sched_lock, MA_OWNED); 301 KASSERT((td->td_state != TDS_RUNQ), ("setrunqueue: bad thread state")); 302 td->td_state = TDS_RUNQ; 303 kg = td->td_ksegrp; 304 kg->kg_runnable++; 305 if ((td->td_flags & TDF_UNBOUND) == 0) { 306 KASSERT((td->td_kse != NULL), 307 ("queueing BAD thread to run queue")); 308 /* 309 * Common path optimisation: Only one of everything 310 * and the KSE is always already attached. 311 * Totally ignore the ksegrp run queue. 312 */ 313 runq_add(&runq, td->td_kse); 314 return; 315 } 316 /* 317 * Ok, so we are threading with this thread. 318 * We don't have a KSE, see if we can get one.. 319 */ 320 tda = kg->kg_last_assigned; 321 if ((ke = td->td_kse) == NULL) { 322 /* 323 * We will need a KSE, see if there is one.. 324 * First look for a free one, before getting desperate. 325 * If we can't get one, our priority is not high enough.. 326 * that's ok.. 327 */ 328 if (kg->kg_idle_kses) { 329 /* 330 * There is a free one so it's ours for the asking.. 331 */ 332 ke = TAILQ_FIRST(&kg->kg_iq); 333 TAILQ_REMOVE(&kg->kg_iq, ke, ke_kgrlist); 334 ke->ke_state = KES_THREAD; 335 kg->kg_idle_kses--; 336 } else if (tda && (tda->td_priority > td->td_priority)) { 337 /* 338 * None free, but there is one we can commandeer. 339 */ 340 ke = tda->td_kse; 341 tda->td_kse = NULL; 342 ke->ke_thread = NULL; 343 tda = kg->kg_last_assigned = 344 TAILQ_PREV(tda, threadqueue, td_runq); 345 runq_remove(&runq, ke); 346 } 347 } else { 348 /* 349 * Temporarily disassociate so it looks like the other cases. 350 */ 351 ke->ke_thread = NULL; 352 td->td_kse = NULL; 353 } 354 355 /* 356 * Add the thread to the ksegrp's run queue at 357 * the appropriate place. 358 */ 359 TAILQ_FOREACH(td2, &kg->kg_runq, td_runq) { 360 if (td2->td_priority > td->td_priority) { 361 TAILQ_INSERT_BEFORE(td2, td, td_runq); 362 break; 363 } 364 } 365 if (td2 == NULL) { 366 /* We ran off the end of the TAILQ or it was empty. */ 367 TAILQ_INSERT_TAIL(&kg->kg_runq, td, td_runq); 368 } 369 370 /* 371 * If we have a ke to use, then put it on the run queue and 372 * If needed, readjust the last_assigned pointer. 373 */ 374 if (ke) { 375 if (tda == NULL) { 376 /* 377 * No pre-existing last assigned so whoever is first 378 * gets the KSE we brought in.. (maybe us) 379 */ 380 td2 = TAILQ_FIRST(&kg->kg_runq); 381 KASSERT((td2->td_kse == NULL), 382 ("unexpected ke present")); 383 td2->td_kse = ke; 384 ke->ke_thread = td2; 385 kg->kg_last_assigned = td2; 386 } else if (tda->td_priority > td->td_priority) { 387 /* 388 * It's ours, grab it, but last_assigned is past us 389 * so don't change it. 390 */ 391 td->td_kse = ke; 392 ke->ke_thread = td; 393 } else { 394 /* 395 * We are past last_assigned, so 396 * put the new kse on whatever is next, 397 * which may or may not be us. 398 */ 399 td2 = TAILQ_NEXT(tda, td_runq); 400 kg->kg_last_assigned = td2; 401 td2->td_kse = ke; 402 ke->ke_thread = td2; 403 } 404 runq_add(&runq, ke); 405 } 406 } 407 408 /************************************************************************ 409 * Critical section marker functions * 410 ************************************************************************/ 411 /* Critical sections that prevent preemption. */ 412 void 413 critical_enter(void) 414 { 415 struct thread *td; 416 417 td = curthread; 418 if (td->td_critnest == 0) 419 cpu_critical_enter(); 420 td->td_critnest++; 421 } 422 423 void 424 critical_exit(void) 425 { 426 struct thread *td; 427 428 td = curthread; 429 if (td->td_critnest == 1) { 430 td->td_critnest = 0; 431 cpu_critical_exit(); 432 } else { 433 td->td_critnest--; 434 } 435 } 436 437 438 /************************************************************************ 439 * SYSTEM RUN QUEUE manipulations and tests * 440 ************************************************************************/ 441 /* 442 * Initialize a run structure. 443 */ 444 void 445 runq_init(struct runq *rq) 446 { 447 int i; 448 449 bzero(rq, sizeof *rq); 450 for (i = 0; i < RQ_NQS; i++) 451 TAILQ_INIT(&rq->rq_queues[i]); 452 } 453 454 /* 455 * Clear the status bit of the queue corresponding to priority level pri, 456 * indicating that it is empty. 457 */ 458 static __inline void 459 runq_clrbit(struct runq *rq, int pri) 460 { 461 struct rqbits *rqb; 462 463 rqb = &rq->rq_status; 464 CTR4(KTR_RUNQ, "runq_clrbit: bits=%#x %#x bit=%#x word=%d", 465 rqb->rqb_bits[RQB_WORD(pri)], 466 rqb->rqb_bits[RQB_WORD(pri)] & ~RQB_BIT(pri), 467 RQB_BIT(pri), RQB_WORD(pri)); 468 rqb->rqb_bits[RQB_WORD(pri)] &= ~RQB_BIT(pri); 469 } 470 471 /* 472 * Find the index of the first non-empty run queue. This is done by 473 * scanning the status bits, a set bit indicates a non-empty queue. 474 */ 475 static __inline int 476 runq_findbit(struct runq *rq) 477 { 478 struct rqbits *rqb; 479 int pri; 480 int i; 481 482 rqb = &rq->rq_status; 483 for (i = 0; i < RQB_LEN; i++) 484 if (rqb->rqb_bits[i]) { 485 pri = RQB_FFS(rqb->rqb_bits[i]) + (i << RQB_L2BPW); 486 CTR3(KTR_RUNQ, "runq_findbit: bits=%#x i=%d pri=%d", 487 rqb->rqb_bits[i], i, pri); 488 return (pri); 489 } 490 491 return (-1); 492 } 493 494 /* 495 * Set the status bit of the queue corresponding to priority level pri, 496 * indicating that it is non-empty. 497 */ 498 static __inline void 499 runq_setbit(struct runq *rq, int pri) 500 { 501 struct rqbits *rqb; 502 503 rqb = &rq->rq_status; 504 CTR4(KTR_RUNQ, "runq_setbit: bits=%#x %#x bit=%#x word=%d", 505 rqb->rqb_bits[RQB_WORD(pri)], 506 rqb->rqb_bits[RQB_WORD(pri)] | RQB_BIT(pri), 507 RQB_BIT(pri), RQB_WORD(pri)); 508 rqb->rqb_bits[RQB_WORD(pri)] |= RQB_BIT(pri); 509 } 510 511 /* 512 * Add the KSE to the queue specified by its priority, and set the 513 * corresponding status bit. 514 */ 515 void 516 runq_add(struct runq *rq, struct kse *ke) 517 { 518 struct rqhead *rqh; 519 int pri; 520 521 mtx_assert(&sched_lock, MA_OWNED); 522 KASSERT((ke->ke_thread != NULL), ("runq_add: No thread on KSE")); 523 KASSERT((ke->ke_thread->td_kse != NULL), 524 ("runq_add: No KSE on thread")); 525 KASSERT(ke->ke_state != KES_ONRUNQ, 526 ("runq_add: kse %p (%s) already in run queue", ke, 527 ke->ke_proc->p_comm)); 528 pri = ke->ke_thread->td_priority / RQ_PPQ; 529 ke->ke_rqindex = pri; 530 runq_setbit(rq, pri); 531 rqh = &rq->rq_queues[pri]; 532 CTR4(KTR_RUNQ, "runq_add: p=%p pri=%d %d rqh=%p", 533 ke->ke_proc, ke->ke_thread->td_priority, pri, rqh); 534 TAILQ_INSERT_TAIL(rqh, ke, ke_procq); 535 ke->ke_ksegrp->kg_runq_kses++; 536 ke->ke_state = KES_ONRUNQ; 537 } 538 539 /* 540 * Return true if there are runnable processes of any priority on the run 541 * queue, false otherwise. Has no side effects, does not modify the run 542 * queue structure. 543 */ 544 int 545 runq_check(struct runq *rq) 546 { 547 struct rqbits *rqb; 548 int i; 549 550 rqb = &rq->rq_status; 551 for (i = 0; i < RQB_LEN; i++) 552 if (rqb->rqb_bits[i]) { 553 CTR2(KTR_RUNQ, "runq_check: bits=%#x i=%d", 554 rqb->rqb_bits[i], i); 555 return (1); 556 } 557 CTR0(KTR_RUNQ, "runq_check: empty"); 558 559 return (0); 560 } 561 562 /* 563 * Find and remove the highest priority process from the run queue. 564 * If there are no runnable processes, the per-cpu idle process is 565 * returned. Will not return NULL under any circumstances. 566 */ 567 struct kse * 568 runq_choose(struct runq *rq) 569 { 570 struct rqhead *rqh; 571 struct kse *ke; 572 int pri; 573 574 mtx_assert(&sched_lock, MA_OWNED); 575 while ((pri = runq_findbit(rq)) != -1) { 576 rqh = &rq->rq_queues[pri]; 577 ke = TAILQ_FIRST(rqh); 578 KASSERT(ke != NULL, ("runq_choose: no proc on busy queue")); 579 CTR3(KTR_RUNQ, 580 "runq_choose: pri=%d kse=%p rqh=%p", pri, ke, rqh); 581 TAILQ_REMOVE(rqh, ke, ke_procq); 582 ke->ke_ksegrp->kg_runq_kses--; 583 if (TAILQ_EMPTY(rqh)) { 584 CTR0(KTR_RUNQ, "runq_choose: empty"); 585 runq_clrbit(rq, pri); 586 } 587 588 ke->ke_state = KES_THREAD; 589 KASSERT((ke->ke_thread != NULL), 590 ("runq_choose: No thread on KSE")); 591 KASSERT((ke->ke_thread->td_kse != NULL), 592 ("runq_choose: No KSE on thread")); 593 return (ke); 594 } 595 CTR1(KTR_RUNQ, "runq_choose: idleproc pri=%d", pri); 596 597 return (NULL); 598 } 599 600 /* 601 * Remove the KSE from the queue specified by its priority, and clear the 602 * corresponding status bit if the queue becomes empty. 603 * Caller must set ke->ke_state afterwards. 604 */ 605 void 606 runq_remove(struct runq *rq, struct kse *ke) 607 { 608 struct rqhead *rqh; 609 int pri; 610 611 KASSERT((ke->ke_state == KES_ONRUNQ), ("KSE not on run queue")); 612 mtx_assert(&sched_lock, MA_OWNED); 613 pri = ke->ke_rqindex; 614 rqh = &rq->rq_queues[pri]; 615 CTR4(KTR_RUNQ, "runq_remove: p=%p pri=%d %d rqh=%p", 616 ke, ke->ke_thread->td_priority, pri, rqh); 617 KASSERT(ke != NULL, ("runq_remove: no proc on busy queue")); 618 TAILQ_REMOVE(rqh, ke, ke_procq); 619 if (TAILQ_EMPTY(rqh)) { 620 CTR0(KTR_RUNQ, "runq_remove: empty"); 621 runq_clrbit(rq, pri); 622 } 623 ke->ke_state = KES_THREAD; 624 ke->ke_ksegrp->kg_runq_kses--; 625 } 626 627 static void 628 runq_readjust(struct runq *rq, struct kse *ke) 629 { 630 631 if (ke->ke_rqindex != (ke->ke_thread->td_priority / RQ_PPQ)) { 632 runq_remove(rq, ke); 633 runq_add(rq, ke); 634 } 635 } 636 637 #if 0 638 void 639 thread_sanity_check(struct thread *td) 640 { 641 struct proc *p; 642 struct ksegrp *kg; 643 struct kse *ke; 644 struct thread *td2; 645 unsigned int prevpri; 646 int saw_lastassigned; 647 int unassigned; 648 int assigned; 649 650 p = td->td_proc; 651 kg = td->td_ksegrp; 652 ke = td->td_kse; 653 654 if (kg != &p->p_ksegrp) { 655 panic ("wrong ksegrp"); 656 } 657 658 if (ke) { 659 if (ke != &p->p_kse) { 660 panic("wrong kse"); 661 } 662 if (ke->ke_thread != td) { 663 panic("wrong thread"); 664 } 665 } 666 667 if ((p->p_flag & P_KSES) == 0) { 668 if (ke == NULL) { 669 panic("non KSE thread lost kse"); 670 } 671 } else { 672 prevpri = 0; 673 saw_lastassigned = 0; 674 unassigned = 0; 675 assigned = 0; 676 TAILQ_FOREACH(td2, &kg->kg_runq, td_runq) { 677 if (td2->td_priority < prevpri) { 678 panic("thread runqueue unosorted"); 679 } 680 prevpri = td2->td_priority; 681 if (td2->td_kse) { 682 assigned++; 683 if (unassigned) { 684 panic("unassigned before assigned"); 685 } 686 if (kg->kg_last_assigned == NULL) { 687 panic("lastassigned corrupt"); 688 } 689 if (saw_lastassigned) { 690 panic("last assigned not last"); 691 } 692 if (td2->td_kse->ke_thread != td2) { 693 panic("mismatched kse/thread"); 694 } 695 } else { 696 unassigned++; 697 } 698 if (td2 == kg->kg_last_assigned) { 699 saw_lastassigned = 1; 700 if (td2->td_kse == NULL) { 701 panic("last assigned not assigned"); 702 } 703 } 704 } 705 if (kg->kg_last_assigned && (saw_lastassigned == 0)) { 706 panic("where on earth does lastassigned point?"); 707 } 708 FOREACH_THREAD_IN_GROUP(kg, td2) { 709 if (((td2->td_flags & TDF_UNBOUND) == 0) && 710 (td2->td_state == TDS_RUNQ)) { 711 assigned++; 712 if (td2->td_kse == NULL) { 713 panic ("BOUND thread with no KSE"); 714 } 715 } 716 } 717 #if 0 718 if ((unassigned + assigned) != kg->kg_runnable) { 719 panic("wrong number in runnable"); 720 } 721 #endif 722 } 723 } 724 #endif 725 726