1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* Copyright (c) 1984, 1986, 1987, 1988, 1989 AT&T */ 23 /* All Rights Reserved */ 24 25 /* 26 * Copyright 2009 Sun Microsystems, Inc. All rights reserved. 27 * Use is subject to license terms. 28 */ 29 30 #include <sys/param.h> 31 #include <sys/types.h> 32 #include <sys/vmparam.h> 33 #include <sys/systm.h> 34 #include <sys/stack.h> 35 #include <sys/frame.h> 36 #include <sys/proc.h> 37 #include <sys/ucontext.h> 38 #include <sys/cpuvar.h> 39 #include <sys/asm_linkage.h> 40 #include <sys/kmem.h> 41 #include <sys/errno.h> 42 #include <sys/bootconf.h> 43 #include <sys/archsystm.h> 44 #include <sys/fpu/fpusystm.h> 45 #include <sys/debug.h> 46 #include <sys/privregs.h> 47 #include <sys/machpcb.h> 48 #include <sys/psr_compat.h> 49 #include <sys/cmn_err.h> 50 #include <sys/asi.h> 51 #include <sys/copyops.h> 52 #include <sys/model.h> 53 #include <sys/panic.h> 54 #include <sys/exec.h> 55 56 /* 57 * By default, set the weakest model to TSO (Total Store Order) 58 * which is the default memory model on SPARC. 59 * If a platform does support a weaker model than TSO, this will be 60 * updated at runtime to reflect that. 61 */ 62 uint_t weakest_mem_model = TSTATE_MM_TSO; 63 64 /* 65 * modify the lower 32bits of a uint64_t 66 */ 67 #define SET_LOWER_32(all, lower) \ 68 (((uint64_t)(all) & 0xffffffff00000000) | (uint32_t)(lower)) 69 70 #define MEMCPY_FPU_EN 2 /* fprs on and fpu_en == 0 */ 71 72 static uint_t mkpsr(uint64_t tstate, uint32_t fprs); 73 74 #ifdef _SYSCALL32_IMPL 75 static void fpuregset_32ton(const fpregset32_t *src, fpregset_t *dest, 76 const struct fq32 *sfq, struct fq *dfq); 77 #endif /* _SYSCALL32_IMPL */ 78 79 /* 80 * Set floating-point registers. 81 * NOTE: 'lwp' might not correspond to 'curthread' since this is 82 * called from code in /proc to set the registers of another lwp. 83 */ 84 void 85 setfpregs(klwp_t *lwp, fpregset_t *fp) 86 { 87 struct machpcb *mpcb; 88 kfpu_t *pfp; 89 uint32_t fprs = (FPRS_FEF|FPRS_DU|FPRS_DL); 90 model_t model = lwp_getdatamodel(lwp); 91 92 mpcb = lwptompcb(lwp); 93 pfp = lwptofpu(lwp); 94 95 /* 96 * This is always true for both "real" fp programs and memcpy fp 97 * programs, because we force fpu_en to MEMCPY_FPU_EN in getfpregs, 98 * for the memcpy and threads cases where (fpu_en == 0) && 99 * (fpu_fprs & FPRS_FEF), if setfpregs is called after getfpregs. 100 */ 101 if (fp->fpu_en) { 102 kpreempt_disable(); 103 104 if (!(pfp->fpu_en) && (!(pfp->fpu_fprs & FPRS_FEF)) && 105 fpu_exists) { 106 /* 107 * He's not currently using the FPU but wants to in his 108 * new context - arrange for this on return to userland. 109 */ 110 pfp->fpu_fprs = (uint32_t)fprs; 111 } 112 /* 113 * Get setfpregs to restore fpu_en to zero 114 * for the memcpy/threads case (where pfp->fpu_en == 0 && 115 * (pfp->fp_fprs & FPRS_FEF) == FPRS_FEF). 116 */ 117 if (fp->fpu_en == MEMCPY_FPU_EN) 118 fp->fpu_en = 0; 119 120 /* 121 * Load up a user's floating point context. 122 */ 123 if (fp->fpu_qcnt > MAXFPQ) /* plug security holes */ 124 fp->fpu_qcnt = MAXFPQ; 125 fp->fpu_q_entrysize = sizeof (struct fq); 126 127 /* 128 * For v9 kernel, copy all of the fp regs. 129 * For v8 kernel, copy v8 fp regs (lower half of v9 fp regs). 130 * Restore entire fsr for v9, only lower half for v8. 131 */ 132 (void) kcopy(fp, pfp, sizeof (fp->fpu_fr)); 133 if (model == DATAMODEL_LP64) 134 pfp->fpu_fsr = fp->fpu_fsr; 135 else 136 pfp->fpu_fsr = SET_LOWER_32(pfp->fpu_fsr, fp->fpu_fsr); 137 pfp->fpu_qcnt = fp->fpu_qcnt; 138 pfp->fpu_q_entrysize = fp->fpu_q_entrysize; 139 pfp->fpu_en = fp->fpu_en; 140 pfp->fpu_q = mpcb->mpcb_fpu_q; 141 if (fp->fpu_qcnt) 142 (void) kcopy(fp->fpu_q, pfp->fpu_q, 143 fp->fpu_qcnt * fp->fpu_q_entrysize); 144 /* FSR ignores these bits on load, so they can not be set */ 145 pfp->fpu_fsr &= ~(FSR_QNE|FSR_FTT); 146 147 /* 148 * If not the current process then resume() will handle it. 149 */ 150 if (lwp != ttolwp(curthread)) { 151 /* force resume to reload fp regs */ 152 pfp->fpu_fprs |= FPRS_FEF; 153 kpreempt_enable(); 154 return; 155 } 156 157 /* 158 * Load up FPU with new floating point context. 159 */ 160 if (fpu_exists) { 161 pfp->fpu_fprs = _fp_read_fprs(); 162 if ((pfp->fpu_fprs & FPRS_FEF) != FPRS_FEF) { 163 _fp_write_fprs(fprs); 164 pfp->fpu_fprs = (uint32_t)fprs; 165 #ifdef DEBUG 166 if (fpdispr) 167 cmn_err(CE_NOTE, 168 "setfpregs with fp disabled!\n"); 169 #endif 170 } 171 /* 172 * Load all fp regs for v9 user programs, but only 173 * load the lower half for v8[plus] programs. 174 */ 175 if (model == DATAMODEL_LP64) 176 fp_restore(pfp); 177 else 178 fp_v8_load(pfp); 179 } 180 181 kpreempt_enable(); 182 } else { 183 if ((pfp->fpu_en) || /* normal fp case */ 184 (pfp->fpu_fprs & FPRS_FEF)) { /* memcpy/threads case */ 185 /* 186 * Currently the lwp has floating point enabled. 187 * Turn off FPRS_FEF in user's fprs, saved and 188 * real copies thereof. 189 */ 190 pfp->fpu_en = 0; 191 if (fpu_exists) { 192 fprs = 0; 193 if (lwp == ttolwp(curthread)) 194 _fp_write_fprs(fprs); 195 pfp->fpu_fprs = (uint32_t)fprs; 196 } 197 } 198 } 199 } 200 201 #ifdef _SYSCALL32_IMPL 202 void 203 setfpregs32(klwp_t *lwp, fpregset32_t *fp) 204 { 205 fpregset_t fpregs; 206 207 fpuregset_32ton(fp, &fpregs, NULL, NULL); 208 setfpregs(lwp, &fpregs); 209 } 210 #endif /* _SYSCALL32_IMPL */ 211 212 /* 213 * NOTE: 'lwp' might not correspond to 'curthread' since this is 214 * called from code in /proc to set the registers of another lwp. 215 */ 216 void 217 run_fpq(klwp_t *lwp, fpregset_t *fp) 218 { 219 /* 220 * If the context being loaded up includes a floating queue, 221 * we need to simulate those instructions (since we can't reload 222 * the fpu) and pass the process any appropriate signals 223 */ 224 225 if (lwp == ttolwp(curthread)) { 226 if (fpu_exists) { 227 if (fp->fpu_qcnt) 228 fp_runq(lwp->lwp_regs); 229 } 230 } 231 } 232 233 /* 234 * Get floating-point registers. 235 * NOTE: 'lwp' might not correspond to 'curthread' since this is 236 * called from code in /proc to set the registers of another lwp. 237 */ 238 void 239 getfpregs(klwp_t *lwp, fpregset_t *fp) 240 { 241 kfpu_t *pfp; 242 model_t model = lwp_getdatamodel(lwp); 243 244 pfp = lwptofpu(lwp); 245 kpreempt_disable(); 246 if (fpu_exists && ttolwp(curthread) == lwp) 247 pfp->fpu_fprs = _fp_read_fprs(); 248 249 /* 250 * First check the fpu_en case, for normal fp programs. 251 * Next check the fprs case, for fp use by memcpy/threads. 252 */ 253 if (((fp->fpu_en = pfp->fpu_en) != 0) || 254 (pfp->fpu_fprs & FPRS_FEF)) { 255 /* 256 * Force setfpregs to restore the fp context in 257 * setfpregs for the memcpy and threads cases (where 258 * pfp->fpu_en == 0 && (pfp->fp_fprs & FPRS_FEF) == FPRS_FEF). 259 */ 260 if (pfp->fpu_en == 0) 261 fp->fpu_en = MEMCPY_FPU_EN; 262 /* 263 * If we have an fpu and the current thread owns the fp 264 * context, flush fp * registers into the pcb. Save all 265 * the fp regs for v9, xregs_getfpregs saves the upper half 266 * for v8plus. Save entire fsr for v9, only lower half for v8. 267 */ 268 if (fpu_exists && ttolwp(curthread) == lwp) { 269 if ((pfp->fpu_fprs & FPRS_FEF) != FPRS_FEF) { 270 uint32_t fprs = (FPRS_FEF|FPRS_DU|FPRS_DL); 271 272 _fp_write_fprs(fprs); 273 pfp->fpu_fprs = fprs; 274 #ifdef DEBUG 275 if (fpdispr) 276 cmn_err(CE_NOTE, 277 "getfpregs with fp disabled!\n"); 278 #endif 279 } 280 if (model == DATAMODEL_LP64) 281 fp_fksave(pfp); 282 else 283 fp_v8_fksave(pfp); 284 } 285 (void) kcopy(pfp, fp, sizeof (fp->fpu_fr)); 286 fp->fpu_q = pfp->fpu_q; 287 if (model == DATAMODEL_LP64) 288 fp->fpu_fsr = pfp->fpu_fsr; 289 else 290 fp->fpu_fsr = (uint32_t)pfp->fpu_fsr; 291 fp->fpu_qcnt = pfp->fpu_qcnt; 292 fp->fpu_q_entrysize = pfp->fpu_q_entrysize; 293 } else { 294 int i; 295 for (i = 0; i < 32; i++) /* NaN */ 296 ((uint32_t *)fp->fpu_fr.fpu_regs)[i] = (uint32_t)-1; 297 if (model == DATAMODEL_LP64) { 298 for (i = 16; i < 32; i++) /* NaN */ 299 ((uint64_t *)fp->fpu_fr.fpu_dregs)[i] = 300 (uint64_t)-1; 301 } 302 fp->fpu_fsr = 0; 303 fp->fpu_qcnt = 0; 304 } 305 kpreempt_enable(); 306 } 307 308 #ifdef _SYSCALL32_IMPL 309 void 310 getfpregs32(klwp_t *lwp, fpregset32_t *fp) 311 { 312 fpregset_t fpregs; 313 314 getfpregs(lwp, &fpregs); 315 fpuregset_nto32(&fpregs, fp, NULL); 316 } 317 #endif /* _SYSCALL32_IMPL */ 318 319 /* 320 * Set general registers. 321 * NOTE: 'lwp' might not correspond to 'curthread' since this is 322 * called from code in /proc to set the registers of another lwp. 323 */ 324 325 /* 64-bit gregset_t */ 326 void 327 setgregs(klwp_t *lwp, gregset_t grp) 328 { 329 struct regs *rp = lwptoregs(lwp); 330 kfpu_t *fp = lwptofpu(lwp); 331 uint64_t tbits; 332 333 int current = (lwp == curthread->t_lwp); 334 335 if (current) 336 (void) save_syscall_args(); /* copy the args first */ 337 338 tbits = (((grp[REG_CCR] & TSTATE_CCR_MASK) << TSTATE_CCR_SHIFT) | 339 ((grp[REG_ASI] & TSTATE_ASI_MASK) << TSTATE_ASI_SHIFT)); 340 rp->r_tstate &= ~(((uint64_t)TSTATE_CCR_MASK << TSTATE_CCR_SHIFT) | 341 ((uint64_t)TSTATE_ASI_MASK << TSTATE_ASI_SHIFT)); 342 rp->r_tstate |= tbits; 343 kpreempt_disable(); 344 fp->fpu_fprs = (uint32_t)grp[REG_FPRS]; 345 if (fpu_exists && (current) && (fp->fpu_fprs & FPRS_FEF)) 346 _fp_write_fprs(fp->fpu_fprs); 347 kpreempt_enable(); 348 349 /* 350 * pc and npc must be 4-byte aligned on sparc. 351 * We silently make it so to avoid a watchdog reset. 352 */ 353 rp->r_pc = grp[REG_PC] & ~03L; 354 rp->r_npc = grp[REG_nPC] & ~03L; 355 rp->r_y = grp[REG_Y]; 356 357 rp->r_g1 = grp[REG_G1]; 358 rp->r_g2 = grp[REG_G2]; 359 rp->r_g3 = grp[REG_G3]; 360 rp->r_g4 = grp[REG_G4]; 361 rp->r_g5 = grp[REG_G5]; 362 rp->r_g6 = grp[REG_G6]; 363 rp->r_g7 = grp[REG_G7]; 364 365 rp->r_o0 = grp[REG_O0]; 366 rp->r_o1 = grp[REG_O1]; 367 rp->r_o2 = grp[REG_O2]; 368 rp->r_o3 = grp[REG_O3]; 369 rp->r_o4 = grp[REG_O4]; 370 rp->r_o5 = grp[REG_O5]; 371 rp->r_o6 = grp[REG_O6]; 372 rp->r_o7 = grp[REG_O7]; 373 374 if (current) { 375 /* 376 * This was called from a system call, but we 377 * do not want to return via the shared window; 378 * restoring the CPU context changes everything. 379 */ 380 lwp->lwp_eosys = JUSTRETURN; 381 curthread->t_post_sys = 1; 382 } 383 } 384 385 /* 386 * Return the general registers. 387 * NOTE: 'lwp' might not correspond to 'curthread' since this is 388 * called from code in /proc to get the registers of another lwp. 389 */ 390 void 391 getgregs(klwp_t *lwp, gregset_t grp) 392 { 393 struct regs *rp = lwptoregs(lwp); 394 uint32_t fprs; 395 396 kpreempt_disable(); 397 if (fpu_exists && ttolwp(curthread) == lwp) { 398 fprs = _fp_read_fprs(); 399 } else { 400 kfpu_t *fp = lwptofpu(lwp); 401 fprs = fp->fpu_fprs; 402 } 403 kpreempt_enable(); 404 grp[REG_CCR] = (rp->r_tstate >> TSTATE_CCR_SHIFT) & TSTATE_CCR_MASK; 405 grp[REG_PC] = rp->r_pc; 406 grp[REG_nPC] = rp->r_npc; 407 grp[REG_Y] = (uint32_t)rp->r_y; 408 grp[REG_G1] = rp->r_g1; 409 grp[REG_G2] = rp->r_g2; 410 grp[REG_G3] = rp->r_g3; 411 grp[REG_G4] = rp->r_g4; 412 grp[REG_G5] = rp->r_g5; 413 grp[REG_G6] = rp->r_g6; 414 grp[REG_G7] = rp->r_g7; 415 grp[REG_O0] = rp->r_o0; 416 grp[REG_O1] = rp->r_o1; 417 grp[REG_O2] = rp->r_o2; 418 grp[REG_O3] = rp->r_o3; 419 grp[REG_O4] = rp->r_o4; 420 grp[REG_O5] = rp->r_o5; 421 grp[REG_O6] = rp->r_o6; 422 grp[REG_O7] = rp->r_o7; 423 grp[REG_ASI] = (rp->r_tstate >> TSTATE_ASI_SHIFT) & TSTATE_ASI_MASK; 424 grp[REG_FPRS] = fprs; 425 } 426 427 void 428 getgregs32(klwp_t *lwp, gregset32_t grp) 429 { 430 struct regs *rp = lwptoregs(lwp); 431 uint32_t fprs; 432 433 kpreempt_disable(); 434 if (fpu_exists && ttolwp(curthread) == lwp) { 435 fprs = _fp_read_fprs(); 436 } else { 437 kfpu_t *fp = lwptofpu(lwp); 438 fprs = fp->fpu_fprs; 439 } 440 kpreempt_enable(); 441 grp[REG_PSR] = mkpsr(rp->r_tstate, fprs); 442 grp[REG_PC] = rp->r_pc; 443 grp[REG_nPC] = rp->r_npc; 444 grp[REG_Y] = rp->r_y; 445 grp[REG_G1] = rp->r_g1; 446 grp[REG_G2] = rp->r_g2; 447 grp[REG_G3] = rp->r_g3; 448 grp[REG_G4] = rp->r_g4; 449 grp[REG_G5] = rp->r_g5; 450 grp[REG_G6] = rp->r_g6; 451 grp[REG_G7] = rp->r_g7; 452 grp[REG_O0] = rp->r_o0; 453 grp[REG_O1] = rp->r_o1; 454 grp[REG_O2] = rp->r_o2; 455 grp[REG_O3] = rp->r_o3; 456 grp[REG_O4] = rp->r_o4; 457 grp[REG_O5] = rp->r_o5; 458 grp[REG_O6] = rp->r_o6; 459 grp[REG_O7] = rp->r_o7; 460 } 461 462 /* 463 * Return the user-level PC. 464 * If in a system call, return the address of the syscall trap. 465 */ 466 greg_t 467 getuserpc() 468 { 469 return (lwptoregs(ttolwp(curthread))->r_pc); 470 } 471 472 /* 473 * Set register windows. 474 */ 475 void 476 setgwins(klwp_t *lwp, gwindows_t *gwins) 477 { 478 struct machpcb *mpcb = lwptompcb(lwp); 479 int wbcnt = gwins->wbcnt; 480 caddr_t sp; 481 int i; 482 struct rwindow32 *rwp; 483 int wbuf_rwindow_size; 484 int is64; 485 486 if (mpcb->mpcb_wstate == WSTATE_USER32) { 487 wbuf_rwindow_size = WINDOWSIZE32; 488 is64 = 0; 489 } else { 490 wbuf_rwindow_size = WINDOWSIZE64; 491 is64 = 1; 492 } 493 ASSERT(wbcnt >= 0 && wbcnt <= SPARC_MAXREGWINDOW); 494 mpcb->mpcb_wbcnt = 0; 495 for (i = 0; i < wbcnt; i++) { 496 sp = (caddr_t)gwins->spbuf[i]; 497 mpcb->mpcb_spbuf[i] = sp; 498 rwp = (struct rwindow32 *) 499 (mpcb->mpcb_wbuf + (i * wbuf_rwindow_size)); 500 if (is64 && IS_V9STACK(sp)) 501 bcopy(&gwins->wbuf[i], rwp, sizeof (struct rwindow)); 502 else 503 rwindow_nto32(&gwins->wbuf[i], rwp); 504 mpcb->mpcb_wbcnt++; 505 } 506 } 507 508 void 509 setgwins32(klwp_t *lwp, gwindows32_t *gwins) 510 { 511 struct machpcb *mpcb = lwptompcb(lwp); 512 int wbcnt = gwins->wbcnt; 513 caddr_t sp; 514 int i; 515 516 struct rwindow *rwp; 517 int wbuf_rwindow_size; 518 int is64; 519 520 if (mpcb->mpcb_wstate == WSTATE_USER32) { 521 wbuf_rwindow_size = WINDOWSIZE32; 522 is64 = 0; 523 } else { 524 wbuf_rwindow_size = WINDOWSIZE64; 525 is64 = 1; 526 } 527 528 ASSERT(wbcnt >= 0 && wbcnt <= SPARC_MAXREGWINDOW); 529 mpcb->mpcb_wbcnt = 0; 530 for (i = 0; i < wbcnt; i++) { 531 sp = (caddr_t)(uintptr_t)gwins->spbuf[i]; 532 mpcb->mpcb_spbuf[i] = sp; 533 rwp = (struct rwindow *) 534 (mpcb->mpcb_wbuf + (i * wbuf_rwindow_size)); 535 if (is64 && IS_V9STACK(sp)) 536 rwindow_32ton(&gwins->wbuf[i], rwp); 537 else 538 bcopy(&gwins->wbuf[i], rwp, sizeof (struct rwindow32)); 539 mpcb->mpcb_wbcnt++; 540 } 541 } 542 543 /* 544 * Get register windows. 545 * NOTE: 'lwp' might not correspond to 'curthread' since this is 546 * called from code in /proc to set the registers of another lwp. 547 */ 548 void 549 getgwins(klwp_t *lwp, gwindows_t *gwp) 550 { 551 struct machpcb *mpcb = lwptompcb(lwp); 552 int wbcnt = mpcb->mpcb_wbcnt; 553 caddr_t sp; 554 int i; 555 struct rwindow32 *rwp; 556 int wbuf_rwindow_size; 557 int is64; 558 559 if (mpcb->mpcb_wstate == WSTATE_USER32) { 560 wbuf_rwindow_size = WINDOWSIZE32; 561 is64 = 0; 562 } else { 563 wbuf_rwindow_size = WINDOWSIZE64; 564 is64 = 1; 565 } 566 ASSERT(wbcnt >= 0 && wbcnt <= SPARC_MAXREGWINDOW); 567 gwp->wbcnt = wbcnt; 568 for (i = 0; i < wbcnt; i++) { 569 sp = mpcb->mpcb_spbuf[i]; 570 gwp->spbuf[i] = (greg_t *)sp; 571 rwp = (struct rwindow32 *) 572 (mpcb->mpcb_wbuf + (i * wbuf_rwindow_size)); 573 if (is64 && IS_V9STACK(sp)) 574 bcopy(rwp, &gwp->wbuf[i], sizeof (struct rwindow)); 575 else 576 rwindow_32ton(rwp, &gwp->wbuf[i]); 577 } 578 } 579 580 void 581 getgwins32(klwp_t *lwp, gwindows32_t *gwp) 582 { 583 struct machpcb *mpcb = lwptompcb(lwp); 584 int wbcnt = mpcb->mpcb_wbcnt; 585 int i; 586 struct rwindow *rwp; 587 int wbuf_rwindow_size; 588 caddr_t sp; 589 int is64; 590 591 if (mpcb->mpcb_wstate == WSTATE_USER32) { 592 wbuf_rwindow_size = WINDOWSIZE32; 593 is64 = 0; 594 } else { 595 wbuf_rwindow_size = WINDOWSIZE64; 596 is64 = 1; 597 } 598 599 ASSERT(wbcnt >= 0 && wbcnt <= SPARC_MAXREGWINDOW); 600 gwp->wbcnt = wbcnt; 601 for (i = 0; i < wbcnt; i++) { 602 sp = mpcb->mpcb_spbuf[i]; 603 rwp = (struct rwindow *) 604 (mpcb->mpcb_wbuf + (i * wbuf_rwindow_size)); 605 gwp->spbuf[i] = (caddr32_t)(uintptr_t)sp; 606 if (is64 && IS_V9STACK(sp)) 607 rwindow_nto32(rwp, &gwp->wbuf[i]); 608 else 609 bcopy(rwp, &gwp->wbuf[i], sizeof (struct rwindow32)); 610 } 611 } 612 613 /* 614 * For things that depend on register state being on the stack, 615 * copy any register windows that get saved into the window buffer 616 * (in the pcb) onto the stack. This normally gets fixed up 617 * before returning to a user program. Callers of this routine 618 * require this to happen immediately because a later kernel 619 * operation depends on window state (like instruction simulation). 620 */ 621 int 622 flush_user_windows_to_stack(caddr_t *psp) 623 { 624 int j, k; 625 caddr_t sp; 626 struct machpcb *mpcb = lwptompcb(ttolwp(curthread)); 627 int err; 628 int error = 0; 629 int wbuf_rwindow_size; 630 int rwindow_size; 631 int stack_align; 632 int watched; 633 634 flush_user_windows(); 635 636 if (mpcb->mpcb_wstate != WSTATE_USER32) 637 wbuf_rwindow_size = WINDOWSIZE64; 638 else 639 wbuf_rwindow_size = WINDOWSIZE32; 640 641 j = mpcb->mpcb_wbcnt; 642 while (j > 0) { 643 sp = mpcb->mpcb_spbuf[--j]; 644 645 if ((mpcb->mpcb_wstate != WSTATE_USER32) && 646 IS_V9STACK(sp)) { 647 sp += V9BIAS64; 648 stack_align = STACK_ALIGN64; 649 rwindow_size = WINDOWSIZE64; 650 } else { 651 /* 652 * Reduce sp to a 32 bit value. This was originally 653 * done by casting down to uint32_t and back up to 654 * caddr_t, but one compiler didn't like that, so the 655 * uintptr_t casts were added. The temporary 32 bit 656 * variable was introduced to avoid depending on all 657 * compilers to generate the desired assembly code for a 658 * quadruple cast in a single expression. 659 */ 660 caddr32_t sp32 = (uint32_t)(uintptr_t)sp; 661 sp = (caddr_t)(uintptr_t)sp32; 662 663 stack_align = STACK_ALIGN32; 664 rwindow_size = WINDOWSIZE32; 665 } 666 if (((uintptr_t)sp & (stack_align - 1)) != 0) 667 continue; 668 669 watched = watch_disable_addr(sp, rwindow_size, S_WRITE); 670 err = xcopyout(mpcb->mpcb_wbuf + 671 (j * wbuf_rwindow_size), sp, rwindow_size); 672 if (err != 0) { 673 if (psp != NULL) { 674 /* 675 * Determine the offending address. 676 * It may not be the stack pointer itself. 677 */ 678 uint_t *kaddr = (uint_t *)(mpcb->mpcb_wbuf + 679 (j * wbuf_rwindow_size)); 680 uint_t *uaddr = (uint_t *)sp; 681 682 for (k = 0; 683 k < rwindow_size / sizeof (int); 684 k++, kaddr++, uaddr++) { 685 if (suword32(uaddr, *kaddr)) 686 break; 687 } 688 689 /* can't happen? */ 690 if (k == rwindow_size / sizeof (int)) 691 uaddr = (uint_t *)sp; 692 693 *psp = (caddr_t)uaddr; 694 } 695 error = err; 696 } else { 697 /* 698 * stack was aligned and copyout succeeded; 699 * move other windows down. 700 */ 701 mpcb->mpcb_wbcnt--; 702 for (k = j; k < mpcb->mpcb_wbcnt; k++) { 703 mpcb->mpcb_spbuf[k] = mpcb->mpcb_spbuf[k+1]; 704 bcopy( 705 mpcb->mpcb_wbuf + 706 ((k+1) * wbuf_rwindow_size), 707 mpcb->mpcb_wbuf + 708 (k * wbuf_rwindow_size), 709 wbuf_rwindow_size); 710 } 711 } 712 if (watched) 713 watch_enable_addr(sp, rwindow_size, S_WRITE); 714 } /* while there are windows in the wbuf */ 715 return (error); 716 } 717 718 static int 719 copy_return_window32(int dotwo) 720 { 721 klwp_t *lwp = ttolwp(curthread); 722 struct machpcb *mpcb = lwptompcb(lwp); 723 struct rwindow32 rwindow32; 724 caddr_t sp1; 725 caddr_t sp2; 726 727 (void) flush_user_windows_to_stack(NULL); 728 if (mpcb->mpcb_rsp[0] == NULL) { 729 /* 730 * Reduce r_sp to a 32 bit value before storing it in sp1. This 731 * was originally done by casting down to uint32_t and back up 732 * to caddr_t, but that generated complaints under one compiler. 733 * The uintptr_t cast was added to address that, and the 734 * temporary 32 bit variable was introduced to avoid depending 735 * on all compilers to generate the desired assembly code for a 736 * triple cast in a single expression. 737 */ 738 caddr32_t sp1_32 = (uint32_t)lwptoregs(lwp)->r_sp; 739 sp1 = (caddr_t)(uintptr_t)sp1_32; 740 741 if ((copyin_nowatch(sp1, &rwindow32, 742 sizeof (struct rwindow32))) == 0) 743 mpcb->mpcb_rsp[0] = sp1; 744 rwindow_32ton(&rwindow32, &mpcb->mpcb_rwin[0]); 745 } 746 mpcb->mpcb_rsp[1] = NULL; 747 if (dotwo && mpcb->mpcb_rsp[0] != NULL && 748 (sp2 = (caddr_t)mpcb->mpcb_rwin[0].rw_fp) != NULL) { 749 if ((copyin_nowatch(sp2, &rwindow32, 750 sizeof (struct rwindow32)) == 0)) 751 mpcb->mpcb_rsp[1] = sp2; 752 rwindow_32ton(&rwindow32, &mpcb->mpcb_rwin[1]); 753 } 754 return (mpcb->mpcb_rsp[0] != NULL); 755 } 756 757 int 758 copy_return_window(int dotwo) 759 { 760 proc_t *p = ttoproc(curthread); 761 klwp_t *lwp; 762 struct machpcb *mpcb; 763 caddr_t sp1; 764 caddr_t sp2; 765 766 if (p->p_model == DATAMODEL_ILP32) 767 return (copy_return_window32(dotwo)); 768 769 lwp = ttolwp(curthread); 770 mpcb = lwptompcb(lwp); 771 (void) flush_user_windows_to_stack(NULL); 772 if (mpcb->mpcb_rsp[0] == NULL) { 773 sp1 = (caddr_t)lwptoregs(lwp)->r_sp + STACK_BIAS; 774 if ((copyin_nowatch(sp1, &mpcb->mpcb_rwin[0], 775 sizeof (struct rwindow)) == 0)) 776 mpcb->mpcb_rsp[0] = sp1 - STACK_BIAS; 777 } 778 mpcb->mpcb_rsp[1] = NULL; 779 if (dotwo && mpcb->mpcb_rsp[0] != NULL && 780 (sp2 = (caddr_t)mpcb->mpcb_rwin[0].rw_fp) != NULL) { 781 sp2 += STACK_BIAS; 782 if ((copyin_nowatch(sp2, &mpcb->mpcb_rwin[1], 783 sizeof (struct rwindow)) == 0)) 784 mpcb->mpcb_rsp[1] = sp2 - STACK_BIAS; 785 } 786 return (mpcb->mpcb_rsp[0] != NULL); 787 } 788 789 /* 790 * Clear registers on exec(2). 791 */ 792 void 793 setregs(uarg_t *args) 794 { 795 struct regs *rp; 796 klwp_t *lwp = ttolwp(curthread); 797 kfpu_t *fpp = lwptofpu(lwp); 798 struct machpcb *mpcb = lwptompcb(lwp); 799 proc_t *p = ttoproc(curthread); 800 801 /* 802 * Initialize user registers. 803 */ 804 (void) save_syscall_args(); /* copy args from registers first */ 805 rp = lwptoregs(lwp); 806 rp->r_g1 = rp->r_g2 = rp->r_g3 = rp->r_g4 = rp->r_g5 = 807 rp->r_g6 = rp->r_o0 = rp->r_o1 = rp->r_o2 = 808 rp->r_o3 = rp->r_o4 = rp->r_o5 = rp->r_o7 = 0; 809 if (p->p_model == DATAMODEL_ILP32) 810 rp->r_tstate = TSTATE_USER32 | weakest_mem_model; 811 else 812 rp->r_tstate = TSTATE_USER64 | weakest_mem_model; 813 if (!fpu_exists) 814 rp->r_tstate &= ~TSTATE_PEF; 815 rp->r_g7 = args->thrptr; 816 rp->r_pc = args->entry; 817 rp->r_npc = args->entry + 4; 818 rp->r_y = 0; 819 curthread->t_post_sys = 1; 820 lwp->lwp_eosys = JUSTRETURN; 821 lwp->lwp_pcb.pcb_trap0addr = NULL; /* no trap 0 handler */ 822 /* 823 * Clear the fixalignment flag 824 */ 825 p->p_fixalignment = 0; 826 827 /* 828 * Throw out old user windows, init window buf. 829 */ 830 trash_user_windows(); 831 832 if (p->p_model == DATAMODEL_LP64 && 833 mpcb->mpcb_wstate != WSTATE_USER64) { 834 ASSERT(mpcb->mpcb_wbcnt == 0); 835 kmem_cache_free(wbuf32_cache, mpcb->mpcb_wbuf); 836 mpcb->mpcb_wbuf = kmem_cache_alloc(wbuf64_cache, KM_SLEEP); 837 ASSERT(((uintptr_t)mpcb->mpcb_wbuf & 7) == 0); 838 mpcb->mpcb_wstate = WSTATE_USER64; 839 } else if (p->p_model == DATAMODEL_ILP32 && 840 mpcb->mpcb_wstate != WSTATE_USER32) { 841 ASSERT(mpcb->mpcb_wbcnt == 0); 842 kmem_cache_free(wbuf64_cache, mpcb->mpcb_wbuf); 843 mpcb->mpcb_wbuf = kmem_cache_alloc(wbuf32_cache, KM_SLEEP); 844 mpcb->mpcb_wstate = WSTATE_USER32; 845 } 846 mpcb->mpcb_pa = va_to_pa(mpcb); 847 mpcb->mpcb_wbuf_pa = va_to_pa(mpcb->mpcb_wbuf); 848 849 /* 850 * Here we initialize minimal fpu state. 851 * The rest is done at the first floating 852 * point instruction that a process executes 853 * or by the lib_psr memcpy routines. 854 */ 855 if (fpu_exists) { 856 extern void _fp_write_fprs(unsigned); 857 _fp_write_fprs(0); 858 } 859 fpp->fpu_en = 0; 860 fpp->fpu_fprs = 0; 861 } 862 863 void 864 lwp_swapin(kthread_t *tp) 865 { 866 struct machpcb *mpcb = lwptompcb(ttolwp(tp)); 867 868 mpcb->mpcb_pa = va_to_pa(mpcb); 869 mpcb->mpcb_wbuf_pa = va_to_pa(mpcb->mpcb_wbuf); 870 } 871 872 /* 873 * Construct the execution environment for the user's signal 874 * handler and arrange for control to be given to it on return 875 * to userland. The library code now calls setcontext() to 876 * clean up after the signal handler, so sigret() is no longer 877 * needed. 878 */ 879 int 880 sendsig(int sig, k_siginfo_t *sip, void (*hdlr)()) 881 { 882 /* 883 * 'volatile' is needed to ensure that values are 884 * correct on the error return from on_fault(). 885 */ 886 volatile int minstacksz; /* min stack required to catch signal */ 887 int newstack = 0; /* if true, switching to altstack */ 888 label_t ljb; 889 caddr_t sp; 890 struct regs *volatile rp; 891 klwp_t *lwp = ttolwp(curthread); 892 proc_t *volatile p = ttoproc(curthread); 893 int fpq_size = 0; 894 struct sigframe { 895 struct frame frwin; 896 ucontext_t uc; 897 }; 898 siginfo_t *sip_addr; 899 struct sigframe *volatile fp; 900 ucontext_t *volatile tuc = NULL; 901 char *volatile xregs = NULL; 902 volatile size_t xregs_size = 0; 903 gwindows_t *volatile gwp = NULL; 904 volatile int gwin_size = 0; 905 kfpu_t *fpp; 906 struct machpcb *mpcb; 907 volatile int watched = 0; 908 volatile int watched2 = 0; 909 caddr_t tos; 910 911 /* 912 * Make sure the current last user window has been flushed to 913 * the stack save area before we change the sp. 914 * Restore register window if a debugger modified it. 915 */ 916 (void) flush_user_windows_to_stack(NULL); 917 if (lwp->lwp_pcb.pcb_xregstat != XREGNONE) 918 xregrestore(lwp, 0); 919 920 mpcb = lwptompcb(lwp); 921 rp = lwptoregs(lwp); 922 923 /* 924 * Clear the watchpoint return stack pointers. 925 */ 926 mpcb->mpcb_rsp[0] = NULL; 927 mpcb->mpcb_rsp[1] = NULL; 928 929 minstacksz = sizeof (struct sigframe); 930 931 /* 932 * We know that sizeof (siginfo_t) is stack-aligned: 933 * 128 bytes for ILP32, 256 bytes for LP64. 934 */ 935 if (sip != NULL) 936 minstacksz += sizeof (siginfo_t); 937 938 /* 939 * These two fields are pointed to by ABI structures and may 940 * be of arbitrary length. Size them now so we know how big 941 * the signal frame has to be. 942 */ 943 fpp = lwptofpu(lwp); 944 fpp->fpu_fprs = _fp_read_fprs(); 945 if ((fpp->fpu_en) || (fpp->fpu_fprs & FPRS_FEF)) { 946 fpq_size = fpp->fpu_q_entrysize * fpp->fpu_qcnt; 947 minstacksz += SA(fpq_size); 948 } 949 950 mpcb = lwptompcb(lwp); 951 if (mpcb->mpcb_wbcnt != 0) { 952 gwin_size = (mpcb->mpcb_wbcnt * sizeof (struct rwindow)) + 953 (SPARC_MAXREGWINDOW * sizeof (caddr_t)) + sizeof (long); 954 minstacksz += SA(gwin_size); 955 } 956 957 /* 958 * Extra registers, if support by this platform, may be of arbitrary 959 * length. Size them now so we know how big the signal frame has to be. 960 * For sparcv9 _LP64 user programs, use asrs instead of the xregs. 961 */ 962 minstacksz += SA(xregs_size); 963 964 /* 965 * Figure out whether we will be handling this signal on 966 * an alternate stack specified by the user. Then allocate 967 * and validate the stack requirements for the signal handler 968 * context. on_fault will catch any faults. 969 */ 970 newstack = (sigismember(&PTOU(curproc)->u_sigonstack, sig) && 971 !(lwp->lwp_sigaltstack.ss_flags & (SS_ONSTACK|SS_DISABLE))); 972 973 tos = (caddr_t)rp->r_sp + STACK_BIAS; 974 /* 975 * Force proper stack pointer alignment, even in the face of a 976 * misaligned stack pointer from user-level before the signal. 977 * Don't use the SA() macro because that rounds up, not down. 978 */ 979 tos = (caddr_t)((uintptr_t)tos & ~(STACK_ALIGN - 1ul)); 980 981 if (newstack != 0) { 982 fp = (struct sigframe *) 983 (SA((uintptr_t)lwp->lwp_sigaltstack.ss_sp) + 984 SA((int)lwp->lwp_sigaltstack.ss_size) - STACK_ALIGN - 985 SA(minstacksz)); 986 } else { 987 /* 988 * If we were unable to flush all register windows to 989 * the stack and we are not now on an alternate stack, 990 * just dump core with a SIGSEGV back in psig(). 991 */ 992 if (sig == SIGSEGV && 993 mpcb->mpcb_wbcnt != 0 && 994 !(lwp->lwp_sigaltstack.ss_flags & SS_ONSTACK)) 995 return (0); 996 fp = (struct sigframe *)(tos - SA(minstacksz)); 997 /* 998 * Could call grow here, but stack growth now handled below 999 * in code protected by on_fault(). 1000 */ 1001 } 1002 sp = (caddr_t)fp + sizeof (struct sigframe); 1003 1004 /* 1005 * Make sure process hasn't trashed its stack. 1006 */ 1007 if ((caddr_t)fp >= p->p_usrstack || 1008 (caddr_t)fp + SA(minstacksz) >= p->p_usrstack) { 1009 #ifdef DEBUG 1010 printf("sendsig: bad signal stack cmd=%s, pid=%d, sig=%d\n", 1011 PTOU(p)->u_comm, p->p_pid, sig); 1012 printf("sigsp = 0x%p, action = 0x%p, upc = 0x%lx\n", 1013 (void *)fp, (void *)hdlr, rp->r_pc); 1014 printf("fp above USRSTACK\n"); 1015 #endif 1016 return (0); 1017 } 1018 1019 watched = watch_disable_addr((caddr_t)fp, SA(minstacksz), S_WRITE); 1020 if (on_fault(&ljb)) 1021 goto badstack; 1022 1023 tuc = kmem_alloc(sizeof (ucontext_t), KM_SLEEP); 1024 savecontext(tuc, lwp->lwp_sigoldmask); 1025 1026 /* 1027 * save extra register state if it exists 1028 */ 1029 if (xregs_size != 0) { 1030 xregs_setptr(lwp, tuc, sp); 1031 xregs = kmem_alloc(xregs_size, KM_SLEEP); 1032 xregs_get(lwp, xregs); 1033 copyout_noerr(xregs, sp, xregs_size); 1034 kmem_free(xregs, xregs_size); 1035 xregs = NULL; 1036 sp += SA(xregs_size); 1037 } 1038 1039 copyout_noerr(tuc, &fp->uc, sizeof (*tuc)); 1040 kmem_free(tuc, sizeof (*tuc)); 1041 tuc = NULL; 1042 1043 if (sip != NULL) { 1044 zoneid_t zoneid; 1045 1046 uzero(sp, sizeof (siginfo_t)); 1047 if (SI_FROMUSER(sip) && 1048 (zoneid = p->p_zone->zone_id) != GLOBAL_ZONEID && 1049 zoneid != sip->si_zoneid) { 1050 k_siginfo_t sani_sip = *sip; 1051 sani_sip.si_pid = p->p_zone->zone_zsched->p_pid; 1052 sani_sip.si_uid = 0; 1053 sani_sip.si_ctid = -1; 1054 sani_sip.si_zoneid = zoneid; 1055 copyout_noerr(&sani_sip, sp, sizeof (sani_sip)); 1056 } else { 1057 copyout_noerr(sip, sp, sizeof (*sip)); 1058 } 1059 sip_addr = (siginfo_t *)sp; 1060 sp += sizeof (siginfo_t); 1061 1062 if (sig == SIGPROF && 1063 curthread->t_rprof != NULL && 1064 curthread->t_rprof->rp_anystate) { 1065 /* 1066 * We stand on our head to deal with 1067 * the real time profiling signal. 1068 * Fill in the stuff that doesn't fit 1069 * in a normal k_siginfo structure. 1070 */ 1071 int i = sip->si_nsysarg; 1072 while (--i >= 0) { 1073 sulword_noerr( 1074 (ulong_t *)&sip_addr->si_sysarg[i], 1075 (ulong_t)lwp->lwp_arg[i]); 1076 } 1077 copyout_noerr(curthread->t_rprof->rp_state, 1078 sip_addr->si_mstate, 1079 sizeof (curthread->t_rprof->rp_state)); 1080 } 1081 } else { 1082 sip_addr = (siginfo_t *)NULL; 1083 } 1084 1085 /* 1086 * When flush_user_windows_to_stack() can't save all the 1087 * windows to the stack, it puts them in the lwp's pcb. 1088 */ 1089 if (gwin_size != 0) { 1090 gwp = kmem_alloc(gwin_size, KM_SLEEP); 1091 getgwins(lwp, gwp); 1092 sulword_noerr(&fp->uc.uc_mcontext.gwins, (ulong_t)sp); 1093 copyout_noerr(gwp, sp, gwin_size); 1094 kmem_free(gwp, gwin_size); 1095 gwp = NULL; 1096 sp += SA(gwin_size); 1097 } else 1098 sulword_noerr(&fp->uc.uc_mcontext.gwins, (ulong_t)NULL); 1099 1100 if (fpq_size != 0) { 1101 struct fq *fqp = (struct fq *)sp; 1102 sulword_noerr(&fp->uc.uc_mcontext.fpregs.fpu_q, (ulong_t)fqp); 1103 copyout_noerr(mpcb->mpcb_fpu_q, fqp, fpq_size); 1104 1105 /* 1106 * forget the fp queue so that the signal handler can run 1107 * without being harrassed--it will do a setcontext that will 1108 * re-establish the queue if there still is one 1109 * 1110 * NOTE: fp_runq() relies on the qcnt field being zeroed here 1111 * to terminate its processing of the queue after signal 1112 * delivery. 1113 */ 1114 mpcb->mpcb_fpu->fpu_qcnt = 0; 1115 sp += SA(fpq_size); 1116 1117 /* Also, syscall needs to know about this */ 1118 mpcb->mpcb_flags |= FP_TRAPPED; 1119 1120 } else { 1121 sulword_noerr(&fp->uc.uc_mcontext.fpregs.fpu_q, (ulong_t)NULL); 1122 suword8_noerr(&fp->uc.uc_mcontext.fpregs.fpu_qcnt, 0); 1123 } 1124 1125 1126 /* 1127 * Since we flushed the user's windows and we are changing his 1128 * stack pointer, the window that the user will return to will 1129 * be restored from the save area in the frame we are setting up. 1130 * We copy in save area for old stack pointer so that debuggers 1131 * can do a proper stack backtrace from the signal handler. 1132 */ 1133 if (mpcb->mpcb_wbcnt == 0) { 1134 watched2 = watch_disable_addr(tos, sizeof (struct rwindow), 1135 S_READ); 1136 ucopy(tos, &fp->frwin, sizeof (struct rwindow)); 1137 } 1138 1139 lwp->lwp_oldcontext = (uintptr_t)&fp->uc; 1140 1141 if (newstack != 0) { 1142 lwp->lwp_sigaltstack.ss_flags |= SS_ONSTACK; 1143 1144 if (lwp->lwp_ustack) { 1145 copyout_noerr(&lwp->lwp_sigaltstack, 1146 (stack_t *)lwp->lwp_ustack, sizeof (stack_t)); 1147 } 1148 } 1149 1150 no_fault(); 1151 mpcb->mpcb_wbcnt = 0; /* let user go on */ 1152 1153 if (watched2) 1154 watch_enable_addr(tos, sizeof (struct rwindow), S_READ); 1155 if (watched) 1156 watch_enable_addr((caddr_t)fp, SA(minstacksz), S_WRITE); 1157 1158 /* 1159 * Set up user registers for execution of signal handler. 1160 */ 1161 rp->r_sp = (uintptr_t)fp - STACK_BIAS; 1162 rp->r_pc = (uintptr_t)hdlr; 1163 rp->r_npc = (uintptr_t)hdlr + 4; 1164 /* make sure %asi is ASI_PNF */ 1165 rp->r_tstate &= ~((uint64_t)TSTATE_ASI_MASK << TSTATE_ASI_SHIFT); 1166 rp->r_tstate |= ((uint64_t)ASI_PNF << TSTATE_ASI_SHIFT); 1167 rp->r_o0 = sig; 1168 rp->r_o1 = (uintptr_t)sip_addr; 1169 rp->r_o2 = (uintptr_t)&fp->uc; 1170 /* 1171 * Don't set lwp_eosys here. sendsig() is called via psig() after 1172 * lwp_eosys is handled, so setting it here would affect the next 1173 * system call. 1174 */ 1175 return (1); 1176 1177 badstack: 1178 no_fault(); 1179 if (watched2) 1180 watch_enable_addr(tos, sizeof (struct rwindow), S_READ); 1181 if (watched) 1182 watch_enable_addr((caddr_t)fp, SA(minstacksz), S_WRITE); 1183 if (tuc) 1184 kmem_free(tuc, sizeof (ucontext_t)); 1185 if (xregs) 1186 kmem_free(xregs, xregs_size); 1187 if (gwp) 1188 kmem_free(gwp, gwin_size); 1189 #ifdef DEBUG 1190 printf("sendsig: bad signal stack cmd=%s, pid=%d, sig=%d\n", 1191 PTOU(p)->u_comm, p->p_pid, sig); 1192 printf("on fault, sigsp = %p, action = %p, upc = 0x%lx\n", 1193 (void *)fp, (void *)hdlr, rp->r_pc); 1194 #endif 1195 return (0); 1196 } 1197 1198 1199 #ifdef _SYSCALL32_IMPL 1200 1201 /* 1202 * Construct the execution environment for the user's signal 1203 * handler and arrange for control to be given to it on return 1204 * to userland. The library code now calls setcontext() to 1205 * clean up after the signal handler, so sigret() is no longer 1206 * needed. 1207 */ 1208 int 1209 sendsig32(int sig, k_siginfo_t *sip, void (*hdlr)()) 1210 { 1211 /* 1212 * 'volatile' is needed to ensure that values are 1213 * correct on the error return from on_fault(). 1214 */ 1215 volatile int minstacksz; /* min stack required to catch signal */ 1216 int newstack = 0; /* if true, switching to altstack */ 1217 label_t ljb; 1218 caddr_t sp; 1219 struct regs *volatile rp; 1220 klwp_t *lwp = ttolwp(curthread); 1221 proc_t *volatile p = ttoproc(curthread); 1222 struct fq32 fpu_q[MAXFPQ]; /* to hold floating queue */ 1223 struct fq32 *dfq = NULL; 1224 size_t fpq_size = 0; 1225 struct sigframe32 { 1226 struct frame32 frwin; 1227 ucontext32_t uc; 1228 }; 1229 struct sigframe32 *volatile fp; 1230 siginfo32_t *sip_addr; 1231 ucontext32_t *volatile tuc = NULL; 1232 char *volatile xregs = NULL; 1233 volatile int xregs_size = 0; 1234 gwindows32_t *volatile gwp = NULL; 1235 volatile size_t gwin_size = 0; 1236 kfpu_t *fpp; 1237 struct machpcb *mpcb; 1238 volatile int watched = 0; 1239 volatile int watched2 = 0; 1240 caddr_t tos; 1241 1242 /* 1243 * Make sure the current last user window has been flushed to 1244 * the stack save area before we change the sp. 1245 * Restore register window if a debugger modified it. 1246 */ 1247 (void) flush_user_windows_to_stack(NULL); 1248 if (lwp->lwp_pcb.pcb_xregstat != XREGNONE) 1249 xregrestore(lwp, 0); 1250 1251 mpcb = lwptompcb(lwp); 1252 rp = lwptoregs(lwp); 1253 1254 /* 1255 * Clear the watchpoint return stack pointers. 1256 */ 1257 mpcb->mpcb_rsp[0] = NULL; 1258 mpcb->mpcb_rsp[1] = NULL; 1259 1260 minstacksz = sizeof (struct sigframe32); 1261 1262 if (sip != NULL) 1263 minstacksz += sizeof (siginfo32_t); 1264 1265 /* 1266 * These two fields are pointed to by ABI structures and may 1267 * be of arbitrary length. Size them now so we know how big 1268 * the signal frame has to be. 1269 */ 1270 fpp = lwptofpu(lwp); 1271 fpp->fpu_fprs = _fp_read_fprs(); 1272 if ((fpp->fpu_en) || (fpp->fpu_fprs & FPRS_FEF)) { 1273 fpq_size = sizeof (struct fpq32) * fpp->fpu_qcnt; 1274 minstacksz += fpq_size; 1275 dfq = fpu_q; 1276 } 1277 1278 mpcb = lwptompcb(lwp); 1279 if (mpcb->mpcb_wbcnt != 0) { 1280 gwin_size = (mpcb->mpcb_wbcnt * sizeof (struct rwindow32)) + 1281 (SPARC_MAXREGWINDOW * sizeof (caddr32_t)) + 1282 sizeof (int32_t); 1283 minstacksz += gwin_size; 1284 } 1285 1286 /* 1287 * Extra registers, if supported by this platform, may be of arbitrary 1288 * length. Size them now so we know how big the signal frame has to be. 1289 */ 1290 xregs_size = xregs_getsize(p); 1291 minstacksz += SA32(xregs_size); 1292 1293 /* 1294 * Figure out whether we will be handling this signal on 1295 * an alternate stack specified by the user. Then allocate 1296 * and validate the stack requirements for the signal handler 1297 * context. on_fault will catch any faults. 1298 */ 1299 newstack = (sigismember(&PTOU(curproc)->u_sigonstack, sig) && 1300 !(lwp->lwp_sigaltstack.ss_flags & (SS_ONSTACK|SS_DISABLE))); 1301 1302 tos = (void *)(uintptr_t)(uint32_t)rp->r_sp; 1303 /* 1304 * Force proper stack pointer alignment, even in the face of a 1305 * misaligned stack pointer from user-level before the signal. 1306 * Don't use the SA32() macro because that rounds up, not down. 1307 */ 1308 tos = (caddr_t)((uintptr_t)tos & ~(STACK_ALIGN32 - 1ul)); 1309 1310 if (newstack != 0) { 1311 fp = (struct sigframe32 *) 1312 (SA32((uintptr_t)lwp->lwp_sigaltstack.ss_sp) + 1313 SA32((int)lwp->lwp_sigaltstack.ss_size) - 1314 STACK_ALIGN32 - 1315 SA32(minstacksz)); 1316 } else { 1317 /* 1318 * If we were unable to flush all register windows to 1319 * the stack and we are not now on an alternate stack, 1320 * just dump core with a SIGSEGV back in psig(). 1321 */ 1322 if (sig == SIGSEGV && 1323 mpcb->mpcb_wbcnt != 0 && 1324 !(lwp->lwp_sigaltstack.ss_flags & SS_ONSTACK)) 1325 return (0); 1326 fp = (struct sigframe32 *)(tos - SA32(minstacksz)); 1327 /* 1328 * Could call grow here, but stack growth now handled below 1329 * in code protected by on_fault(). 1330 */ 1331 } 1332 sp = (caddr_t)fp + sizeof (struct sigframe32); 1333 1334 /* 1335 * Make sure process hasn't trashed its stack. 1336 */ 1337 if ((caddr_t)fp >= p->p_usrstack || 1338 (caddr_t)fp + SA32(minstacksz) >= p->p_usrstack) { 1339 #ifdef DEBUG 1340 printf("sendsig32: bad signal stack cmd=%s, pid=%d, sig=%d\n", 1341 PTOU(p)->u_comm, p->p_pid, sig); 1342 printf("sigsp = 0x%p, action = 0x%p, upc = 0x%lx\n", 1343 (void *)fp, (void *)hdlr, rp->r_pc); 1344 printf("fp above USRSTACK32\n"); 1345 #endif 1346 return (0); 1347 } 1348 1349 watched = watch_disable_addr((caddr_t)fp, SA32(minstacksz), S_WRITE); 1350 if (on_fault(&ljb)) 1351 goto badstack; 1352 1353 tuc = kmem_alloc(sizeof (ucontext32_t), KM_SLEEP); 1354 savecontext32(tuc, lwp->lwp_sigoldmask, dfq); 1355 1356 /* 1357 * save extra register state if it exists 1358 */ 1359 if (xregs_size != 0) { 1360 xregs_setptr32(lwp, tuc, (caddr32_t)(uintptr_t)sp); 1361 xregs = kmem_alloc(xregs_size, KM_SLEEP); 1362 xregs_get(lwp, xregs); 1363 copyout_noerr(xregs, sp, xregs_size); 1364 kmem_free(xregs, xregs_size); 1365 xregs = NULL; 1366 sp += SA32(xregs_size); 1367 } 1368 1369 copyout_noerr(tuc, &fp->uc, sizeof (*tuc)); 1370 kmem_free(tuc, sizeof (*tuc)); 1371 tuc = NULL; 1372 1373 if (sip != NULL) { 1374 siginfo32_t si32; 1375 zoneid_t zoneid; 1376 1377 siginfo_kto32(sip, &si32); 1378 if (SI_FROMUSER(sip) && 1379 (zoneid = p->p_zone->zone_id) != GLOBAL_ZONEID && 1380 zoneid != sip->si_zoneid) { 1381 si32.si_pid = p->p_zone->zone_zsched->p_pid; 1382 si32.si_uid = 0; 1383 si32.si_ctid = -1; 1384 si32.si_zoneid = zoneid; 1385 } 1386 uzero(sp, sizeof (siginfo32_t)); 1387 copyout_noerr(&si32, sp, sizeof (siginfo32_t)); 1388 sip_addr = (siginfo32_t *)sp; 1389 sp += sizeof (siginfo32_t); 1390 1391 if (sig == SIGPROF && 1392 curthread->t_rprof != NULL && 1393 curthread->t_rprof->rp_anystate) { 1394 /* 1395 * We stand on our head to deal with 1396 * the real time profiling signal. 1397 * Fill in the stuff that doesn't fit 1398 * in a normal k_siginfo structure. 1399 */ 1400 int i = sip->si_nsysarg; 1401 while (--i >= 0) { 1402 suword32_noerr(&sip_addr->si_sysarg[i], 1403 (uint32_t)lwp->lwp_arg[i]); 1404 } 1405 copyout_noerr(curthread->t_rprof->rp_state, 1406 sip_addr->si_mstate, 1407 sizeof (curthread->t_rprof->rp_state)); 1408 } 1409 } else { 1410 sip_addr = NULL; 1411 } 1412 1413 /* 1414 * When flush_user_windows_to_stack() can't save all the 1415 * windows to the stack, it puts them in the lwp's pcb. 1416 */ 1417 if (gwin_size != 0) { 1418 gwp = kmem_alloc(gwin_size, KM_SLEEP); 1419 getgwins32(lwp, gwp); 1420 suword32_noerr(&fp->uc.uc_mcontext.gwins, 1421 (uint32_t)(uintptr_t)sp); 1422 copyout_noerr(gwp, sp, gwin_size); 1423 kmem_free(gwp, gwin_size); 1424 gwp = NULL; 1425 sp += gwin_size; 1426 } else { 1427 suword32_noerr(&fp->uc.uc_mcontext.gwins, (uint32_t)NULL); 1428 } 1429 1430 if (fpq_size != 0) { 1431 /* 1432 * Update the (already copied out) fpu32.fpu_q pointer 1433 * from NULL to the 32-bit address on the user's stack 1434 * where we then copyout the fq32 to. 1435 */ 1436 struct fq32 *fqp = (struct fq32 *)sp; 1437 suword32_noerr(&fp->uc.uc_mcontext.fpregs.fpu_q, 1438 (uint32_t)(uintptr_t)fqp); 1439 copyout_noerr(dfq, fqp, fpq_size); 1440 1441 /* 1442 * forget the fp queue so that the signal handler can run 1443 * without being harrassed--it will do a setcontext that will 1444 * re-establish the queue if there still is one 1445 * 1446 * NOTE: fp_runq() relies on the qcnt field being zeroed here 1447 * to terminate its processing of the queue after signal 1448 * delivery. 1449 */ 1450 mpcb->mpcb_fpu->fpu_qcnt = 0; 1451 sp += fpq_size; 1452 1453 /* Also, syscall needs to know about this */ 1454 mpcb->mpcb_flags |= FP_TRAPPED; 1455 1456 } else { 1457 suword32_noerr(&fp->uc.uc_mcontext.fpregs.fpu_q, 1458 (uint32_t)NULL); 1459 suword8_noerr(&fp->uc.uc_mcontext.fpregs.fpu_qcnt, 0); 1460 } 1461 1462 1463 /* 1464 * Since we flushed the user's windows and we are changing his 1465 * stack pointer, the window that the user will return to will 1466 * be restored from the save area in the frame we are setting up. 1467 * We copy in save area for old stack pointer so that debuggers 1468 * can do a proper stack backtrace from the signal handler. 1469 */ 1470 if (mpcb->mpcb_wbcnt == 0) { 1471 watched2 = watch_disable_addr(tos, sizeof (struct rwindow32), 1472 S_READ); 1473 ucopy(tos, &fp->frwin, sizeof (struct rwindow32)); 1474 } 1475 1476 lwp->lwp_oldcontext = (uintptr_t)&fp->uc; 1477 1478 if (newstack != 0) { 1479 lwp->lwp_sigaltstack.ss_flags |= SS_ONSTACK; 1480 if (lwp->lwp_ustack) { 1481 stack32_t stk32; 1482 1483 stk32.ss_sp = 1484 (caddr32_t)(uintptr_t)lwp->lwp_sigaltstack.ss_sp; 1485 stk32.ss_size = (size32_t)lwp->lwp_sigaltstack.ss_size; 1486 stk32.ss_flags = (int32_t)lwp->lwp_sigaltstack.ss_flags; 1487 1488 copyout_noerr(&stk32, (stack32_t *)lwp->lwp_ustack, 1489 sizeof (stack32_t)); 1490 } 1491 } 1492 1493 no_fault(); 1494 mpcb->mpcb_wbcnt = 0; /* let user go on */ 1495 1496 if (watched2) 1497 watch_enable_addr(tos, sizeof (struct rwindow32), S_READ); 1498 if (watched) 1499 watch_enable_addr((caddr_t)fp, SA32(minstacksz), S_WRITE); 1500 1501 /* 1502 * Set up user registers for execution of signal handler. 1503 */ 1504 rp->r_sp = (uintptr_t)fp; 1505 rp->r_pc = (uintptr_t)hdlr; 1506 rp->r_npc = (uintptr_t)hdlr + 4; 1507 /* make sure %asi is ASI_PNF */ 1508 rp->r_tstate &= ~((uint64_t)TSTATE_ASI_MASK << TSTATE_ASI_SHIFT); 1509 rp->r_tstate |= ((uint64_t)ASI_PNF << TSTATE_ASI_SHIFT); 1510 rp->r_o0 = sig; 1511 rp->r_o1 = (uintptr_t)sip_addr; 1512 rp->r_o2 = (uintptr_t)&fp->uc; 1513 /* 1514 * Don't set lwp_eosys here. sendsig() is called via psig() after 1515 * lwp_eosys is handled, so setting it here would affect the next 1516 * system call. 1517 */ 1518 return (1); 1519 1520 badstack: 1521 no_fault(); 1522 if (watched2) 1523 watch_enable_addr(tos, sizeof (struct rwindow32), S_READ); 1524 if (watched) 1525 watch_enable_addr((caddr_t)fp, SA32(minstacksz), S_WRITE); 1526 if (tuc) 1527 kmem_free(tuc, sizeof (*tuc)); 1528 if (xregs) 1529 kmem_free(xregs, xregs_size); 1530 if (gwp) 1531 kmem_free(gwp, gwin_size); 1532 #ifdef DEBUG 1533 printf("sendsig32: bad signal stack cmd=%s, pid=%d, sig=%d\n", 1534 PTOU(p)->u_comm, p->p_pid, sig); 1535 printf("on fault, sigsp = 0x%p, action = 0x%p, upc = 0x%lx\n", 1536 (void *)fp, (void *)hdlr, rp->r_pc); 1537 #endif 1538 return (0); 1539 } 1540 1541 #endif /* _SYSCALL32_IMPL */ 1542 1543 1544 /* 1545 * Load user registers into lwp. Called only from syslwp_create(). 1546 * thrptr ignored for sparc. 1547 */ 1548 /* ARGSUSED2 */ 1549 void 1550 lwp_load(klwp_t *lwp, gregset_t grp, uintptr_t thrptr) 1551 { 1552 setgregs(lwp, grp); 1553 if (lwptoproc(lwp)->p_model == DATAMODEL_ILP32) 1554 lwptoregs(lwp)->r_tstate = TSTATE_USER32 | TSTATE_MM_TSO; 1555 else 1556 lwptoregs(lwp)->r_tstate = TSTATE_USER64 | TSTATE_MM_TSO; 1557 1558 if (!fpu_exists) 1559 lwptoregs(lwp)->r_tstate &= ~TSTATE_PEF; 1560 lwp->lwp_eosys = JUSTRETURN; 1561 lwptot(lwp)->t_post_sys = 1; 1562 } 1563 1564 /* 1565 * set syscall()'s return values for a lwp. 1566 */ 1567 void 1568 lwp_setrval(klwp_t *lwp, int v1, int v2) 1569 { 1570 struct regs *rp = lwptoregs(lwp); 1571 1572 rp->r_tstate &= ~TSTATE_IC; 1573 rp->r_o0 = v1; 1574 rp->r_o1 = v2; 1575 } 1576 1577 /* 1578 * set stack pointer for a lwp 1579 */ 1580 void 1581 lwp_setsp(klwp_t *lwp, caddr_t sp) 1582 { 1583 struct regs *rp = lwptoregs(lwp); 1584 rp->r_sp = (uintptr_t)sp; 1585 } 1586 1587 /* 1588 * Take any PCB specific actions that are required or flagged in the PCB. 1589 */ 1590 extern void trap_async_hwerr(void); 1591 #pragma weak trap_async_hwerr 1592 1593 void 1594 lwp_pcb_exit(void) 1595 { 1596 klwp_t *lwp = ttolwp(curthread); 1597 1598 if (lwp->lwp_pcb.pcb_flags & ASYNC_HWERR) { 1599 trap_async_hwerr(); 1600 } 1601 } 1602 1603 /* 1604 * Invalidate the saved user register windows in the pcb struct 1605 * for the current thread. They will no longer be preserved. 1606 */ 1607 void 1608 lwp_clear_uwin(void) 1609 { 1610 struct machpcb *m = lwptompcb(ttolwp(curthread)); 1611 1612 /* 1613 * This has the effect of invalidating all (any) of the 1614 * user level windows that are currently sitting in the 1615 * kernel buffer. 1616 */ 1617 m->mpcb_wbcnt = 0; 1618 } 1619 1620 /* 1621 * Set memory model to Total Store Order (TSO). 1622 */ 1623 static void 1624 mmodel_set_tso(void) 1625 { 1626 struct regs *rp = lwptoregs(ttolwp(curthread)); 1627 1628 /* 1629 * The thread is doing something which requires TSO semantics 1630 * (creating a 2nd thread, or mapping writable shared memory). 1631 * It's no longer safe to run in WC mode. 1632 */ 1633 rp->r_tstate &= ~TSTATE_MM; 1634 /* LINTED E_EXPR_NULL_EFFECT */ 1635 rp->r_tstate |= TSTATE_MM_TSO; 1636 } 1637 1638 /* 1639 * When this routine is invoked, the process is just about to add a new lwp; 1640 * making it multi threaded. 1641 * 1642 * If the program requires default stronger/legacy memory model semantics, 1643 * this is an indication that the processor memory model 1644 * should be altered to provide those semantics. 1645 */ 1646 void 1647 lwp_mmodel_newlwp(void) 1648 { 1649 /* 1650 * New thread has been created and it's no longer safe 1651 * to run in WC mode, so revert back to TSO. 1652 */ 1653 mmodel_set_tso(); 1654 } 1655 1656 /* 1657 * This routine is invoked immediately after the lwp has added a mapping 1658 * to shared memory to its address space. The mapping starts at address 1659 * 'addr' and extends for 'size' bytes. 1660 * 1661 * Unless we can (somehow) guarantee that all the processes we're sharing 1662 * the underlying mapped object with, are using the same memory model that 1663 * this process is using, this call should change the memory model 1664 * configuration of the processor to be the most pessimistic available. 1665 */ 1666 /* ARGSUSED */ 1667 void 1668 lwp_mmodel_shared_as(caddr_t addr, size_t sz) 1669 { 1670 /* 1671 * lwp has mapped shared memory and is no longer safe 1672 * to run in WC mode, so revert back to TSO. 1673 * For now, any shared memory access is enough to get back to TSO 1674 * and hence not checking on 'addr' & 'sz'. 1675 */ 1676 mmodel_set_tso(); 1677 } 1678 1679 static uint_t 1680 mkpsr(uint64_t tstate, uint_t fprs) 1681 { 1682 uint_t psr, icc; 1683 1684 psr = tstate & TSTATE_CWP_MASK; 1685 if (tstate & TSTATE_PRIV) 1686 psr |= PSR_PS; 1687 if (fprs & FPRS_FEF) 1688 psr |= PSR_EF; 1689 icc = (uint_t)(tstate >> PSR_TSTATE_CC_SHIFT) & PSR_ICC; 1690 psr |= icc; 1691 psr |= V9_PSR_IMPLVER; 1692 return (psr); 1693 } 1694 1695 void 1696 sync_icache(caddr_t va, uint_t len) 1697 { 1698 caddr_t end; 1699 1700 end = va + len; 1701 va = (caddr_t)((uintptr_t)va & -8l); /* sparc needs 8-byte align */ 1702 while (va < end) { 1703 doflush(va); 1704 va += 8; 1705 } 1706 } 1707 1708 #ifdef _SYSCALL32_IMPL 1709 1710 /* 1711 * Copy the floating point queue if and only if there is a queue and a place 1712 * to copy it to. Let xregs take care of the other fp regs, for v8plus. 1713 * The issue is that while we are handling the fq32 in sendsig, we 1714 * still need a 64-bit pointer to it, and the caddr32_t in fpregset32_t 1715 * will not suffice, so we have the third parameter to this function. 1716 */ 1717 void 1718 fpuregset_nto32(const fpregset_t *src, fpregset32_t *dest, struct fq32 *dfq) 1719 { 1720 int i; 1721 1722 bzero(dest, sizeof (*dest)); 1723 for (i = 0; i < 32; i++) 1724 dest->fpu_fr.fpu_regs[i] = src->fpu_fr.fpu_regs[i]; 1725 dest->fpu_q = NULL; 1726 dest->fpu_fsr = (uint32_t)src->fpu_fsr; 1727 dest->fpu_qcnt = src->fpu_qcnt; 1728 dest->fpu_q_entrysize = sizeof (struct fpq32); 1729 dest->fpu_en = src->fpu_en; 1730 1731 if ((src->fpu_qcnt) && (dfq != NULL)) { 1732 struct fq *sfq = src->fpu_q; 1733 for (i = 0; i < src->fpu_qcnt; i++, dfq++, sfq++) { 1734 dfq->FQu.fpq.fpq_addr = 1735 (caddr32_t)(uintptr_t)sfq->FQu.fpq.fpq_addr; 1736 dfq->FQu.fpq.fpq_instr = sfq->FQu.fpq.fpq_instr; 1737 } 1738 } 1739 } 1740 1741 /* 1742 * Copy the floating point queue if and only if there is a queue and a place 1743 * to copy it to. Let xregs take care of the other fp regs, for v8plus. 1744 * The *dfq is required to escape the bzero in both this function and in 1745 * ucontext_32ton. The *sfq is required because once the fq32 is copied 1746 * into the kernel, in setcontext, then we need a 64-bit pointer to it. 1747 */ 1748 static void 1749 fpuregset_32ton(const fpregset32_t *src, fpregset_t *dest, 1750 const struct fq32 *sfq, struct fq *dfq) 1751 { 1752 int i; 1753 1754 bzero(dest, sizeof (*dest)); 1755 for (i = 0; i < 32; i++) 1756 dest->fpu_fr.fpu_regs[i] = src->fpu_fr.fpu_regs[i]; 1757 dest->fpu_q = dfq; 1758 dest->fpu_fsr = (uint64_t)src->fpu_fsr; 1759 if ((dest->fpu_qcnt = src->fpu_qcnt) > 0) 1760 dest->fpu_q_entrysize = sizeof (struct fpq); 1761 else 1762 dest->fpu_q_entrysize = 0; 1763 dest->fpu_en = src->fpu_en; 1764 1765 if ((src->fpu_qcnt) && (sfq) && (dfq)) { 1766 for (i = 0; i < src->fpu_qcnt; i++, dfq++, sfq++) { 1767 dfq->FQu.fpq.fpq_addr = 1768 (unsigned int *)(uintptr_t)sfq->FQu.fpq.fpq_addr; 1769 dfq->FQu.fpq.fpq_instr = sfq->FQu.fpq.fpq_instr; 1770 } 1771 } 1772 } 1773 1774 void 1775 ucontext_32ton(const ucontext32_t *src, ucontext_t *dest, 1776 const struct fq32 *sfq, struct fq *dfq) 1777 { 1778 int i; 1779 1780 bzero(dest, sizeof (*dest)); 1781 1782 dest->uc_flags = src->uc_flags; 1783 dest->uc_link = (ucontext_t *)(uintptr_t)src->uc_link; 1784 1785 for (i = 0; i < 4; i++) { 1786 dest->uc_sigmask.__sigbits[i] = src->uc_sigmask.__sigbits[i]; 1787 } 1788 1789 dest->uc_stack.ss_sp = (void *)(uintptr_t)src->uc_stack.ss_sp; 1790 dest->uc_stack.ss_size = (size_t)src->uc_stack.ss_size; 1791 dest->uc_stack.ss_flags = src->uc_stack.ss_flags; 1792 1793 /* REG_CCR is 0, skip over it and handle it after this loop */ 1794 for (i = 1; i < _NGREG32; i++) 1795 dest->uc_mcontext.gregs[i] = 1796 (greg_t)(uint32_t)src->uc_mcontext.gregs[i]; 1797 dest->uc_mcontext.gregs[REG_CCR] = 1798 (src->uc_mcontext.gregs[REG_PSR] & PSR_ICC) >> PSR_ICC_SHIFT; 1799 dest->uc_mcontext.gregs[REG_ASI] = ASI_PNF; 1800 /* 1801 * A valid fpregs is only copied in if (uc.uc_flags & UC_FPU), 1802 * otherwise there is no guarantee that anything in fpregs is valid. 1803 */ 1804 if (src->uc_flags & UC_FPU) { 1805 dest->uc_mcontext.gregs[REG_FPRS] = 1806 ((src->uc_mcontext.fpregs.fpu_en) ? 1807 (FPRS_DU|FPRS_DL|FPRS_FEF) : 0); 1808 } else { 1809 dest->uc_mcontext.gregs[REG_FPRS] = 0; 1810 } 1811 dest->uc_mcontext.gwins = 1812 (gwindows_t *)(uintptr_t)src->uc_mcontext.gwins; 1813 if (src->uc_flags & UC_FPU) { 1814 fpuregset_32ton(&src->uc_mcontext.fpregs, 1815 &dest->uc_mcontext.fpregs, sfq, dfq); 1816 } 1817 } 1818 1819 void 1820 rwindow_nto32(struct rwindow *src, struct rwindow32 *dest) 1821 { 1822 greg_t *s = (greg_t *)src; 1823 greg32_t *d = (greg32_t *)dest; 1824 int i; 1825 1826 for (i = 0; i < 16; i++) 1827 *d++ = (greg32_t)*s++; 1828 } 1829 1830 void 1831 rwindow_32ton(struct rwindow32 *src, struct rwindow *dest) 1832 { 1833 greg32_t *s = (greg32_t *)src; 1834 greg_t *d = (greg_t *)dest; 1835 int i; 1836 1837 for (i = 0; i < 16; i++) 1838 *d++ = (uint32_t)*s++; 1839 } 1840 1841 #endif /* _SYSCALL32_IMPL */ 1842 1843 /* 1844 * The panic code invokes panic_saveregs() to record the contents of a 1845 * regs structure into the specified panic_data structure for debuggers. 1846 */ 1847 void 1848 panic_saveregs(panic_data_t *pdp, struct regs *rp) 1849 { 1850 panic_nv_t *pnv = PANICNVGET(pdp); 1851 1852 PANICNVADD(pnv, "tstate", rp->r_tstate); 1853 PANICNVADD(pnv, "g1", rp->r_g1); 1854 PANICNVADD(pnv, "g2", rp->r_g2); 1855 PANICNVADD(pnv, "g3", rp->r_g3); 1856 PANICNVADD(pnv, "g4", rp->r_g4); 1857 PANICNVADD(pnv, "g5", rp->r_g5); 1858 PANICNVADD(pnv, "g6", rp->r_g6); 1859 PANICNVADD(pnv, "g7", rp->r_g7); 1860 PANICNVADD(pnv, "o0", rp->r_o0); 1861 PANICNVADD(pnv, "o1", rp->r_o1); 1862 PANICNVADD(pnv, "o2", rp->r_o2); 1863 PANICNVADD(pnv, "o3", rp->r_o3); 1864 PANICNVADD(pnv, "o4", rp->r_o4); 1865 PANICNVADD(pnv, "o5", rp->r_o5); 1866 PANICNVADD(pnv, "o6", rp->r_o6); 1867 PANICNVADD(pnv, "o7", rp->r_o7); 1868 PANICNVADD(pnv, "pc", (ulong_t)rp->r_pc); 1869 PANICNVADD(pnv, "npc", (ulong_t)rp->r_npc); 1870 PANICNVADD(pnv, "y", (uint32_t)rp->r_y); 1871 1872 PANICNVSET(pdp, pnv); 1873 } 1874