1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License, Version 1.0 only 6 * (the "License"). You may not use this file except in compliance 7 * with the License. 8 * 9 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 10 * or http://www.opensolaris.org/os/licensing. 11 * See the License for the specific language governing permissions 12 * and limitations under the License. 13 * 14 * When distributing Covered Code, include this CDDL HEADER in each 15 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 16 * If applicable, add the following below this CDDL HEADER, with the 17 * fields enclosed by brackets "[]" replaced with your own identifying 18 * information: Portions Copyright [yyyy] [name of copyright owner] 19 * 20 * CDDL HEADER END 21 * 22 * $FreeBSD$ 23 */ 24 /* 25 * Copyright 2005 Sun Microsystems, Inc. All rights reserved. 26 * Use is subject to license terms. 27 */ 28 #include <sys/cdefs.h> 29 30 #include <sys/param.h> 31 #include <sys/systm.h> 32 #include <sys/kernel.h> 33 #include <sys/stack.h> 34 #include <sys/pcpu.h> 35 36 #include <machine/frame.h> 37 #include <machine/md_var.h> 38 #include <machine/reg.h> 39 #include <machine/stack.h> 40 #include <x86/ifunc.h> 41 42 #include <vm/vm.h> 43 #include <vm/vm_param.h> 44 #include <vm/pmap.h> 45 46 #include "regset.h" 47 48 uint8_t dtrace_fuword8_nocheck(void *); 49 uint16_t dtrace_fuword16_nocheck(void *); 50 uint32_t dtrace_fuword32_nocheck(void *); 51 uint64_t dtrace_fuword64_nocheck(void *); 52 53 int dtrace_ustackdepth_max = 2048; 54 55 void 56 dtrace_getpcstack(pc_t *pcstack, int pcstack_limit, int aframes, 57 uint32_t *intrpc) 58 { 59 int depth = 0; 60 register_t rbp; 61 struct amd64_frame *frame; 62 vm_offset_t callpc; 63 pc_t caller = (pc_t) solaris_cpu[curcpu].cpu_dtrace_caller; 64 65 if (intrpc != 0) 66 pcstack[depth++] = (pc_t) intrpc; 67 68 aframes++; 69 70 __asm __volatile("movq %%rbp,%0" : "=r" (rbp)); 71 72 frame = (struct amd64_frame *)rbp; 73 while (depth < pcstack_limit) { 74 if (!INKERNEL((long) frame)) 75 break; 76 77 callpc = frame->f_retaddr; 78 79 if (!INKERNEL(callpc)) 80 break; 81 82 if (aframes > 0) { 83 aframes--; 84 if ((aframes == 0) && (caller != 0)) { 85 pcstack[depth++] = caller; 86 } 87 } 88 else { 89 pcstack[depth++] = callpc; 90 } 91 92 if (frame->f_frame <= frame || 93 (vm_offset_t)frame->f_frame >= curthread->td_kstack + 94 curthread->td_kstack_pages * PAGE_SIZE) 95 break; 96 frame = frame->f_frame; 97 } 98 99 for (; depth < pcstack_limit; depth++) { 100 pcstack[depth] = 0; 101 } 102 } 103 104 static int 105 dtrace_getustack_common(uint64_t *pcstack, int pcstack_limit, uintptr_t pc, 106 uintptr_t sp) 107 { 108 uintptr_t oldsp; 109 volatile uint16_t *flags = 110 (volatile uint16_t *)&cpu_core[curcpu].cpuc_dtrace_flags; 111 int ret = 0; 112 113 ASSERT(pcstack == NULL || pcstack_limit > 0); 114 ASSERT(dtrace_ustackdepth_max > 0); 115 116 while (pc != 0) { 117 /* 118 * We limit the number of times we can go around this 119 * loop to account for a circular stack. 120 */ 121 if (ret++ >= dtrace_ustackdepth_max) { 122 *flags |= CPU_DTRACE_BADSTACK; 123 cpu_core[curcpu].cpuc_dtrace_illval = sp; 124 break; 125 } 126 127 if (pcstack != NULL) { 128 *pcstack++ = (uint64_t)pc; 129 pcstack_limit--; 130 if (pcstack_limit <= 0) 131 break; 132 } 133 134 if (sp == 0) 135 break; 136 137 oldsp = sp; 138 139 pc = dtrace_fuword64((void *)(sp + 140 offsetof(struct amd64_frame, f_retaddr))); 141 sp = dtrace_fuword64((void *)sp); 142 143 if (sp == oldsp) { 144 *flags |= CPU_DTRACE_BADSTACK; 145 cpu_core[curcpu].cpuc_dtrace_illval = sp; 146 break; 147 } 148 149 /* 150 * This is totally bogus: if we faulted, we're going to clear 151 * the fault and break. This is to deal with the apparently 152 * broken Java stacks on x86. 153 */ 154 if (*flags & CPU_DTRACE_FAULT) { 155 *flags &= ~CPU_DTRACE_FAULT; 156 break; 157 } 158 } 159 160 return (ret); 161 } 162 163 void 164 dtrace_getupcstack(uint64_t *pcstack, int pcstack_limit) 165 { 166 proc_t *p = curproc; 167 struct trapframe *tf; 168 uintptr_t pc, sp, fp; 169 volatile uint16_t *flags = 170 (volatile uint16_t *)&cpu_core[curcpu].cpuc_dtrace_flags; 171 int n; 172 173 if (*flags & CPU_DTRACE_FAULT) 174 return; 175 176 if (pcstack_limit <= 0) 177 return; 178 179 /* 180 * If there's no user context we still need to zero the stack. 181 */ 182 if (p == NULL || (tf = curthread->td_frame) == NULL) 183 goto zero; 184 185 *pcstack++ = (uint64_t)p->p_pid; 186 pcstack_limit--; 187 188 if (pcstack_limit <= 0) 189 return; 190 191 pc = tf->tf_rip; 192 fp = tf->tf_rbp; 193 sp = tf->tf_rsp; 194 195 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_ENTRY)) { 196 /* 197 * In an entry probe. The frame pointer has not yet been 198 * pushed (that happens in the function prologue). The 199 * best approach is to add the current pc as a missing top 200 * of stack and back the pc up to the caller, which is stored 201 * at the current stack pointer address since the call 202 * instruction puts it there right before the branch. 203 */ 204 205 *pcstack++ = (uint64_t)pc; 206 pcstack_limit--; 207 if (pcstack_limit <= 0) 208 return; 209 210 pc = dtrace_fuword64((void *) sp); 211 } 212 213 n = dtrace_getustack_common(pcstack, pcstack_limit, pc, fp); 214 ASSERT(n >= 0); 215 ASSERT(n <= pcstack_limit); 216 217 pcstack += n; 218 pcstack_limit -= n; 219 220 zero: 221 while (pcstack_limit-- > 0) 222 *pcstack++ = 0; 223 } 224 225 int 226 dtrace_getustackdepth(void) 227 { 228 proc_t *p = curproc; 229 struct trapframe *tf; 230 uintptr_t pc, fp, sp; 231 int n = 0; 232 233 if (p == NULL || (tf = curthread->td_frame) == NULL) 234 return (0); 235 236 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_FAULT)) 237 return (-1); 238 239 pc = tf->tf_rip; 240 fp = tf->tf_rbp; 241 sp = tf->tf_rsp; 242 243 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_ENTRY)) { 244 /* 245 * In an entry probe. The frame pointer has not yet been 246 * pushed (that happens in the function prologue). The 247 * best approach is to add the current pc as a missing top 248 * of stack and back the pc up to the caller, which is stored 249 * at the current stack pointer address since the call 250 * instruction puts it there right before the branch. 251 */ 252 253 pc = dtrace_fuword64((void *) sp); 254 n++; 255 } 256 257 n += dtrace_getustack_common(NULL, 0, pc, fp); 258 259 return (n); 260 } 261 262 void 263 dtrace_getufpstack(uint64_t *pcstack, uint64_t *fpstack, int pcstack_limit) 264 { 265 proc_t *p = curproc; 266 struct trapframe *tf; 267 uintptr_t pc, sp, fp; 268 volatile uint16_t *flags = 269 (volatile uint16_t *)&cpu_core[curcpu].cpuc_dtrace_flags; 270 #ifdef notyet /* XXX signal stack */ 271 uintptr_t oldcontext; 272 size_t s1, s2; 273 #endif 274 275 if (*flags & CPU_DTRACE_FAULT) 276 return; 277 278 if (pcstack_limit <= 0) 279 return; 280 281 /* 282 * If there's no user context we still need to zero the stack. 283 */ 284 if (p == NULL || (tf = curthread->td_frame) == NULL) 285 goto zero; 286 287 *pcstack++ = (uint64_t)p->p_pid; 288 pcstack_limit--; 289 290 if (pcstack_limit <= 0) 291 return; 292 293 pc = tf->tf_rip; 294 sp = tf->tf_rsp; 295 fp = tf->tf_rbp; 296 297 #ifdef notyet /* XXX signal stack */ 298 oldcontext = lwp->lwp_oldcontext; 299 s1 = sizeof (struct xframe) + 2 * sizeof (long); 300 s2 = s1 + sizeof (siginfo_t); 301 #endif 302 303 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_ENTRY)) { 304 *pcstack++ = (uint64_t)pc; 305 *fpstack++ = 0; 306 pcstack_limit--; 307 if (pcstack_limit <= 0) 308 return; 309 310 pc = dtrace_fuword64((void *)sp); 311 } 312 313 while (pc != 0) { 314 *pcstack++ = (uint64_t)pc; 315 *fpstack++ = fp; 316 pcstack_limit--; 317 if (pcstack_limit <= 0) 318 break; 319 320 if (fp == 0) 321 break; 322 323 #ifdef notyet /* XXX signal stack */ 324 if (oldcontext == sp + s1 || oldcontext == sp + s2) { 325 ucontext_t *ucp = (ucontext_t *)oldcontext; 326 greg_t *gregs = ucp->uc_mcontext.gregs; 327 328 sp = dtrace_fulword(&gregs[REG_FP]); 329 pc = dtrace_fulword(&gregs[REG_PC]); 330 331 oldcontext = dtrace_fulword(&ucp->uc_link); 332 } else 333 #endif /* XXX */ 334 { 335 pc = dtrace_fuword64((void *)(fp + 336 offsetof(struct amd64_frame, f_retaddr))); 337 fp = dtrace_fuword64((void *)fp); 338 } 339 340 /* 341 * This is totally bogus: if we faulted, we're going to clear 342 * the fault and break. This is to deal with the apparently 343 * broken Java stacks on x86. 344 */ 345 if (*flags & CPU_DTRACE_FAULT) { 346 *flags &= ~CPU_DTRACE_FAULT; 347 break; 348 } 349 } 350 351 zero: 352 while (pcstack_limit-- > 0) 353 *pcstack++ = 0; 354 } 355 356 /*ARGSUSED*/ 357 uint64_t 358 dtrace_getarg(int arg, int aframes) 359 { 360 uintptr_t val; 361 struct amd64_frame *fp = (struct amd64_frame *)dtrace_getfp(); 362 uintptr_t *stack; 363 int i; 364 365 /* 366 * A total of 6 arguments are passed via registers; any argument with 367 * index of 5 or lower is therefore in a register. 368 */ 369 int inreg = 5; 370 371 for (i = 1; i <= aframes; i++) { 372 fp = fp->f_frame; 373 374 if (P2ROUNDUP(fp->f_retaddr, 16) == 375 (long)dtrace_invop_callsite) { 376 /* 377 * In the case of amd64, we will use the pointer to the 378 * regs structure that was pushed when we took the 379 * trap. To get this structure, we must increment 380 * beyond the frame structure, and then again beyond 381 * the calling RIP stored in dtrace_invop(). If the 382 * argument that we're seeking is passed on the stack, 383 * we'll pull the true stack pointer out of the saved 384 * registers and decrement our argument by the number 385 * of arguments passed in registers; if the argument 386 * we're seeking is passed in registers, we can just 387 * load it directly. 388 */ 389 struct trapframe *tf = (struct trapframe *)&fp[1]; 390 391 if (arg <= inreg) { 392 switch (arg) { 393 case 0: 394 stack = (uintptr_t *)&tf->tf_rdi; 395 break; 396 case 1: 397 stack = (uintptr_t *)&tf->tf_rsi; 398 break; 399 case 2: 400 stack = (uintptr_t *)&tf->tf_rdx; 401 break; 402 case 3: 403 stack = (uintptr_t *)&tf->tf_rcx; 404 break; 405 case 4: 406 stack = (uintptr_t *)&tf->tf_r8; 407 break; 408 case 5: 409 stack = (uintptr_t *)&tf->tf_r9; 410 break; 411 } 412 arg = 0; 413 } else { 414 stack = (uintptr_t *)(tf->tf_rsp); 415 arg -= inreg; 416 } 417 goto load; 418 } 419 420 } 421 422 /* 423 * We know that we did not come through a trap to get into 424 * dtrace_probe() -- the provider simply called dtrace_probe() 425 * directly. As this is the case, we need to shift the argument 426 * that we're looking for: the probe ID is the first argument to 427 * dtrace_probe(), so the argument n will actually be found where 428 * one would expect to find argument (n + 1). 429 */ 430 arg++; 431 432 if (arg <= inreg) { 433 /* 434 * This shouldn't happen. If the argument is passed in a 435 * register then it should have been, well, passed in a 436 * register... 437 */ 438 DTRACE_CPUFLAG_SET(CPU_DTRACE_ILLOP); 439 return (0); 440 } 441 442 arg -= (inreg + 1); 443 stack = (uintptr_t *)&fp[1]; 444 445 load: 446 DTRACE_CPUFLAG_SET(CPU_DTRACE_NOFAULT); 447 val = stack[arg]; 448 DTRACE_CPUFLAG_CLEAR(CPU_DTRACE_NOFAULT); 449 450 return (val); 451 } 452 453 int 454 dtrace_getstackdepth(int aframes) 455 { 456 int depth = 0; 457 struct amd64_frame *frame; 458 vm_offset_t rbp; 459 460 aframes++; 461 rbp = dtrace_getfp(); 462 frame = (struct amd64_frame *)rbp; 463 depth++; 464 for(;;) { 465 if (!INKERNEL((long) frame)) 466 break; 467 if (!INKERNEL((long) frame->f_frame)) 468 break; 469 depth++; 470 if (frame->f_frame <= frame || 471 (vm_offset_t)frame->f_frame >= curthread->td_kstack + 472 curthread->td_kstack_pages * PAGE_SIZE) 473 break; 474 frame = frame->f_frame; 475 } 476 if (depth < aframes) 477 return 0; 478 else 479 return depth - aframes; 480 } 481 482 ulong_t 483 dtrace_getreg(struct trapframe *rp, uint_t reg) 484 { 485 /* This table is dependent on reg.d. */ 486 int regmap[] = { 487 REG_GS, /* 0 GS */ 488 REG_FS, /* 1 FS */ 489 REG_ES, /* 2 ES */ 490 REG_DS, /* 3 DS */ 491 REG_RDI, /* 4 EDI */ 492 REG_RSI, /* 5 ESI */ 493 REG_RBP, /* 6 EBP, REG_FP */ 494 REG_RSP, /* 7 ESP */ 495 REG_RBX, /* 8 EBX, REG_R1 */ 496 REG_RDX, /* 9 EDX */ 497 REG_RCX, /* 10 ECX */ 498 REG_RAX, /* 11 EAX, REG_R0 */ 499 REG_TRAPNO, /* 12 TRAPNO */ 500 REG_ERR, /* 13 ERR */ 501 REG_RIP, /* 14 EIP, REG_PC */ 502 REG_CS, /* 15 CS */ 503 REG_RFL, /* 16 EFL, REG_PS */ 504 REG_RSP, /* 17 UESP, REG_SP */ 505 REG_SS /* 18 SS */ 506 }; 507 508 #ifdef illumos 509 if (reg <= SS) { 510 #else /* !illumos */ 511 if (reg <= GS) { 512 #endif 513 if (reg >= sizeof (regmap) / sizeof (int)) { 514 DTRACE_CPUFLAG_SET(CPU_DTRACE_ILLOP); 515 return (0); 516 } 517 518 reg = regmap[reg]; 519 } else { 520 /* This is dependent on reg.d. */ 521 #ifdef illumos 522 reg -= SS + 1; 523 #else /* !illumos */ 524 reg -= GS + 1; 525 #endif 526 } 527 528 switch (reg) { 529 case REG_RDI: 530 return (rp->tf_rdi); 531 case REG_RSI: 532 return (rp->tf_rsi); 533 case REG_RDX: 534 return (rp->tf_rdx); 535 case REG_RCX: 536 return (rp->tf_rcx); 537 case REG_R8: 538 return (rp->tf_r8); 539 case REG_R9: 540 return (rp->tf_r9); 541 case REG_RAX: 542 return (rp->tf_rax); 543 case REG_RBX: 544 return (rp->tf_rbx); 545 case REG_RBP: 546 return (rp->tf_rbp); 547 case REG_R10: 548 return (rp->tf_r10); 549 case REG_R11: 550 return (rp->tf_r11); 551 case REG_R12: 552 return (rp->tf_r12); 553 case REG_R13: 554 return (rp->tf_r13); 555 case REG_R14: 556 return (rp->tf_r14); 557 case REG_R15: 558 return (rp->tf_r15); 559 case REG_DS: 560 return (rp->tf_ds); 561 case REG_ES: 562 return (rp->tf_es); 563 case REG_FS: 564 return (rp->tf_fs); 565 case REG_GS: 566 return (rp->tf_gs); 567 case REG_TRAPNO: 568 return (rp->tf_trapno); 569 case REG_ERR: 570 return (rp->tf_err); 571 case REG_RIP: 572 return (rp->tf_rip); 573 case REG_CS: 574 return (rp->tf_cs); 575 case REG_SS: 576 return (rp->tf_ss); 577 case REG_RFL: 578 return (rp->tf_rflags); 579 case REG_RSP: 580 return (rp->tf_rsp); 581 default: 582 DTRACE_CPUFLAG_SET(CPU_DTRACE_ILLOP); 583 return (0); 584 } 585 } 586 587 static int 588 dtrace_copycheck(uintptr_t uaddr, uintptr_t kaddr, size_t size) 589 { 590 ASSERT(INKERNEL(kaddr) && kaddr + size >= kaddr); 591 592 if (uaddr + size > VM_MAXUSER_ADDRESS || uaddr + size < uaddr) { 593 DTRACE_CPUFLAG_SET(CPU_DTRACE_BADADDR); 594 cpu_core[curcpu].cpuc_dtrace_illval = uaddr; 595 return (0); 596 } 597 598 return (1); 599 } 600 601 void 602 dtrace_copyin(uintptr_t uaddr, uintptr_t kaddr, size_t size, 603 volatile uint16_t *flags) 604 { 605 if (dtrace_copycheck(uaddr, kaddr, size)) 606 dtrace_copy(uaddr, kaddr, size); 607 } 608 609 void 610 dtrace_copyout(uintptr_t kaddr, uintptr_t uaddr, size_t size, 611 volatile uint16_t *flags) 612 { 613 if (dtrace_copycheck(uaddr, kaddr, size)) 614 dtrace_copy(kaddr, uaddr, size); 615 } 616 617 void 618 dtrace_copyinstr(uintptr_t uaddr, uintptr_t kaddr, size_t size, 619 volatile uint16_t *flags) 620 { 621 if (dtrace_copycheck(uaddr, kaddr, size)) 622 dtrace_copystr(uaddr, kaddr, size, flags); 623 } 624 625 void 626 dtrace_copyoutstr(uintptr_t kaddr, uintptr_t uaddr, size_t size, 627 volatile uint16_t *flags) 628 { 629 if (dtrace_copycheck(uaddr, kaddr, size)) 630 dtrace_copystr(kaddr, uaddr, size, flags); 631 } 632 633 uint8_t 634 dtrace_fuword8(void *uaddr) 635 { 636 if ((uintptr_t)uaddr > VM_MAXUSER_ADDRESS) { 637 DTRACE_CPUFLAG_SET(CPU_DTRACE_BADADDR); 638 cpu_core[curcpu].cpuc_dtrace_illval = (uintptr_t)uaddr; 639 return (0); 640 } 641 return (dtrace_fuword8_nocheck(uaddr)); 642 } 643 644 uint16_t 645 dtrace_fuword16(void *uaddr) 646 { 647 if ((uintptr_t)uaddr > VM_MAXUSER_ADDRESS) { 648 DTRACE_CPUFLAG_SET(CPU_DTRACE_BADADDR); 649 cpu_core[curcpu].cpuc_dtrace_illval = (uintptr_t)uaddr; 650 return (0); 651 } 652 return (dtrace_fuword16_nocheck(uaddr)); 653 } 654 655 uint32_t 656 dtrace_fuword32(void *uaddr) 657 { 658 if ((uintptr_t)uaddr > VM_MAXUSER_ADDRESS) { 659 DTRACE_CPUFLAG_SET(CPU_DTRACE_BADADDR); 660 cpu_core[curcpu].cpuc_dtrace_illval = (uintptr_t)uaddr; 661 return (0); 662 } 663 return (dtrace_fuword32_nocheck(uaddr)); 664 } 665 666 uint64_t 667 dtrace_fuword64(void *uaddr) 668 { 669 if ((uintptr_t)uaddr > VM_MAXUSER_ADDRESS) { 670 DTRACE_CPUFLAG_SET(CPU_DTRACE_BADADDR); 671 cpu_core[curcpu].cpuc_dtrace_illval = (uintptr_t)uaddr; 672 return (0); 673 } 674 return (dtrace_fuword64_nocheck(uaddr)); 675 } 676 677 /* 678 * ifunc resolvers for SMAP support 679 */ 680 void dtrace_copy_nosmap(uintptr_t, uintptr_t, size_t); 681 void dtrace_copy_smap(uintptr_t, uintptr_t, size_t); 682 DEFINE_IFUNC(, void, dtrace_copy, (uintptr_t, uintptr_t, size_t)) 683 { 684 685 return ((cpu_stdext_feature & CPUID_STDEXT_SMAP) != 0 ? 686 dtrace_copy_smap : dtrace_copy_nosmap); 687 } 688 689 void dtrace_copystr_nosmap(uintptr_t, uintptr_t, size_t, volatile uint16_t *); 690 void dtrace_copystr_smap(uintptr_t, uintptr_t, size_t, volatile uint16_t *); 691 DEFINE_IFUNC(, void, dtrace_copystr, (uintptr_t, uintptr_t, size_t, 692 volatile uint16_t *)) 693 { 694 695 return ((cpu_stdext_feature & CPUID_STDEXT_SMAP) != 0 ? 696 dtrace_copystr_smap : dtrace_copystr_nosmap); 697 } 698 699 uintptr_t dtrace_fulword_nosmap(void *); 700 uintptr_t dtrace_fulword_smap(void *); 701 DEFINE_IFUNC(, uintptr_t, dtrace_fulword, (void *)) 702 { 703 704 return ((cpu_stdext_feature & CPUID_STDEXT_SMAP) != 0 ? 705 dtrace_fulword_smap : dtrace_fulword_nosmap); 706 } 707 708 uint8_t dtrace_fuword8_nocheck_nosmap(void *); 709 uint8_t dtrace_fuword8_nocheck_smap(void *); 710 DEFINE_IFUNC(, uint8_t, dtrace_fuword8_nocheck, (void *)) 711 { 712 713 return ((cpu_stdext_feature & CPUID_STDEXT_SMAP) != 0 ? 714 dtrace_fuword8_nocheck_smap : dtrace_fuword8_nocheck_nosmap); 715 } 716 717 uint16_t dtrace_fuword16_nocheck_nosmap(void *); 718 uint16_t dtrace_fuword16_nocheck_smap(void *); 719 DEFINE_IFUNC(, uint16_t, dtrace_fuword16_nocheck, (void *)) 720 { 721 722 return ((cpu_stdext_feature & CPUID_STDEXT_SMAP) != 0 ? 723 dtrace_fuword16_nocheck_smap : dtrace_fuword16_nocheck_nosmap); 724 } 725 726 uint32_t dtrace_fuword32_nocheck_nosmap(void *); 727 uint32_t dtrace_fuword32_nocheck_smap(void *); 728 DEFINE_IFUNC(, uint32_t, dtrace_fuword32_nocheck, (void *)) 729 { 730 731 return ((cpu_stdext_feature & CPUID_STDEXT_SMAP) != 0 ? 732 dtrace_fuword32_nocheck_smap : dtrace_fuword32_nocheck_nosmap); 733 } 734 735 uint64_t dtrace_fuword64_nocheck_nosmap(void *); 736 uint64_t dtrace_fuword64_nocheck_smap(void *); 737 DEFINE_IFUNC(, uint64_t, dtrace_fuword64_nocheck, (void *)) 738 { 739 740 return ((cpu_stdext_feature & CPUID_STDEXT_SMAP) != 0 ? 741 dtrace_fuword64_nocheck_smap : dtrace_fuword64_nocheck_nosmap); 742 } 743