1 /*- 2 * Copyright (c) 2003 Peter Wemm. 3 * Copyright (c) 1993 The Regents of the University of California. 4 * All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions 8 * are met: 9 * 1. Redistributions of source code must retain the above copyright 10 * notice, this list of conditions and the following disclaimer. 11 * 2. Redistributions in binary form must reproduce the above copyright 12 * notice, this list of conditions and the following disclaimer in the 13 * documentation and/or other materials provided with the distribution. 14 * 3. Neither the name of the University nor the names of its contributors 15 * may be used to endorse or promote products derived from this software 16 * without specific prior written permission. 17 * 18 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 19 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 20 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 21 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 22 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 23 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 24 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 25 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 26 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 27 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 28 * SUCH DAMAGE. 29 * 30 * $FreeBSD$ 31 */ 32 33 /* 34 * Functions to provide access to special i386 instructions. 35 * This in included in sys/systm.h, and that file should be 36 * used in preference to this. 37 */ 38 39 #ifndef _MACHINE_CPUFUNC_H_ 40 #define _MACHINE_CPUFUNC_H_ 41 42 #ifndef _SYS_CDEFS_H_ 43 #error this file needs sys/cdefs.h as a prerequisite 44 #endif 45 46 struct region_descriptor; 47 48 #define readb(va) (*(volatile uint8_t *) (va)) 49 #define readw(va) (*(volatile uint16_t *) (va)) 50 #define readl(va) (*(volatile uint32_t *) (va)) 51 #define readq(va) (*(volatile uint64_t *) (va)) 52 53 #define writeb(va, d) (*(volatile uint8_t *) (va) = (d)) 54 #define writew(va, d) (*(volatile uint16_t *) (va) = (d)) 55 #define writel(va, d) (*(volatile uint32_t *) (va) = (d)) 56 #define writeq(va, d) (*(volatile uint64_t *) (va) = (d)) 57 58 #if defined(__GNUCLIKE_ASM) && defined(__CC_SUPPORTS___INLINE) 59 60 static __inline void 61 breakpoint(void) 62 { 63 __asm __volatile("int $3"); 64 } 65 66 static __inline __pure2 u_int 67 bsfl(u_int mask) 68 { 69 u_int result; 70 71 __asm __volatile("bsfl %1,%0" : "=r" (result) : "rm" (mask)); 72 return (result); 73 } 74 75 static __inline __pure2 u_long 76 bsfq(u_long mask) 77 { 78 u_long result; 79 80 __asm __volatile("bsfq %1,%0" : "=r" (result) : "rm" (mask)); 81 return (result); 82 } 83 84 static __inline __pure2 u_int 85 bsrl(u_int mask) 86 { 87 u_int result; 88 89 __asm __volatile("bsrl %1,%0" : "=r" (result) : "rm" (mask)); 90 return (result); 91 } 92 93 static __inline __pure2 u_long 94 bsrq(u_long mask) 95 { 96 u_long result; 97 98 __asm __volatile("bsrq %1,%0" : "=r" (result) : "rm" (mask)); 99 return (result); 100 } 101 102 static __inline void 103 clflush(u_long addr) 104 { 105 106 __asm __volatile("clflush %0" : : "m" (*(char *)addr)); 107 } 108 109 static __inline void 110 clflushopt(u_long addr) 111 { 112 113 __asm __volatile(".byte 0x66;clflush %0" : : "m" (*(char *)addr)); 114 } 115 116 static __inline void 117 clts(void) 118 { 119 120 __asm __volatile("clts"); 121 } 122 123 static __inline void 124 disable_intr(void) 125 { 126 __asm __volatile("cli" : : : "memory"); 127 } 128 129 static __inline void 130 do_cpuid(u_int ax, u_int *p) 131 { 132 __asm __volatile("cpuid" 133 : "=a" (p[0]), "=b" (p[1]), "=c" (p[2]), "=d" (p[3]) 134 : "0" (ax)); 135 } 136 137 static __inline void 138 cpuid_count(u_int ax, u_int cx, u_int *p) 139 { 140 __asm __volatile("cpuid" 141 : "=a" (p[0]), "=b" (p[1]), "=c" (p[2]), "=d" (p[3]) 142 : "0" (ax), "c" (cx)); 143 } 144 145 static __inline void 146 enable_intr(void) 147 { 148 __asm __volatile("sti"); 149 } 150 151 #ifdef _KERNEL 152 153 #define HAVE_INLINE_FFS 154 #define ffs(x) __builtin_ffs(x) 155 156 #define HAVE_INLINE_FFSL 157 158 static __inline __pure2 int 159 ffsl(long mask) 160 { 161 return (mask == 0 ? mask : (int)bsfq((u_long)mask) + 1); 162 } 163 164 #define HAVE_INLINE_FFSLL 165 166 static __inline __pure2 int 167 ffsll(long long mask) 168 { 169 return (ffsl((long)mask)); 170 } 171 172 #define HAVE_INLINE_FLS 173 174 static __inline __pure2 int 175 fls(int mask) 176 { 177 return (mask == 0 ? mask : (int)bsrl((u_int)mask) + 1); 178 } 179 180 #define HAVE_INLINE_FLSL 181 182 static __inline __pure2 int 183 flsl(long mask) 184 { 185 return (mask == 0 ? mask : (int)bsrq((u_long)mask) + 1); 186 } 187 188 #define HAVE_INLINE_FLSLL 189 190 static __inline __pure2 int 191 flsll(long long mask) 192 { 193 return (flsl((long)mask)); 194 } 195 196 #endif /* _KERNEL */ 197 198 static __inline void 199 halt(void) 200 { 201 __asm __volatile("hlt"); 202 } 203 204 static __inline u_char 205 inb(u_int port) 206 { 207 u_char data; 208 209 __asm __volatile("inb %w1, %0" : "=a" (data) : "Nd" (port)); 210 return (data); 211 } 212 213 static __inline u_int 214 inl(u_int port) 215 { 216 u_int data; 217 218 __asm __volatile("inl %w1, %0" : "=a" (data) : "Nd" (port)); 219 return (data); 220 } 221 222 static __inline void 223 insb(u_int port, void *addr, size_t count) 224 { 225 __asm __volatile("cld; rep; insb" 226 : "+D" (addr), "+c" (count) 227 : "d" (port) 228 : "memory"); 229 } 230 231 static __inline void 232 insw(u_int port, void *addr, size_t count) 233 { 234 __asm __volatile("cld; rep; insw" 235 : "+D" (addr), "+c" (count) 236 : "d" (port) 237 : "memory"); 238 } 239 240 static __inline void 241 insl(u_int port, void *addr, size_t count) 242 { 243 __asm __volatile("cld; rep; insl" 244 : "+D" (addr), "+c" (count) 245 : "d" (port) 246 : "memory"); 247 } 248 249 static __inline void 250 invd(void) 251 { 252 __asm __volatile("invd"); 253 } 254 255 static __inline u_short 256 inw(u_int port) 257 { 258 u_short data; 259 260 __asm __volatile("inw %w1, %0" : "=a" (data) : "Nd" (port)); 261 return (data); 262 } 263 264 static __inline void 265 outb(u_int port, u_char data) 266 { 267 __asm __volatile("outb %0, %w1" : : "a" (data), "Nd" (port)); 268 } 269 270 static __inline void 271 outl(u_int port, u_int data) 272 { 273 __asm __volatile("outl %0, %w1" : : "a" (data), "Nd" (port)); 274 } 275 276 static __inline void 277 outsb(u_int port, const void *addr, size_t count) 278 { 279 __asm __volatile("cld; rep; outsb" 280 : "+S" (addr), "+c" (count) 281 : "d" (port)); 282 } 283 284 static __inline void 285 outsw(u_int port, const void *addr, size_t count) 286 { 287 __asm __volatile("cld; rep; outsw" 288 : "+S" (addr), "+c" (count) 289 : "d" (port)); 290 } 291 292 static __inline void 293 outsl(u_int port, const void *addr, size_t count) 294 { 295 __asm __volatile("cld; rep; outsl" 296 : "+S" (addr), "+c" (count) 297 : "d" (port)); 298 } 299 300 static __inline void 301 outw(u_int port, u_short data) 302 { 303 __asm __volatile("outw %0, %w1" : : "a" (data), "Nd" (port)); 304 } 305 306 static __inline u_long 307 popcntq(u_long mask) 308 { 309 u_long result; 310 311 __asm __volatile("popcntq %1,%0" : "=r" (result) : "rm" (mask)); 312 return (result); 313 } 314 315 static __inline void 316 lfence(void) 317 { 318 319 __asm __volatile("lfence" : : : "memory"); 320 } 321 322 static __inline void 323 mfence(void) 324 { 325 326 __asm __volatile("mfence" : : : "memory"); 327 } 328 329 static __inline void 330 sfence(void) 331 { 332 333 __asm __volatile("sfence" : : : "memory"); 334 } 335 336 static __inline void 337 ia32_pause(void) 338 { 339 __asm __volatile("pause"); 340 } 341 342 static __inline u_long 343 read_rflags(void) 344 { 345 u_long rf; 346 347 __asm __volatile("pushfq; popq %0" : "=r" (rf)); 348 return (rf); 349 } 350 351 static __inline uint64_t 352 rdmsr(u_int msr) 353 { 354 uint32_t low, high; 355 356 __asm __volatile("rdmsr" : "=a" (low), "=d" (high) : "c" (msr)); 357 return (low | ((uint64_t)high << 32)); 358 } 359 360 static __inline uint32_t 361 rdmsr32(u_int msr) 362 { 363 uint32_t low; 364 365 __asm __volatile("rdmsr" : "=a" (low) : "c" (msr) : "rdx"); 366 return (low); 367 } 368 369 static __inline uint64_t 370 rdpmc(u_int pmc) 371 { 372 uint32_t low, high; 373 374 __asm __volatile("rdpmc" : "=a" (low), "=d" (high) : "c" (pmc)); 375 return (low | ((uint64_t)high << 32)); 376 } 377 378 static __inline uint64_t 379 rdtsc(void) 380 { 381 uint32_t low, high; 382 383 __asm __volatile("rdtsc" : "=a" (low), "=d" (high)); 384 return (low | ((uint64_t)high << 32)); 385 } 386 387 static __inline uint32_t 388 rdtsc32(void) 389 { 390 uint32_t rv; 391 392 __asm __volatile("rdtsc" : "=a" (rv) : : "edx"); 393 return (rv); 394 } 395 396 static __inline void 397 wbinvd(void) 398 { 399 __asm __volatile("wbinvd"); 400 } 401 402 static __inline void 403 write_rflags(u_long rf) 404 { 405 __asm __volatile("pushq %0; popfq" : : "r" (rf)); 406 } 407 408 static __inline void 409 wrmsr(u_int msr, uint64_t newval) 410 { 411 uint32_t low, high; 412 413 low = newval; 414 high = newval >> 32; 415 __asm __volatile("wrmsr" : : "a" (low), "d" (high), "c" (msr)); 416 } 417 418 static __inline void 419 load_cr0(u_long data) 420 { 421 422 __asm __volatile("movq %0,%%cr0" : : "r" (data)); 423 } 424 425 static __inline u_long 426 rcr0(void) 427 { 428 u_long data; 429 430 __asm __volatile("movq %%cr0,%0" : "=r" (data)); 431 return (data); 432 } 433 434 static __inline u_long 435 rcr2(void) 436 { 437 u_long data; 438 439 __asm __volatile("movq %%cr2,%0" : "=r" (data)); 440 return (data); 441 } 442 443 static __inline void 444 load_cr3(u_long data) 445 { 446 447 __asm __volatile("movq %0,%%cr3" : : "r" (data) : "memory"); 448 } 449 450 static __inline u_long 451 rcr3(void) 452 { 453 u_long data; 454 455 __asm __volatile("movq %%cr3,%0" : "=r" (data)); 456 return (data); 457 } 458 459 static __inline void 460 load_cr4(u_long data) 461 { 462 __asm __volatile("movq %0,%%cr4" : : "r" (data)); 463 } 464 465 static __inline u_long 466 rcr4(void) 467 { 468 u_long data; 469 470 __asm __volatile("movq %%cr4,%0" : "=r" (data)); 471 return (data); 472 } 473 474 static __inline u_long 475 rxcr(u_int reg) 476 { 477 u_int low, high; 478 479 __asm __volatile("xgetbv" : "=a" (low), "=d" (high) : "c" (reg)); 480 return (low | ((uint64_t)high << 32)); 481 } 482 483 static __inline void 484 load_xcr(u_int reg, u_long val) 485 { 486 u_int low, high; 487 488 low = val; 489 high = val >> 32; 490 __asm __volatile("xsetbv" : : "c" (reg), "a" (low), "d" (high)); 491 } 492 493 /* 494 * Global TLB flush (except for thise for pages marked PG_G) 495 */ 496 static __inline void 497 invltlb(void) 498 { 499 500 load_cr3(rcr3()); 501 } 502 503 #ifndef CR4_PGE 504 #define CR4_PGE 0x00000080 /* Page global enable */ 505 #endif 506 507 /* 508 * Perform the guaranteed invalidation of all TLB entries. This 509 * includes the global entries, and entries in all PCIDs, not only the 510 * current context. The function works both on non-PCID CPUs and CPUs 511 * with the PCID turned off or on. See IA-32 SDM Vol. 3a 4.10.4.1 512 * Operations that Invalidate TLBs and Paging-Structure Caches. 513 */ 514 static __inline void 515 invltlb_glob(void) 516 { 517 uint64_t cr4; 518 519 cr4 = rcr4(); 520 load_cr4(cr4 & ~CR4_PGE); 521 /* 522 * Although preemption at this point could be detrimental to 523 * performance, it would not lead to an error. PG_G is simply 524 * ignored if CR4.PGE is clear. Moreover, in case this block 525 * is re-entered, the load_cr4() either above or below will 526 * modify CR4.PGE flushing the TLB. 527 */ 528 load_cr4(cr4 | CR4_PGE); 529 } 530 531 /* 532 * TLB flush for an individual page (even if it has PG_G). 533 * Only works on 486+ CPUs (i386 does not have PG_G). 534 */ 535 static __inline void 536 invlpg(u_long addr) 537 { 538 539 __asm __volatile("invlpg %0" : : "m" (*(char *)addr) : "memory"); 540 } 541 542 #define INVPCID_ADDR 0 543 #define INVPCID_CTX 1 544 #define INVPCID_CTXGLOB 2 545 #define INVPCID_ALLCTX 3 546 547 struct invpcid_descr { 548 uint64_t pcid:12 __packed; 549 uint64_t pad:52 __packed; 550 uint64_t addr; 551 } __packed; 552 553 static __inline void 554 invpcid(struct invpcid_descr *d, int type) 555 { 556 557 __asm __volatile("invpcid (%0),%1" 558 : : "r" (d), "r" ((u_long)type) : "memory"); 559 } 560 561 static __inline u_short 562 rfs(void) 563 { 564 u_short sel; 565 __asm __volatile("movw %%fs,%0" : "=rm" (sel)); 566 return (sel); 567 } 568 569 static __inline u_short 570 rgs(void) 571 { 572 u_short sel; 573 __asm __volatile("movw %%gs,%0" : "=rm" (sel)); 574 return (sel); 575 } 576 577 static __inline u_short 578 rss(void) 579 { 580 u_short sel; 581 __asm __volatile("movw %%ss,%0" : "=rm" (sel)); 582 return (sel); 583 } 584 585 static __inline void 586 load_ds(u_short sel) 587 { 588 __asm __volatile("movw %0,%%ds" : : "rm" (sel)); 589 } 590 591 static __inline void 592 load_es(u_short sel) 593 { 594 __asm __volatile("movw %0,%%es" : : "rm" (sel)); 595 } 596 597 static __inline void 598 cpu_monitor(const void *addr, u_long extensions, u_int hints) 599 { 600 601 __asm __volatile("monitor" 602 : : "a" (addr), "c" (extensions), "d" (hints)); 603 } 604 605 static __inline void 606 cpu_mwait(u_long extensions, u_int hints) 607 { 608 609 __asm __volatile("mwait" : : "a" (hints), "c" (extensions)); 610 } 611 612 #ifdef _KERNEL 613 /* This is defined in <machine/specialreg.h> but is too painful to get to */ 614 #ifndef MSR_FSBASE 615 #define MSR_FSBASE 0xc0000100 616 #endif 617 static __inline void 618 load_fs(u_short sel) 619 { 620 /* Preserve the fsbase value across the selector load */ 621 __asm __volatile("rdmsr; movw %0,%%fs; wrmsr" 622 : : "rm" (sel), "c" (MSR_FSBASE) : "eax", "edx"); 623 } 624 625 #ifndef MSR_GSBASE 626 #define MSR_GSBASE 0xc0000101 627 #endif 628 static __inline void 629 load_gs(u_short sel) 630 { 631 /* 632 * Preserve the gsbase value across the selector load. 633 * Note that we have to disable interrupts because the gsbase 634 * being trashed happens to be the kernel gsbase at the time. 635 */ 636 __asm __volatile("pushfq; cli; rdmsr; movw %0,%%gs; wrmsr; popfq" 637 : : "rm" (sel), "c" (MSR_GSBASE) : "eax", "edx"); 638 } 639 #else 640 /* Usable by userland */ 641 static __inline void 642 load_fs(u_short sel) 643 { 644 __asm __volatile("movw %0,%%fs" : : "rm" (sel)); 645 } 646 647 static __inline void 648 load_gs(u_short sel) 649 { 650 __asm __volatile("movw %0,%%gs" : : "rm" (sel)); 651 } 652 #endif 653 654 static __inline uint64_t 655 rdfsbase(void) 656 { 657 uint64_t x; 658 659 __asm __volatile("rdfsbase %0" : "=r" (x)); 660 return (x); 661 } 662 663 static __inline void 664 wrfsbase(uint64_t x) 665 { 666 667 __asm __volatile("wrfsbase %0" : : "r" (x)); 668 } 669 670 static __inline uint64_t 671 rdgsbase(void) 672 { 673 uint64_t x; 674 675 __asm __volatile("rdgsbase %0" : "=r" (x)); 676 return (x); 677 } 678 679 static __inline void 680 wrgsbase(uint64_t x) 681 { 682 683 __asm __volatile("wrgsbase %0" : : "r" (x)); 684 } 685 686 static __inline void 687 bare_lgdt(struct region_descriptor *addr) 688 { 689 __asm __volatile("lgdt (%0)" : : "r" (addr)); 690 } 691 692 static __inline void 693 sgdt(struct region_descriptor *addr) 694 { 695 char *loc; 696 697 loc = (char *)addr; 698 __asm __volatile("sgdt %0" : "=m" (*loc) : : "memory"); 699 } 700 701 static __inline void 702 lidt(struct region_descriptor *addr) 703 { 704 __asm __volatile("lidt (%0)" : : "r" (addr)); 705 } 706 707 static __inline void 708 sidt(struct region_descriptor *addr) 709 { 710 char *loc; 711 712 loc = (char *)addr; 713 __asm __volatile("sidt %0" : "=m" (*loc) : : "memory"); 714 } 715 716 static __inline void 717 lldt(u_short sel) 718 { 719 __asm __volatile("lldt %0" : : "r" (sel)); 720 } 721 722 static __inline void 723 ltr(u_short sel) 724 { 725 __asm __volatile("ltr %0" : : "r" (sel)); 726 } 727 728 static __inline uint32_t 729 read_tr(void) 730 { 731 u_short sel; 732 733 __asm __volatile("str %0" : "=r" (sel)); 734 return (sel); 735 } 736 737 static __inline uint64_t 738 rdr0(void) 739 { 740 uint64_t data; 741 __asm __volatile("movq %%dr0,%0" : "=r" (data)); 742 return (data); 743 } 744 745 static __inline void 746 load_dr0(uint64_t dr0) 747 { 748 __asm __volatile("movq %0,%%dr0" : : "r" (dr0)); 749 } 750 751 static __inline uint64_t 752 rdr1(void) 753 { 754 uint64_t data; 755 __asm __volatile("movq %%dr1,%0" : "=r" (data)); 756 return (data); 757 } 758 759 static __inline void 760 load_dr1(uint64_t dr1) 761 { 762 __asm __volatile("movq %0,%%dr1" : : "r" (dr1)); 763 } 764 765 static __inline uint64_t 766 rdr2(void) 767 { 768 uint64_t data; 769 __asm __volatile("movq %%dr2,%0" : "=r" (data)); 770 return (data); 771 } 772 773 static __inline void 774 load_dr2(uint64_t dr2) 775 { 776 __asm __volatile("movq %0,%%dr2" : : "r" (dr2)); 777 } 778 779 static __inline uint64_t 780 rdr3(void) 781 { 782 uint64_t data; 783 __asm __volatile("movq %%dr3,%0" : "=r" (data)); 784 return (data); 785 } 786 787 static __inline void 788 load_dr3(uint64_t dr3) 789 { 790 __asm __volatile("movq %0,%%dr3" : : "r" (dr3)); 791 } 792 793 static __inline uint64_t 794 rdr6(void) 795 { 796 uint64_t data; 797 __asm __volatile("movq %%dr6,%0" : "=r" (data)); 798 return (data); 799 } 800 801 static __inline void 802 load_dr6(uint64_t dr6) 803 { 804 __asm __volatile("movq %0,%%dr6" : : "r" (dr6)); 805 } 806 807 static __inline uint64_t 808 rdr7(void) 809 { 810 uint64_t data; 811 __asm __volatile("movq %%dr7,%0" : "=r" (data)); 812 return (data); 813 } 814 815 static __inline void 816 load_dr7(uint64_t dr7) 817 { 818 __asm __volatile("movq %0,%%dr7" : : "r" (dr7)); 819 } 820 821 static __inline register_t 822 intr_disable(void) 823 { 824 register_t rflags; 825 826 rflags = read_rflags(); 827 disable_intr(); 828 return (rflags); 829 } 830 831 static __inline void 832 intr_restore(register_t rflags) 833 { 834 write_rflags(rflags); 835 } 836 837 enum { 838 SGX_ECREATE = 0x0, 839 SGX_EADD = 0x1, 840 SGX_EINIT = 0x2, 841 SGX_EREMOVE = 0x3, 842 SGX_EDGBRD = 0x4, 843 SGX_EDGBWR = 0x5, 844 SGX_EEXTEND = 0x6, 845 SGX_ELDU = 0x8, 846 SGX_EBLOCK = 0x9, 847 SGX_EPA = 0xA, 848 SGX_EWB = 0xB, 849 SGX_ETRACK = 0xC, 850 }; 851 852 enum { 853 SGX_PT_SECS = 0x00, 854 SGX_PT_TCS = 0x01, 855 SGX_PT_REG = 0x02, 856 SGX_PT_VA = 0x03, 857 SGX_PT_TRIM = 0x04, 858 }; 859 860 int sgx_encls(uint32_t eax, uint64_t rbx, uint64_t rcx, uint64_t rdx); 861 862 static __inline int 863 sgx_ecreate(void *pginfo, void *secs) 864 { 865 866 return (sgx_encls(SGX_ECREATE, (uint64_t)pginfo, 867 (uint64_t)secs, 0)); 868 } 869 870 static __inline int 871 sgx_eadd(void *pginfo, void *epc) 872 { 873 874 return (sgx_encls(SGX_EADD, (uint64_t)pginfo, 875 (uint64_t)epc, 0)); 876 } 877 878 static __inline int 879 sgx_einit(void *sigstruct, void *secs, void *einittoken) 880 { 881 882 return (sgx_encls(SGX_EINIT, (uint64_t)sigstruct, 883 (uint64_t)secs, (uint64_t)einittoken)); 884 } 885 886 static __inline int 887 sgx_eextend(void *secs, void *epc) 888 { 889 890 return (sgx_encls(SGX_EEXTEND, (uint64_t)secs, 891 (uint64_t)epc, 0)); 892 } 893 894 static __inline int 895 sgx_epa(void *epc) 896 { 897 898 return (sgx_encls(SGX_EPA, SGX_PT_VA, (uint64_t)epc, 0)); 899 } 900 901 static __inline int 902 sgx_eldu(uint64_t rbx, uint64_t rcx, 903 uint64_t rdx) 904 { 905 906 return (sgx_encls(SGX_ELDU, rbx, rcx, rdx)); 907 } 908 909 static __inline int 910 sgx_eremove(void *epc) 911 { 912 913 return (sgx_encls(SGX_EREMOVE, 0, (uint64_t)epc, 0)); 914 } 915 916 #else /* !(__GNUCLIKE_ASM && __CC_SUPPORTS___INLINE) */ 917 918 int breakpoint(void); 919 u_int bsfl(u_int mask); 920 u_int bsrl(u_int mask); 921 void clflush(u_long addr); 922 void clts(void); 923 void cpuid_count(u_int ax, u_int cx, u_int *p); 924 void disable_intr(void); 925 void do_cpuid(u_int ax, u_int *p); 926 void enable_intr(void); 927 void halt(void); 928 void ia32_pause(void); 929 u_char inb(u_int port); 930 u_int inl(u_int port); 931 void insb(u_int port, void *addr, size_t count); 932 void insl(u_int port, void *addr, size_t count); 933 void insw(u_int port, void *addr, size_t count); 934 register_t intr_disable(void); 935 void intr_restore(register_t rf); 936 void invd(void); 937 void invlpg(u_int addr); 938 void invltlb(void); 939 u_short inw(u_int port); 940 void lidt(struct region_descriptor *addr); 941 void lldt(u_short sel); 942 void load_cr0(u_long cr0); 943 void load_cr3(u_long cr3); 944 void load_cr4(u_long cr4); 945 void load_dr0(uint64_t dr0); 946 void load_dr1(uint64_t dr1); 947 void load_dr2(uint64_t dr2); 948 void load_dr3(uint64_t dr3); 949 void load_dr6(uint64_t dr6); 950 void load_dr7(uint64_t dr7); 951 void load_fs(u_short sel); 952 void load_gs(u_short sel); 953 void ltr(u_short sel); 954 void outb(u_int port, u_char data); 955 void outl(u_int port, u_int data); 956 void outsb(u_int port, const void *addr, size_t count); 957 void outsl(u_int port, const void *addr, size_t count); 958 void outsw(u_int port, const void *addr, size_t count); 959 void outw(u_int port, u_short data); 960 u_long rcr0(void); 961 u_long rcr2(void); 962 u_long rcr3(void); 963 u_long rcr4(void); 964 uint64_t rdmsr(u_int msr); 965 uint32_t rdmsr32(u_int msr); 966 uint64_t rdpmc(u_int pmc); 967 uint64_t rdr0(void); 968 uint64_t rdr1(void); 969 uint64_t rdr2(void); 970 uint64_t rdr3(void); 971 uint64_t rdr6(void); 972 uint64_t rdr7(void); 973 uint64_t rdtsc(void); 974 u_long read_rflags(void); 975 u_int rfs(void); 976 u_int rgs(void); 977 void wbinvd(void); 978 void write_rflags(u_int rf); 979 void wrmsr(u_int msr, uint64_t newval); 980 981 #endif /* __GNUCLIKE_ASM && __CC_SUPPORTS___INLINE */ 982 983 void reset_dbregs(void); 984 985 #ifdef _KERNEL 986 int rdmsr_safe(u_int msr, uint64_t *val); 987 int wrmsr_safe(u_int msr, uint64_t newval); 988 #endif 989 990 #endif /* !_MACHINE_CPUFUNC_H_ */ 991