1e1743d02SSøren Schmidt /*- 28a36da99SPedro F. Giffuni * SPDX-License-Identifier: BSD-3-Clause 38a36da99SPedro F. Giffuni * 486be94fcSTycho Nightingale * Copyright (c) 2017 Dell EMC 5455d3589SDavid E. O'Brien * Copyright (c) 2000-2001, 2003 David O'Brien 69a14aa01SUlrich Spörlein * Copyright (c) 1995-1996 Søren Schmidt 7e1743d02SSøren Schmidt * Copyright (c) 1996 Peter Wemm 8e1743d02SSøren Schmidt * All rights reserved. 9e1743d02SSøren Schmidt * 10e1743d02SSøren Schmidt * Redistribution and use in source and binary forms, with or without 11e1743d02SSøren Schmidt * modification, are permitted provided that the following conditions 12e1743d02SSøren Schmidt * are met: 13e1743d02SSøren Schmidt * 1. Redistributions of source code must retain the above copyright 14e1743d02SSøren Schmidt * notice, this list of conditions and the following disclaimer 15e1743d02SSøren Schmidt * in this position and unchanged. 16e1743d02SSøren Schmidt * 2. Redistributions in binary form must reproduce the above copyright 17e1743d02SSøren Schmidt * notice, this list of conditions and the following disclaimer in the 18e1743d02SSøren Schmidt * documentation and/or other materials provided with the distribution. 19e1743d02SSøren Schmidt * 3. The name of the author may not be used to endorse or promote products 2021dc7d4fSJens Schweikhardt * derived from this software without specific prior written permission 21e1743d02SSøren Schmidt * 22e1743d02SSøren Schmidt * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 23e1743d02SSøren Schmidt * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 24e1743d02SSøren Schmidt * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 25e1743d02SSøren Schmidt * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 26e1743d02SSøren Schmidt * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 27e1743d02SSøren Schmidt * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 28e1743d02SSøren Schmidt * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 29e1743d02SSøren Schmidt * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 30e1743d02SSøren Schmidt * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 31e1743d02SSøren Schmidt * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 32e1743d02SSøren Schmidt */ 33e1743d02SSøren Schmidt 3412bc222eSJonathan Anderson #include "opt_capsicum.h" 3562919d78SPeter Wemm 36e1743d02SSøren Schmidt #include <sys/param.h> 374a144410SRobert Watson #include <sys/capsicum.h> 3878f57a9cSMark Johnston #include <sys/compressor.h> 39e1743d02SSøren Schmidt #include <sys/exec.h> 408c64af4fSJohn Polstra #include <sys/fcntl.h> 41e1743d02SSøren Schmidt #include <sys/imgact.h> 42e1743d02SSøren Schmidt #include <sys/imgact_elf.h> 43b96bd95bSIan Lepore #include <sys/jail.h> 44e1743d02SSøren Schmidt #include <sys/kernel.h> 45f34fa851SJohn Baldwin #include <sys/lock.h> 46e1743d02SSøren Schmidt #include <sys/malloc.h> 4768ff2a43SChristian S.J. Peron #include <sys/mount.h> 488c64af4fSJohn Polstra #include <sys/mman.h> 49a794e791SBruce Evans #include <sys/namei.h> 50a794e791SBruce Evans #include <sys/proc.h> 518c64af4fSJohn Polstra #include <sys/procfs.h> 5286be94fcSTycho Nightingale #include <sys/ptrace.h> 531ba5ad42SEdward Tomasz Napierala #include <sys/racct.h> 54b7924341SAndrew Turner #include <sys/reg.h> 558c64af4fSJohn Polstra #include <sys/resourcevar.h> 5689f6b863SAttilio Rao #include <sys/rwlock.h> 57bd390213SMikolaj Golub #include <sys/sbuf.h> 58da61b9a6SAlan Cox #include <sys/sf_buf.h> 59ee235befSKonstantin Belousov #include <sys/smp.h> 6036240ea5SDoug Rabson #include <sys/systm.h> 61e1743d02SSøren Schmidt #include <sys/signalvar.h> 628c64af4fSJohn Polstra #include <sys/stat.h> 631005a129SJohn Baldwin #include <sys/sx.h> 648c64af4fSJohn Polstra #include <sys/syscall.h> 65e1743d02SSøren Schmidt #include <sys/sysctl.h> 668c64af4fSJohn Polstra #include <sys/sysent.h> 67a794e791SBruce Evans #include <sys/vnode.h> 68e7228204SAlfred Perlstein #include <sys/syslog.h> 69e7228204SAlfred Perlstein #include <sys/eventhandler.h> 70f1fca82eSMikolaj Golub #include <sys/user.h> 71e7228204SAlfred Perlstein 72e1743d02SSøren Schmidt #include <vm/vm.h> 73e1743d02SSøren Schmidt #include <vm/vm_kern.h> 74e1743d02SSøren Schmidt #include <vm/vm_param.h> 75e1743d02SSøren Schmidt #include <vm/pmap.h> 76e1743d02SSøren Schmidt #include <vm/vm_map.h> 770ff27d31SJohn Polstra #include <vm/vm_object.h> 78e1743d02SSøren Schmidt #include <vm/vm_extern.h> 79e1743d02SSøren Schmidt 8052c24af7SPeter Wemm #include <machine/elf.h> 81e1743d02SSøren Schmidt #include <machine/md_var.h> 82e1743d02SSøren Schmidt 831b8388cdSMikolaj Golub #define ELF_NOTE_ROUNDSIZE 4 84c815a20cSDavid E. O'Brien #define OLD_EI_BRAND 8 85c815a20cSDavid E. O'Brien 863ebc1248SPeter Wemm static int __elfN(check_header)(const Elf_Ehdr *hdr); 8732c01de2SDmitry Chagin static Elf_Brandinfo *__elfN(get_brandinfo)(struct image_params *imgp, 8809c78d53SEdward Tomasz Napierala const char *interp, int32_t *osrel, uint32_t *fctl0); 893ebc1248SPeter Wemm static int __elfN(load_file)(struct proc *p, const char *file, u_long *addr, 901699546dSEdward Tomasz Napierala u_long *entry); 91169641f7SAlex Richardson static int __elfN(load_section)(const struct image_params *imgp, 92169641f7SAlex Richardson vm_ooffset_t offset, caddr_t vmaddr, size_t memsz, size_t filsz, 93169641f7SAlex Richardson vm_prot_t prot); 943ebc1248SPeter Wemm static int __CONCAT(exec_, __elfN(imgact))(struct image_params *imgp); 95a95659f7SEd Maste static bool __elfN(freebsd_trans_osrel)(const Elf_Note *note, 9689ffc202SBjoern A. Zeeb int32_t *osrel); 97a95659f7SEd Maste static bool kfreebsd_trans_osrel(const Elf_Note *note, int32_t *osrel); 9819621645SAlex Richardson static bool __elfN(check_note)(struct image_params *imgp, 9919621645SAlex Richardson Elf_Brandnote *checknote, int32_t *osrel, bool *has_fctl0, 1000cad2aa2SKonstantin Belousov uint32_t *fctl0); 101ed167eaaSKonstantin Belousov static vm_prot_t __elfN(trans_prot)(Elf_Word); 102ed167eaaSKonstantin Belousov static Elf_Word __elfN(untrans_prot)(vm_prot_t); 1036b71405bSJohn Baldwin static size_t __elfN(prepare_register_notes)(struct thread *td, 1046b71405bSJohn Baldwin struct note_info_list *list, struct thread *target_td); 105e1743d02SSøren Schmidt 1067029da5cSPawel Biernacki SYSCTL_NODE(_kern, OID_AUTO, __CONCAT(elf, __ELF_WORD_SIZE), 1077029da5cSPawel Biernacki CTLFLAG_RW | CTLFLAG_MPSAFE, 0, 108a360a43dSJake Burkholder ""); 109a360a43dSJake Burkholder 110e548a1d4SJake Burkholder int __elfN(fallback_brand) = -1; 111e548a1d4SJake Burkholder SYSCTL_INT(__CONCAT(_kern_elf, __ELF_WORD_SIZE), OID_AUTO, 112af3b2549SHans Petter Selasky fallback_brand, CTLFLAG_RWTUN, &__elfN(fallback_brand), 0, 113a360a43dSJake Burkholder __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) " brand of last resort"); 114a360a43dSJake Burkholder 115551d79e1SMarcel Moolenaar static int elf_legacy_coredump = 0; 116a360a43dSJake Burkholder SYSCTL_INT(_debug, OID_AUTO, __elfN(legacy_coredump), CTLFLAG_RW, 1171cbb879dSEd Maste &elf_legacy_coredump, 0, 1181cbb879dSEd Maste "include all and only RW pages in core dumps"); 119e1743d02SSøren Schmidt 12062c625fdSKonstantin Belousov int __elfN(nxstack) = 1214d22d07aSKonstantin Belousov #if defined(__amd64__) || defined(__powerpc64__) /* both 64 and 32 bit */ || \ 122*d29771a7SAndrew Turner defined(__arm__) || defined(__aarch64__) || \ 1234bf4b0f1SJohn Baldwin defined(__riscv) 12462c625fdSKonstantin Belousov 1; 12562c625fdSKonstantin Belousov #else 12662c625fdSKonstantin Belousov 0; 12762c625fdSKonstantin Belousov #endif 128291c06a1SKonstantin Belousov SYSCTL_INT(__CONCAT(_kern_elf, __ELF_WORD_SIZE), OID_AUTO, 129291c06a1SKonstantin Belousov nxstack, CTLFLAG_RW, &__elfN(nxstack), 0, 130291c06a1SKonstantin Belousov __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) ": enable non-executable stack"); 131291c06a1SKonstantin Belousov 132eb029587SKonstantin Belousov #if defined(__amd64__) 133eb029587SKonstantin Belousov static int __elfN(vdso) = 1; 134eb029587SKonstantin Belousov SYSCTL_INT(__CONCAT(_kern_elf, __ELF_WORD_SIZE), OID_AUTO, 135eb029587SKonstantin Belousov vdso, CTLFLAG_RWTUN, &__elfN(vdso), 0, 136eb029587SKonstantin Belousov __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) ": enable vdso preloading"); 137eb029587SKonstantin Belousov #else 138eb029587SKonstantin Belousov static int __elfN(vdso) = 0; 139eb029587SKonstantin Belousov #endif 140eb029587SKonstantin Belousov 141eb785fabSKonstantin Belousov #if __ELF_WORD_SIZE == 32 && (defined(__amd64__) || defined(__i386__)) 142126b36a2SKonstantin Belousov int i386_read_exec = 0; 143126b36a2SKonstantin Belousov SYSCTL_INT(_kern_elf32, OID_AUTO, read_exec, CTLFLAG_RW, &i386_read_exec, 0, 144126b36a2SKonstantin Belousov "enable execution from readable segments"); 145126b36a2SKonstantin Belousov #endif 146126b36a2SKonstantin Belousov 14795aafd69SKonstantin Belousov static u_long __elfN(pie_base) = ET_DYN_LOAD_ADDR; 148f33533daSKonstantin Belousov static int 149f33533daSKonstantin Belousov sysctl_pie_base(SYSCTL_HANDLER_ARGS) 150f33533daSKonstantin Belousov { 151f33533daSKonstantin Belousov u_long val; 152f33533daSKonstantin Belousov int error; 153f33533daSKonstantin Belousov 154f33533daSKonstantin Belousov val = __elfN(pie_base); 155f33533daSKonstantin Belousov error = sysctl_handle_long(oidp, &val, 0, req); 156f33533daSKonstantin Belousov if (error != 0 || req->newptr == NULL) 157f33533daSKonstantin Belousov return (error); 158f33533daSKonstantin Belousov if ((val & PAGE_MASK) != 0) 159f33533daSKonstantin Belousov return (EINVAL); 160f33533daSKonstantin Belousov __elfN(pie_base) = val; 161f33533daSKonstantin Belousov return (0); 162f33533daSKonstantin Belousov } 163f33533daSKonstantin Belousov SYSCTL_PROC(__CONCAT(_kern_elf, __ELF_WORD_SIZE), OID_AUTO, pie_base, 164f33533daSKonstantin Belousov CTLTYPE_ULONG | CTLFLAG_MPSAFE | CTLFLAG_RW, NULL, 0, 165f33533daSKonstantin Belousov sysctl_pie_base, "LU", 16695aafd69SKonstantin Belousov "PIE load base without randomization"); 16795aafd69SKonstantin Belousov 1687029da5cSPawel Biernacki SYSCTL_NODE(__CONCAT(_kern_elf, __ELF_WORD_SIZE), OID_AUTO, aslr, 1697029da5cSPawel Biernacki CTLFLAG_RW | CTLFLAG_MPSAFE, 0, 170fa50a355SKonstantin Belousov ""); 171fa50a355SKonstantin Belousov #define ASLR_NODE_OID __CONCAT(__CONCAT(_kern_elf, __ELF_WORD_SIZE), _aslr) 172fa50a355SKonstantin Belousov 173b014e0f1SMarcin Wojtas /* 174f0687f3eSEd Maste * Enable ASLR by default for 64-bit non-PIE binaries. 32-bit architectures 175f0687f3eSEd Maste * have limited address space (which can cause issues for applications with 176f0687f3eSEd Maste * high memory use) so we leave it off there. 177b014e0f1SMarcin Wojtas */ 178b014e0f1SMarcin Wojtas static int __elfN(aslr_enabled) = __ELF_WORD_SIZE == 64; 179fa50a355SKonstantin Belousov SYSCTL_INT(ASLR_NODE_OID, OID_AUTO, enable, CTLFLAG_RWTUN, 180fa50a355SKonstantin Belousov &__elfN(aslr_enabled), 0, 181fa50a355SKonstantin Belousov __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) 182fa50a355SKonstantin Belousov ": enable address map randomization"); 183fa50a355SKonstantin Belousov 184b014e0f1SMarcin Wojtas /* 185f0687f3eSEd Maste * Enable ASLR by default for 64-bit PIE binaries. 186b014e0f1SMarcin Wojtas */ 187b014e0f1SMarcin Wojtas static int __elfN(pie_aslr_enabled) = __ELF_WORD_SIZE == 64; 188fa50a355SKonstantin Belousov SYSCTL_INT(ASLR_NODE_OID, OID_AUTO, pie_enable, CTLFLAG_RWTUN, 189fa50a355SKonstantin Belousov &__elfN(pie_aslr_enabled), 0, 190fa50a355SKonstantin Belousov __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) 191fa50a355SKonstantin Belousov ": enable address map randomization for PIE binaries"); 192fa50a355SKonstantin Belousov 193b014e0f1SMarcin Wojtas /* 194f0687f3eSEd Maste * Sbrk is deprecated and it can be assumed that in most cases it will not be 195f0687f3eSEd Maste * used anyway. This setting is valid only with ASLR enabled, and allows ASLR 196f0687f3eSEd Maste * to use the bss grow region. 197b014e0f1SMarcin Wojtas */ 198b014e0f1SMarcin Wojtas static int __elfN(aslr_honor_sbrk) = 0; 199fa50a355SKonstantin Belousov SYSCTL_INT(ASLR_NODE_OID, OID_AUTO, honor_sbrk, CTLFLAG_RW, 200fa50a355SKonstantin Belousov &__elfN(aslr_honor_sbrk), 0, 201fa50a355SKonstantin Belousov __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) ": assume sbrk is used"); 202fa50a355SKonstantin Belousov 2031798b44fSKonstantin Belousov static int __elfN(aslr_stack) = __ELF_WORD_SIZE == 64; 2041811c1e9SMark Johnston SYSCTL_INT(ASLR_NODE_OID, OID_AUTO, stack, CTLFLAG_RWTUN, 2051811c1e9SMark Johnston &__elfN(aslr_stack), 0, 206fc83c5a7SKonstantin Belousov __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) 2071811c1e9SMark Johnston ": enable stack address randomization"); 208fc83c5a7SKonstantin Belousov 209939f0b63SKornel Dulęba static int __elfN(aslr_shared_page) = __ELF_WORD_SIZE == 64; 210939f0b63SKornel Dulęba SYSCTL_INT(ASLR_NODE_OID, OID_AUTO, shared_page, CTLFLAG_RWTUN, 211939f0b63SKornel Dulęba &__elfN(aslr_shared_page), 0, 212939f0b63SKornel Dulęba __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) 213939f0b63SKornel Dulęba ": enable shared page address randomization"); 214939f0b63SKornel Dulęba 215944cf37bSKonstantin Belousov static int __elfN(sigfastblock) = 1; 216944cf37bSKonstantin Belousov SYSCTL_INT(__CONCAT(_kern_elf, __ELF_WORD_SIZE), OID_AUTO, sigfastblock, 217944cf37bSKonstantin Belousov CTLFLAG_RWTUN, &__elfN(sigfastblock), 0, 218944cf37bSKonstantin Belousov "enable sigfastblock for new processes"); 219944cf37bSKonstantin Belousov 2202e1c94aaSKonstantin Belousov static bool __elfN(allow_wx) = true; 2212e1c94aaSKonstantin Belousov SYSCTL_BOOL(__CONCAT(_kern_elf, __ELF_WORD_SIZE), OID_AUTO, allow_wx, 2222e1c94aaSKonstantin Belousov CTLFLAG_RWTUN, &__elfN(allow_wx), 0, 2232e1c94aaSKonstantin Belousov "Allow pages to be mapped simultaneously writable and executable"); 2242e1c94aaSKonstantin Belousov 2253ebc1248SPeter Wemm static Elf_Brandinfo *elf_brand_list[MAX_BRANDS]; 226e1743d02SSøren Schmidt 227545517f1SEdward Tomasz Napierala #define aligned(a, t) (rounddown2((u_long)(a), sizeof(t)) == (u_long)(a)) 22893d1c728SKonstantin Belousov 22932c01de2SDmitry Chagin Elf_Brandnote __elfN(freebsd_brandnote) = { 23032c01de2SDmitry Chagin .hdr.n_namesz = sizeof(FREEBSD_ABI_VENDOR), 23132c01de2SDmitry Chagin .hdr.n_descsz = sizeof(int32_t), 2324c22b468SEd Maste .hdr.n_type = NT_FREEBSD_ABI_TAG, 23332c01de2SDmitry Chagin .vendor = FREEBSD_ABI_VENDOR, 23489ffc202SBjoern A. Zeeb .flags = BN_TRANSLATE_OSREL, 23589ffc202SBjoern A. Zeeb .trans_osrel = __elfN(freebsd_trans_osrel) 23632c01de2SDmitry Chagin }; 23732c01de2SDmitry Chagin 238a95659f7SEd Maste static bool 23989ffc202SBjoern A. Zeeb __elfN(freebsd_trans_osrel)(const Elf_Note *note, int32_t *osrel) 24089ffc202SBjoern A. Zeeb { 24189ffc202SBjoern A. Zeeb uintptr_t p; 24289ffc202SBjoern A. Zeeb 24389ffc202SBjoern A. Zeeb p = (uintptr_t)(note + 1); 2441b8388cdSMikolaj Golub p += roundup2(note->n_namesz, ELF_NOTE_ROUNDSIZE); 24589ffc202SBjoern A. Zeeb *osrel = *(const int32_t *)(p); 24689ffc202SBjoern A. Zeeb 247a95659f7SEd Maste return (true); 24889ffc202SBjoern A. Zeeb } 24989ffc202SBjoern A. Zeeb 25089ffc202SBjoern A. Zeeb static int GNU_KFREEBSD_ABI_DESC = 3; 25189ffc202SBjoern A. Zeeb 25289ffc202SBjoern A. Zeeb Elf_Brandnote __elfN(kfreebsd_brandnote) = { 25389ffc202SBjoern A. Zeeb .hdr.n_namesz = sizeof(GNU_ABI_VENDOR), 25489ffc202SBjoern A. Zeeb .hdr.n_descsz = 16, /* XXX at least 16 */ 25589ffc202SBjoern A. Zeeb .hdr.n_type = 1, 25689ffc202SBjoern A. Zeeb .vendor = GNU_ABI_VENDOR, 25789ffc202SBjoern A. Zeeb .flags = BN_TRANSLATE_OSREL, 25889ffc202SBjoern A. Zeeb .trans_osrel = kfreebsd_trans_osrel 25989ffc202SBjoern A. Zeeb }; 26089ffc202SBjoern A. Zeeb 261a95659f7SEd Maste static bool 26289ffc202SBjoern A. Zeeb kfreebsd_trans_osrel(const Elf_Note *note, int32_t *osrel) 26389ffc202SBjoern A. Zeeb { 26489ffc202SBjoern A. Zeeb const Elf32_Word *desc; 26589ffc202SBjoern A. Zeeb uintptr_t p; 26689ffc202SBjoern A. Zeeb 26789ffc202SBjoern A. Zeeb p = (uintptr_t)(note + 1); 2681b8388cdSMikolaj Golub p += roundup2(note->n_namesz, ELF_NOTE_ROUNDSIZE); 26989ffc202SBjoern A. Zeeb 27089ffc202SBjoern A. Zeeb desc = (const Elf32_Word *)p; 27189ffc202SBjoern A. Zeeb if (desc[0] != GNU_KFREEBSD_ABI_DESC) 272a95659f7SEd Maste return (false); 27389ffc202SBjoern A. Zeeb 27489ffc202SBjoern A. Zeeb /* 27589ffc202SBjoern A. Zeeb * Debian GNU/kFreeBSD embed the earliest compatible kernel version 27689ffc202SBjoern A. Zeeb * (__FreeBSD_version: <major><two digit minor>Rxx) in the LSB way. 27789ffc202SBjoern A. Zeeb */ 27889ffc202SBjoern A. Zeeb *osrel = desc[1] * 100000 + desc[2] * 1000 + desc[3]; 27989ffc202SBjoern A. Zeeb 280a95659f7SEd Maste return (true); 28189ffc202SBjoern A. Zeeb } 28289ffc202SBjoern A. Zeeb 283e1743d02SSøren Schmidt int 2843ebc1248SPeter Wemm __elfN(insert_brand_entry)(Elf_Brandinfo *entry) 285e1743d02SSøren Schmidt { 286e1743d02SSøren Schmidt int i; 287e1743d02SSøren Schmidt 2883ebc1248SPeter Wemm for (i = 0; i < MAX_BRANDS; i++) { 289ea5a2b2eSSøren Schmidt if (elf_brand_list[i] == NULL) { 290ea5a2b2eSSøren Schmidt elf_brand_list[i] = entry; 291e1743d02SSøren Schmidt break; 292e1743d02SSøren Schmidt } 293e1743d02SSøren Schmidt } 294925c8b5bSBjoern A. Zeeb if (i == MAX_BRANDS) { 295925c8b5bSBjoern A. Zeeb printf("WARNING: %s: could not insert brandinfo entry: %p\n", 296925c8b5bSBjoern A. Zeeb __func__, entry); 297a7cddfedSJake Burkholder return (-1); 298925c8b5bSBjoern A. Zeeb } 299a7cddfedSJake Burkholder return (0); 300e1743d02SSøren Schmidt } 301e1743d02SSøren Schmidt 302e1743d02SSøren Schmidt int 3033ebc1248SPeter Wemm __elfN(remove_brand_entry)(Elf_Brandinfo *entry) 304e1743d02SSøren Schmidt { 305e1743d02SSøren Schmidt int i; 306e1743d02SSøren Schmidt 3073ebc1248SPeter Wemm for (i = 0; i < MAX_BRANDS; i++) { 308ea5a2b2eSSøren Schmidt if (elf_brand_list[i] == entry) { 309ea5a2b2eSSøren Schmidt elf_brand_list[i] = NULL; 310e1743d02SSøren Schmidt break; 311e1743d02SSøren Schmidt } 312e1743d02SSøren Schmidt } 313ea5a2b2eSSøren Schmidt if (i == MAX_BRANDS) 314a7cddfedSJake Burkholder return (-1); 315a7cddfedSJake Burkholder return (0); 316e1743d02SSøren Schmidt } 317e1743d02SSøren Schmidt 3184082b189SAlex Richardson bool 3193ebc1248SPeter Wemm __elfN(brand_inuse)(Elf_Brandinfo *entry) 320096977faSMark Newton { 321096977faSMark Newton struct proc *p; 3224082b189SAlex Richardson bool rval = false; 323096977faSMark Newton 3241005a129SJohn Baldwin sx_slock(&allproc_lock); 3254f506694SXin LI FOREACH_PROC_IN_SYSTEM(p) { 326553629ebSJake Burkholder if (p->p_sysent == entry->sysvec) { 3274082b189SAlex Richardson rval = true; 328553629ebSJake Burkholder break; 329096977faSMark Newton } 330553629ebSJake Burkholder } 3311005a129SJohn Baldwin sx_sunlock(&allproc_lock); 332096977faSMark Newton 333553629ebSJake Burkholder return (rval); 334096977faSMark Newton } 335096977faSMark Newton 3365fe3ed62SJake Burkholder static Elf_Brandinfo * 33732c01de2SDmitry Chagin __elfN(get_brandinfo)(struct image_params *imgp, const char *interp, 33809c78d53SEdward Tomasz Napierala int32_t *osrel, uint32_t *fctl0) 3395fe3ed62SJake Burkholder { 34032c01de2SDmitry Chagin const Elf_Ehdr *hdr = (const Elf_Ehdr *)imgp->image_header; 341af582aaeSKonstantin Belousov Elf_Brandinfo *bi, *bi_m; 34219621645SAlex Richardson bool ret, has_fctl0; 34309c78d53SEdward Tomasz Napierala int i, interp_name_len; 34409c78d53SEdward Tomasz Napierala 345be7808dcSKonstantin Belousov interp_name_len = interp != NULL ? strlen(interp) + 1 : 0; 3465fe3ed62SJake Burkholder 3475fe3ed62SJake Burkholder /* 34832c01de2SDmitry Chagin * We support four types of branding -- (1) the ELF EI_OSABI field 3495fe3ed62SJake Burkholder * that SCO added to the ELF spec, (2) FreeBSD 3.x's traditional string 35032c01de2SDmitry Chagin * branding w/in the ELF header, (3) path of the `interp_path' 35132c01de2SDmitry Chagin * field, and (4) the ".note.ABI-tag" ELF section. 3525fe3ed62SJake Burkholder */ 3535fe3ed62SJake Burkholder 35432c01de2SDmitry Chagin /* Look for an ".note.ABI-tag" ELF section */ 355af582aaeSKonstantin Belousov bi_m = NULL; 35632c01de2SDmitry Chagin for (i = 0; i < MAX_BRANDS; i++) { 35732c01de2SDmitry Chagin bi = elf_brand_list[i]; 358ecc2fda8SBjoern A. Zeeb if (bi == NULL) 359ecc2fda8SBjoern A. Zeeb continue; 3602274ab3dSKonstantin Belousov if (interp != NULL && (bi->flags & BI_BRAND_ONLY_STATIC) != 0) 3611438fe3cSKonstantin Belousov continue; 362ecc2fda8SBjoern A. Zeeb if (hdr->e_machine == bi->machine && (bi->flags & 363ecc2fda8SBjoern A. Zeeb (BI_BRAND_NOTE|BI_BRAND_NOTE_MANDATORY)) != 0) { 3640cad2aa2SKonstantin Belousov has_fctl0 = false; 3650cad2aa2SKonstantin Belousov *fctl0 = 0; 3660cad2aa2SKonstantin Belousov *osrel = 0; 367cefb93f2SKonstantin Belousov ret = __elfN(check_note)(imgp, bi->brand_note, osrel, 3680cad2aa2SKonstantin Belousov &has_fctl0, fctl0); 369f19d421aSNathan Whitehorn /* Give brand a chance to veto check_note's guess */ 3700cad2aa2SKonstantin Belousov if (ret && bi->header_supported) { 3710cad2aa2SKonstantin Belousov ret = bi->header_supported(imgp, osrel, 3720cad2aa2SKonstantin Belousov has_fctl0 ? fctl0 : NULL); 3730cad2aa2SKonstantin Belousov } 374af582aaeSKonstantin Belousov /* 375af582aaeSKonstantin Belousov * If note checker claimed the binary, but the 376af582aaeSKonstantin Belousov * interpreter path in the image does not 377af582aaeSKonstantin Belousov * match default one for the brand, try to 378af582aaeSKonstantin Belousov * search for other brands with the same 379af582aaeSKonstantin Belousov * interpreter. Either there is better brand 380af582aaeSKonstantin Belousov * with the right interpreter, or, failing 381af582aaeSKonstantin Belousov * this, we return first brand which accepted 382af582aaeSKonstantin Belousov * our note and, optionally, header. 383af582aaeSKonstantin Belousov */ 3843aeacc55SKonstantin Belousov if (ret && bi_m == NULL && interp != NULL && 3853aeacc55SKonstantin Belousov (bi->interp_path == NULL || 3863aeacc55SKonstantin Belousov (strlen(bi->interp_path) + 1 != interp_name_len || 3873aeacc55SKonstantin Belousov strncmp(interp, bi->interp_path, interp_name_len) 3883aeacc55SKonstantin Belousov != 0))) { 389af582aaeSKonstantin Belousov bi_m = bi; 390af582aaeSKonstantin Belousov ret = 0; 391af582aaeSKonstantin Belousov } 39232c01de2SDmitry Chagin if (ret) 39332c01de2SDmitry Chagin return (bi); 39432c01de2SDmitry Chagin } 39532c01de2SDmitry Chagin } 396af582aaeSKonstantin Belousov if (bi_m != NULL) 397af582aaeSKonstantin Belousov return (bi_m); 39832c01de2SDmitry Chagin 3995fe3ed62SJake Burkholder /* If the executable has a brand, search for it in the brand list. */ 4005fe3ed62SJake Burkholder for (i = 0; i < MAX_BRANDS; i++) { 4015fe3ed62SJake Burkholder bi = elf_brand_list[i]; 4021438fe3cSKonstantin Belousov if (bi == NULL || (bi->flags & BI_BRAND_NOTE_MANDATORY) != 0 || 4032274ab3dSKonstantin Belousov (interp != NULL && (bi->flags & BI_BRAND_ONLY_STATIC) != 0)) 404ecc2fda8SBjoern A. Zeeb continue; 405ecc2fda8SBjoern A. Zeeb if (hdr->e_machine == bi->machine && 4065fe3ed62SJake Burkholder (hdr->e_ident[EI_OSABI] == bi->brand || 4070fe98320SEd Schouten (bi->compat_3_brand != NULL && 4083d560b4bSKonstantin Belousov strcmp((const char *)&hdr->e_ident[OLD_EI_BRAND], 4090fe98320SEd Schouten bi->compat_3_brand) == 0))) { 410686d2f31SNathan Whitehorn /* Looks good, but give brand a chance to veto */ 411d722231bSJohn Baldwin if (bi->header_supported == NULL || 4120cad2aa2SKonstantin Belousov bi->header_supported(imgp, NULL, NULL)) { 41315a9aedfSKonstantin Belousov /* 41415a9aedfSKonstantin Belousov * Again, prefer strictly matching 41515a9aedfSKonstantin Belousov * interpreter path. 41615a9aedfSKonstantin Belousov */ 4177aab7a80SKonstantin Belousov if (interp_name_len == 0 && 4187aab7a80SKonstantin Belousov bi->interp_path == NULL) 4197aab7a80SKonstantin Belousov return (bi); 4207aab7a80SKonstantin Belousov if (bi->interp_path != NULL && 4217aab7a80SKonstantin Belousov strlen(bi->interp_path) + 1 == 42215a9aedfSKonstantin Belousov interp_name_len && strncmp(interp, 42315a9aedfSKonstantin Belousov bi->interp_path, interp_name_len) == 0) 4245fe3ed62SJake Burkholder return (bi); 42515a9aedfSKonstantin Belousov if (bi_m == NULL) 42615a9aedfSKonstantin Belousov bi_m = bi; 4275fe3ed62SJake Burkholder } 428686d2f31SNathan Whitehorn } 42915a9aedfSKonstantin Belousov } 43015a9aedfSKonstantin Belousov if (bi_m != NULL) 43115a9aedfSKonstantin Belousov return (bi_m); 4325fe3ed62SJake Burkholder 433817dc004SWarner Losh /* No known brand, see if the header is recognized by any brand */ 434817dc004SWarner Losh for (i = 0; i < MAX_BRANDS; i++) { 435817dc004SWarner Losh bi = elf_brand_list[i]; 436817dc004SWarner Losh if (bi == NULL || bi->flags & BI_BRAND_NOTE_MANDATORY || 437817dc004SWarner Losh bi->header_supported == NULL) 438817dc004SWarner Losh continue; 439817dc004SWarner Losh if (hdr->e_machine == bi->machine) { 4400cad2aa2SKonstantin Belousov ret = bi->header_supported(imgp, NULL, NULL); 441817dc004SWarner Losh if (ret) 442817dc004SWarner Losh return (bi); 443817dc004SWarner Losh } 444817dc004SWarner Losh } 445817dc004SWarner Losh 4465fe3ed62SJake Burkholder /* Lacking a known brand, search for a recognized interpreter. */ 4475fe3ed62SJake Burkholder if (interp != NULL) { 4485fe3ed62SJake Burkholder for (i = 0; i < MAX_BRANDS; i++) { 4495fe3ed62SJake Burkholder bi = elf_brand_list[i]; 4502274ab3dSKonstantin Belousov if (bi == NULL || (bi->flags & 4512274ab3dSKonstantin Belousov (BI_BRAND_NOTE_MANDATORY | BI_BRAND_ONLY_STATIC)) 4522274ab3dSKonstantin Belousov != 0) 453ecc2fda8SBjoern A. Zeeb continue; 454ecc2fda8SBjoern A. Zeeb if (hdr->e_machine == bi->machine && 4553aeacc55SKonstantin Belousov bi->interp_path != NULL && 456d1ae5c83SKonstantin Belousov /* ELF image p_filesz includes terminating zero */ 457d1ae5c83SKonstantin Belousov strlen(bi->interp_path) + 1 == interp_name_len && 458d1ae5c83SKonstantin Belousov strncmp(interp, bi->interp_path, interp_name_len) 459d722231bSJohn Baldwin == 0 && (bi->header_supported == NULL || 4600cad2aa2SKonstantin Belousov bi->header_supported(imgp, NULL, NULL))) 4615fe3ed62SJake Burkholder return (bi); 4625fe3ed62SJake Burkholder } 4635fe3ed62SJake Burkholder } 4645fe3ed62SJake Burkholder 4655fe3ed62SJake Burkholder /* Lacking a recognized interpreter, try the default brand */ 4665fe3ed62SJake Burkholder for (i = 0; i < MAX_BRANDS; i++) { 4675fe3ed62SJake Burkholder bi = elf_brand_list[i]; 4681438fe3cSKonstantin Belousov if (bi == NULL || (bi->flags & BI_BRAND_NOTE_MANDATORY) != 0 || 4692274ab3dSKonstantin Belousov (interp != NULL && (bi->flags & BI_BRAND_ONLY_STATIC) != 0)) 470ecc2fda8SBjoern A. Zeeb continue; 471ecc2fda8SBjoern A. Zeeb if (hdr->e_machine == bi->machine && 472d722231bSJohn Baldwin __elfN(fallback_brand) == bi->brand && 473d722231bSJohn Baldwin (bi->header_supported == NULL || 4740cad2aa2SKonstantin Belousov bi->header_supported(imgp, NULL, NULL))) 4755fe3ed62SJake Burkholder return (bi); 4765fe3ed62SJake Burkholder } 4775fe3ed62SJake Burkholder return (NULL); 4785fe3ed62SJake Burkholder } 4795fe3ed62SJake Burkholder 4807de1bc13SKonstantin Belousov static bool 4817de1bc13SKonstantin Belousov __elfN(phdr_in_zero_page)(const Elf_Ehdr *hdr) 4827de1bc13SKonstantin Belousov { 4837de1bc13SKonstantin Belousov return (hdr->e_phoff <= PAGE_SIZE && 4847de1bc13SKonstantin Belousov (u_int)hdr->e_phentsize * hdr->e_phnum <= PAGE_SIZE - hdr->e_phoff); 4857de1bc13SKonstantin Belousov } 4867de1bc13SKonstantin Belousov 487e1743d02SSøren Schmidt static int 4883ebc1248SPeter Wemm __elfN(check_header)(const Elf_Ehdr *hdr) 489e1743d02SSøren Schmidt { 490d0ca7c29SPeter Wemm Elf_Brandinfo *bi; 4913ebc1248SPeter Wemm int i; 4923ebc1248SPeter Wemm 49352c24af7SPeter Wemm if (!IS_ELF(*hdr) || 49452c24af7SPeter Wemm hdr->e_ident[EI_CLASS] != ELF_TARG_CLASS || 49552c24af7SPeter Wemm hdr->e_ident[EI_DATA] != ELF_TARG_DATA || 4963dc19c46SJacques Vidrine hdr->e_ident[EI_VERSION] != EV_CURRENT || 4973dc19c46SJacques Vidrine hdr->e_phentsize != sizeof(Elf_Phdr) || 4983dc19c46SJacques Vidrine hdr->e_version != ELF_TARG_VER) 499a7cddfedSJake Burkholder return (ENOEXEC); 500e1743d02SSøren Schmidt 5013ebc1248SPeter Wemm /* 5023ebc1248SPeter Wemm * Make sure we have at least one brand for this machine. 5033ebc1248SPeter Wemm */ 5043ebc1248SPeter Wemm 5053ebc1248SPeter Wemm for (i = 0; i < MAX_BRANDS; i++) { 506d0ca7c29SPeter Wemm bi = elf_brand_list[i]; 507d0ca7c29SPeter Wemm if (bi != NULL && bi->machine == hdr->e_machine) 5083ebc1248SPeter Wemm break; 5093ebc1248SPeter Wemm } 5103ebc1248SPeter Wemm if (i == MAX_BRANDS) 511a7cddfedSJake Burkholder return (ENOEXEC); 512e1743d02SSøren Schmidt 513a7cddfedSJake Burkholder return (0); 514e1743d02SSøren Schmidt } 515e1743d02SSøren Schmidt 516e1743d02SSøren Schmidt static int 5173ebc1248SPeter Wemm __elfN(map_partial)(vm_map_t map, vm_object_t object, vm_ooffset_t offset, 518ff6f03c7SAlan Cox vm_offset_t start, vm_offset_t end, vm_prot_t prot) 5193ebc1248SPeter Wemm { 520da61b9a6SAlan Cox struct sf_buf *sf; 521da61b9a6SAlan Cox int error; 5223ebc1248SPeter Wemm vm_offset_t off; 5233ebc1248SPeter Wemm 5243ebc1248SPeter Wemm /* 5253ebc1248SPeter Wemm * Create the page if it doesn't exist yet. Ignore errors. 5263ebc1248SPeter Wemm */ 527aaadc41fSKonstantin Belousov vm_map_fixed(map, NULL, 0, trunc_page(start), round_page(end) - 528aaadc41fSKonstantin Belousov trunc_page(start), VM_PROT_ALL, VM_PROT_ALL, MAP_CHECK_EXCL); 5293ebc1248SPeter Wemm 5303ebc1248SPeter Wemm /* 5313ebc1248SPeter Wemm * Find the page from the underlying object. 5323ebc1248SPeter Wemm */ 53328e8da65SAlan Cox if (object != NULL) { 534da61b9a6SAlan Cox sf = vm_imgact_map_page(object, offset); 535da61b9a6SAlan Cox if (sf == NULL) 536da61b9a6SAlan Cox return (KERN_FAILURE); 5373ebc1248SPeter Wemm off = offset - trunc_page(offset); 538da61b9a6SAlan Cox error = copyout((caddr_t)sf_buf_kva(sf) + off, (caddr_t)start, 539ca0387efSJake Burkholder end - start); 540be996836SAttilio Rao vm_imgact_unmap_page(sf); 541fe0a8a39SKonstantin Belousov if (error != 0) 542a7cddfedSJake Burkholder return (KERN_FAILURE); 5433ebc1248SPeter Wemm } 5443ebc1248SPeter Wemm 545a7cddfedSJake Burkholder return (KERN_SUCCESS); 5463ebc1248SPeter Wemm } 5473ebc1248SPeter Wemm 5483ebc1248SPeter Wemm static int 549169641f7SAlex Richardson __elfN(map_insert)(const struct image_params *imgp, vm_map_t map, 550169641f7SAlex Richardson vm_object_t object, vm_ooffset_t offset, vm_offset_t start, vm_offset_t end, 551169641f7SAlex Richardson vm_prot_t prot, int cow) 5523ebc1248SPeter Wemm { 553da61b9a6SAlan Cox struct sf_buf *sf; 554da61b9a6SAlan Cox vm_offset_t off; 555a063facbSMarcel Moolenaar vm_size_t sz; 556e3d8f8feSKonstantin Belousov int error, locked, rv; 5573ebc1248SPeter Wemm 5583ebc1248SPeter Wemm if (start != trunc_page(start)) { 55981f223caSJake Burkholder rv = __elfN(map_partial)(map, object, offset, start, 560ff6f03c7SAlan Cox round_page(start), prot); 56128e8da65SAlan Cox if (rv != KERN_SUCCESS) 562a7cddfedSJake Burkholder return (rv); 5633ebc1248SPeter Wemm offset += round_page(start) - start; 5643ebc1248SPeter Wemm start = round_page(start); 5653ebc1248SPeter Wemm } 5663ebc1248SPeter Wemm if (end != round_page(end)) { 56781f223caSJake Burkholder rv = __elfN(map_partial)(map, object, offset + 568ff6f03c7SAlan Cox trunc_page(end) - start, trunc_page(end), end, prot); 56928e8da65SAlan Cox if (rv != KERN_SUCCESS) 570a7cddfedSJake Burkholder return (rv); 5713ebc1248SPeter Wemm end = trunc_page(end); 5723ebc1248SPeter Wemm } 573e383e820SAlan Cox if (start >= end) 574e383e820SAlan Cox return (KERN_SUCCESS); 575e383e820SAlan Cox if ((offset & PAGE_MASK) != 0) { 5763ebc1248SPeter Wemm /* 577e383e820SAlan Cox * The mapping is not page aligned. This means that we have 578e383e820SAlan Cox * to copy the data. 5793ebc1248SPeter Wemm */ 580aaadc41fSKonstantin Belousov rv = vm_map_fixed(map, NULL, 0, start, end - start, 581aaadc41fSKonstantin Belousov prot | VM_PROT_WRITE, VM_PROT_ALL, MAP_CHECK_EXCL); 5825420f76bSKonstantin Belousov if (rv != KERN_SUCCESS) 583a7cddfedSJake Burkholder return (rv); 584da61b9a6SAlan Cox if (object == NULL) 585da61b9a6SAlan Cox return (KERN_SUCCESS); 586da61b9a6SAlan Cox for (; start < end; start += sz) { 587da61b9a6SAlan Cox sf = vm_imgact_map_page(object, offset); 588da61b9a6SAlan Cox if (sf == NULL) 589da61b9a6SAlan Cox return (KERN_FAILURE); 5903ebc1248SPeter Wemm off = offset - trunc_page(offset); 5913ebc1248SPeter Wemm sz = end - start; 592da61b9a6SAlan Cox if (sz > PAGE_SIZE - off) 593da61b9a6SAlan Cox sz = PAGE_SIZE - off; 594da61b9a6SAlan Cox error = copyout((caddr_t)sf_buf_kva(sf) + off, 5953ebc1248SPeter Wemm (caddr_t)start, sz); 596be996836SAttilio Rao vm_imgact_unmap_page(sf); 5975420f76bSKonstantin Belousov if (error != 0) 598a7cddfedSJake Burkholder return (KERN_FAILURE); 599da61b9a6SAlan Cox offset += sz; 6003ebc1248SPeter Wemm } 6013ebc1248SPeter Wemm } else { 602e5e6093bSAlan Cox vm_object_reference(object); 603e383e820SAlan Cox rv = vm_map_fixed(map, object, offset, start, end - start, 60478022527SKonstantin Belousov prot, VM_PROT_ALL, cow | MAP_CHECK_EXCL | 60578022527SKonstantin Belousov (object != NULL ? MAP_VN_EXEC : 0)); 606e3d8f8feSKonstantin Belousov if (rv != KERN_SUCCESS) { 607e3d8f8feSKonstantin Belousov locked = VOP_ISLOCKED(imgp->vp); 608b249ce48SMateusz Guzik VOP_UNLOCK(imgp->vp); 609e5e6093bSAlan Cox vm_object_deallocate(object); 610e3d8f8feSKonstantin Belousov vn_lock(imgp->vp, locked | LK_RETRY); 611a7cddfedSJake Burkholder return (rv); 61278022527SKonstantin Belousov } else if (object != NULL) { 61378022527SKonstantin Belousov MPASS(imgp->vp->v_object == object); 61478022527SKonstantin Belousov VOP_SET_TEXT_CHECKED(imgp->vp); 6153ebc1248SPeter Wemm } 6163ebc1248SPeter Wemm } 617e383e820SAlan Cox return (KERN_SUCCESS); 618e383e820SAlan Cox } 6193ebc1248SPeter Wemm 620102b4e33SJohn Baldwin static int 621102b4e33SJohn Baldwin __elfN(load_section)(const struct image_params *imgp, vm_ooffset_t offset, 622102b4e33SJohn Baldwin caddr_t vmaddr, size_t memsz, size_t filsz, vm_prot_t prot) 623e1743d02SSøren Schmidt { 624da61b9a6SAlan Cox struct sf_buf *sf; 625e1743d02SSøren Schmidt size_t map_len; 626292177e6SAlan Cox vm_map_t map; 627292177e6SAlan Cox vm_object_t object; 628e020a35fSMark Johnston vm_offset_t map_addr; 629fa7dd9c5SMatthew Dillon int error, rv, cow; 630e1743d02SSøren Schmidt size_t copy_len; 6310bbee4cdSKonstantin Belousov vm_ooffset_t file_addr; 63252c24af7SPeter Wemm 63325ead034SBrian Feldman /* 63425ead034SBrian Feldman * It's necessary to fail if the filsz + offset taken from the 63525ead034SBrian Feldman * header is greater than the actual file pager object's size. 63625ead034SBrian Feldman * If we were to allow this, then the vm_map_find() below would 63725ead034SBrian Feldman * walk right off the end of the file object and into the ether. 63825ead034SBrian Feldman * 63925ead034SBrian Feldman * While I'm here, might as well check for something else that 64025ead034SBrian Feldman * is invalid: filsz cannot be greater than memsz. 64125ead034SBrian Feldman */ 6429bcf2f2dSKonstantin Belousov if ((filsz != 0 && (off_t)filsz + offset > imgp->attr->va_size) || 6439bcf2f2dSKonstantin Belousov filsz > memsz) { 64425ead034SBrian Feldman uprintf("elf_load_section: truncated ELF file\n"); 64525ead034SBrian Feldman return (ENOEXEC); 64625ead034SBrian Feldman } 64725ead034SBrian Feldman 648292177e6SAlan Cox object = imgp->object; 649292177e6SAlan Cox map = &imgp->proc->p_vmspace->vm_map; 650545517f1SEdward Tomasz Napierala map_addr = trunc_page((vm_offset_t)vmaddr); 651545517f1SEdward Tomasz Napierala file_addr = trunc_page(offset); 652e1743d02SSøren Schmidt 653e1743d02SSøren Schmidt /* 65452c24af7SPeter Wemm * We have two choices. We can either clear the data in the last page 65552c24af7SPeter Wemm * of an oversized mapping, or we can start the anon mapping a page 65652c24af7SPeter Wemm * early and copy the initialized data into that first page. We 65728e8da65SAlan Cox * choose the second. 65852c24af7SPeter Wemm */ 6599bcf2f2dSKonstantin Belousov if (filsz == 0) 6609bcf2f2dSKonstantin Belousov map_len = 0; 6619bcf2f2dSKonstantin Belousov else if (memsz > filsz) 662545517f1SEdward Tomasz Napierala map_len = trunc_page(offset + filsz) - file_addr; 66352c24af7SPeter Wemm else 664545517f1SEdward Tomasz Napierala map_len = round_page(offset + filsz) - file_addr; 66552c24af7SPeter Wemm 66652c24af7SPeter Wemm if (map_len != 0) { 667fa7dd9c5SMatthew Dillon /* cow flags: don't dump readonly sections in core */ 668fa7dd9c5SMatthew Dillon cow = MAP_COPY_ON_WRITE | MAP_PREFAULT | 669fa7dd9c5SMatthew Dillon (prot & VM_PROT_WRITE ? 0 : MAP_DISABLE_COREDUMP); 670fa7dd9c5SMatthew Dillon 67178022527SKonstantin Belousov rv = __elfN(map_insert)(imgp, map, object, file_addr, 67278022527SKonstantin Belousov map_addr, map_addr + map_len, prot, cow); 673e5e6093bSAlan Cox if (rv != KERN_SUCCESS) 674a7cddfedSJake Burkholder return (EINVAL); 67552c24af7SPeter Wemm 67652c24af7SPeter Wemm /* we can stop now if we've covered it all */ 677973d67c4SKonstantin Belousov if (memsz == filsz) 678a7cddfedSJake Burkholder return (0); 67952c24af7SPeter Wemm } 68052c24af7SPeter Wemm 68152c24af7SPeter Wemm /* 68252c24af7SPeter Wemm * We have to get the remaining bit of the file into the first part 68352c24af7SPeter Wemm * of the oversized map segment. This is normally because the .data 68452c24af7SPeter Wemm * segment in the file is extended to provide bss. It's a neat idea 68552c24af7SPeter Wemm * to try and save a page, but it's a pain in the behind to implement. 686e1743d02SSøren Schmidt */ 687545517f1SEdward Tomasz Napierala copy_len = filsz == 0 ? 0 : (offset + filsz) - trunc_page(offset + 688545517f1SEdward Tomasz Napierala filsz); 689545517f1SEdward Tomasz Napierala map_addr = trunc_page((vm_offset_t)vmaddr + filsz); 690545517f1SEdward Tomasz Napierala map_len = round_page((vm_offset_t)vmaddr + memsz) - map_addr; 691e1743d02SSøren Schmidt 69252c24af7SPeter Wemm /* This had damn well better be true! */ 6938191d577SPeter Wemm if (map_len != 0) { 694e3d8f8feSKonstantin Belousov rv = __elfN(map_insert)(imgp, map, NULL, 0, map_addr, 695c547cbb4SAlan Cox map_addr + map_len, prot, 0); 696973d67c4SKonstantin Belousov if (rv != KERN_SUCCESS) 697a7cddfedSJake Burkholder return (EINVAL); 6988191d577SPeter Wemm } 699e1743d02SSøren Schmidt 70052c24af7SPeter Wemm if (copy_len != 0) { 701da61b9a6SAlan Cox sf = vm_imgact_map_page(object, offset + filsz); 702da61b9a6SAlan Cox if (sf == NULL) 703da61b9a6SAlan Cox return (EIO); 704e1743d02SSøren Schmidt 70552c24af7SPeter Wemm /* send the page fragment to user space */ 706e020a35fSMark Johnston error = copyout((caddr_t)sf_buf_kva(sf), (caddr_t)map_addr, 707e020a35fSMark Johnston copy_len); 708be996836SAttilio Rao vm_imgact_unmap_page(sf); 709973d67c4SKonstantin Belousov if (error != 0) 71052c24af7SPeter Wemm return (error); 71152c24af7SPeter Wemm } 712e1743d02SSøren Schmidt 713e1743d02SSøren Schmidt /* 714c547cbb4SAlan Cox * Remove write access to the page if it was only granted by map_insert 715c547cbb4SAlan Cox * to allow copyout. 716e1743d02SSøren Schmidt */ 717c547cbb4SAlan Cox if ((prot & VM_PROT_WRITE) == 0) 718292177e6SAlan Cox vm_map_protect(map, trunc_page(map_addr), round_page(map_addr + 7190659df6fSKonstantin Belousov map_len), prot, 0, VM_MAP_PROTECT_SET_PROT); 7208191d577SPeter Wemm 721ff6f03c7SAlan Cox return (0); 722e1743d02SSøren Schmidt } 723e1743d02SSøren Schmidt 7249bcd7482SEdward Tomasz Napierala static int 725169641f7SAlex Richardson __elfN(load_sections)(const struct image_params *imgp, const Elf_Ehdr *hdr, 7269bcd7482SEdward Tomasz Napierala const Elf_Phdr *phdr, u_long rbase, u_long *base_addrp) 7279bcd7482SEdward Tomasz Napierala { 7289bcd7482SEdward Tomasz Napierala vm_prot_t prot; 7299bcd7482SEdward Tomasz Napierala u_long base_addr; 7309bcd7482SEdward Tomasz Napierala bool first; 7319bcd7482SEdward Tomasz Napierala int error, i; 7329bcd7482SEdward Tomasz Napierala 733b65ca345SEdward Tomasz Napierala ASSERT_VOP_LOCKED(imgp->vp, __func__); 734b65ca345SEdward Tomasz Napierala 7359bcd7482SEdward Tomasz Napierala base_addr = 0; 7369bcd7482SEdward Tomasz Napierala first = true; 7379bcd7482SEdward Tomasz Napierala 7389bcd7482SEdward Tomasz Napierala for (i = 0; i < hdr->e_phnum; i++) { 7399bcd7482SEdward Tomasz Napierala if (phdr[i].p_type != PT_LOAD || phdr[i].p_memsz == 0) 7409bcd7482SEdward Tomasz Napierala continue; 7419bcd7482SEdward Tomasz Napierala 7429bcd7482SEdward Tomasz Napierala /* Loadable segment */ 7439bcd7482SEdward Tomasz Napierala prot = __elfN(trans_prot)(phdr[i].p_flags); 7449bcd7482SEdward Tomasz Napierala error = __elfN(load_section)(imgp, phdr[i].p_offset, 7459bcd7482SEdward Tomasz Napierala (caddr_t)(uintptr_t)phdr[i].p_vaddr + rbase, 7469bcd7482SEdward Tomasz Napierala phdr[i].p_memsz, phdr[i].p_filesz, prot); 7479bcd7482SEdward Tomasz Napierala if (error != 0) 7489bcd7482SEdward Tomasz Napierala return (error); 7499bcd7482SEdward Tomasz Napierala 7509bcd7482SEdward Tomasz Napierala /* 7519bcd7482SEdward Tomasz Napierala * Establish the base address if this is the first segment. 7529bcd7482SEdward Tomasz Napierala */ 7539bcd7482SEdward Tomasz Napierala if (first) { 7549bcd7482SEdward Tomasz Napierala base_addr = trunc_page(phdr[i].p_vaddr + rbase); 7559bcd7482SEdward Tomasz Napierala first = false; 7569bcd7482SEdward Tomasz Napierala } 7579bcd7482SEdward Tomasz Napierala } 7589bcd7482SEdward Tomasz Napierala 7599bcd7482SEdward Tomasz Napierala if (base_addrp != NULL) 7609bcd7482SEdward Tomasz Napierala *base_addrp = base_addr; 7619bcd7482SEdward Tomasz Napierala 7629bcd7482SEdward Tomasz Napierala return (0); 7639bcd7482SEdward Tomasz Napierala } 7649bcd7482SEdward Tomasz Napierala 765c33fe779SJohn Polstra /* 766c33fe779SJohn Polstra * Load the file "file" into memory. It may be either a shared object 767c33fe779SJohn Polstra * or an executable. 768c33fe779SJohn Polstra * 769c33fe779SJohn Polstra * The "addr" reference parameter is in/out. On entry, it specifies 770c33fe779SJohn Polstra * the address where a shared object should be loaded. If the file is 771c33fe779SJohn Polstra * an executable, this value is ignored. On exit, "addr" specifies 772c33fe779SJohn Polstra * where the file was actually loaded. 773c33fe779SJohn Polstra * 774c33fe779SJohn Polstra * The "entry" reference parameter is out only. On exit, it specifies 775c33fe779SJohn Polstra * the entry point for the loaded file. 776c33fe779SJohn Polstra */ 777e1743d02SSøren Schmidt static int 7783ebc1248SPeter Wemm __elfN(load_file)(struct proc *p, const char *file, u_long *addr, 7791699546dSEdward Tomasz Napierala u_long *entry) 780e1743d02SSøren Schmidt { 781911c2be0SMark Peek struct { 782911c2be0SMark Peek struct nameidata nd; 783911c2be0SMark Peek struct vattr attr; 784911c2be0SMark Peek struct image_params image_params; 785911c2be0SMark Peek } *tempdata; 786d254af07SMatthew Dillon const Elf_Ehdr *hdr = NULL; 787d254af07SMatthew Dillon const Elf_Phdr *phdr = NULL; 788911c2be0SMark Peek struct nameidata *nd; 789911c2be0SMark Peek struct vattr *attr; 790911c2be0SMark Peek struct image_params *imgp; 79178022527SKonstantin Belousov u_long rbase; 792c33fe779SJohn Polstra u_long base_addr = 0; 7939bcd7482SEdward Tomasz Napierala int error; 794e1743d02SSøren Schmidt 79512bc222eSJonathan Anderson #ifdef CAPABILITY_MODE 79612bc222eSJonathan Anderson /* 79712bc222eSJonathan Anderson * XXXJA: This check can go away once we are sufficiently confident 79812bc222eSJonathan Anderson * that the checks in namei() are correct. 79912bc222eSJonathan Anderson */ 80012bc222eSJonathan Anderson if (IN_CAPABILITY_MODE(curthread)) 80112bc222eSJonathan Anderson return (ECAPMODE); 80212bc222eSJonathan Anderson #endif 80312bc222eSJonathan Anderson 8041073d17eSKonstantin Belousov tempdata = malloc(sizeof(*tempdata), M_TEMP, M_WAITOK | M_ZERO); 805911c2be0SMark Peek nd = &tempdata->nd; 806911c2be0SMark Peek attr = &tempdata->attr; 807911c2be0SMark Peek imgp = &tempdata->image_params; 808911c2be0SMark Peek 809c8a79999SPeter Wemm /* 810c8a79999SPeter Wemm * Initialize part of the common data 811c8a79999SPeter Wemm */ 812c8a79999SPeter Wemm imgp->proc = p; 813911c2be0SMark Peek imgp->attr = attr; 814c8a79999SPeter Wemm 815f422bc30SJohn Baldwin NDINIT(nd, LOOKUP, ISOPEN | FOLLOW | LOCKSHARED | LOCKLEAF, 8167e1d3eefSMateusz Guzik UIO_SYSSPACE, file); 817911c2be0SMark Peek if ((error = namei(nd)) != 0) { 818911c2be0SMark Peek nd->ni_vp = NULL; 819e1743d02SSøren Schmidt goto fail; 820e1743d02SSøren Schmidt } 821bb92cd7bSMateusz Guzik NDFREE_PNBUF(nd); 822911c2be0SMark Peek imgp->vp = nd->ni_vp; 823c8a79999SPeter Wemm 824e1743d02SSøren Schmidt /* 825e1743d02SSøren Schmidt * Check permissions, modes, uid, etc on the file, and "open" it. 826e1743d02SSøren Schmidt */ 827c8a79999SPeter Wemm error = exec_check_permissions(imgp); 828373d1a3fSAlan Cox if (error) 829c8a79999SPeter Wemm goto fail; 830e1743d02SSøren Schmidt 831c8a79999SPeter Wemm error = exec_map_first_page(imgp); 832373d1a3fSAlan Cox if (error) 833373d1a3fSAlan Cox goto fail; 834373d1a3fSAlan Cox 8358516dd18SPoul-Henning Kamp imgp->object = nd->ni_vp->v_object; 836e1743d02SSøren Schmidt 837d254af07SMatthew Dillon hdr = (const Elf_Ehdr *)imgp->image_header; 8383ebc1248SPeter Wemm if ((error = __elfN(check_header)(hdr)) != 0) 839e1743d02SSøren Schmidt goto fail; 840c33fe779SJohn Polstra if (hdr->e_type == ET_DYN) 841c33fe779SJohn Polstra rbase = *addr; 842c33fe779SJohn Polstra else if (hdr->e_type == ET_EXEC) 843c33fe779SJohn Polstra rbase = 0; 844c33fe779SJohn Polstra else { 845c33fe779SJohn Polstra error = ENOEXEC; 846c33fe779SJohn Polstra goto fail; 847c33fe779SJohn Polstra } 848e1743d02SSøren Schmidt 849c8a79999SPeter Wemm /* Only support headers that fit within first page for now */ 8507de1bc13SKonstantin Belousov if (!__elfN(phdr_in_zero_page)(hdr)) { 851c8a79999SPeter Wemm error = ENOEXEC; 852e1743d02SSøren Schmidt goto fail; 853c8a79999SPeter Wemm } 854c8a79999SPeter Wemm 855d254af07SMatthew Dillon phdr = (const Elf_Phdr *)(imgp->image_header + hdr->e_phoff); 85693d1c728SKonstantin Belousov if (!aligned(phdr, Elf_Addr)) { 85793d1c728SKonstantin Belousov error = ENOEXEC; 85893d1c728SKonstantin Belousov goto fail; 85993d1c728SKonstantin Belousov } 860e1743d02SSøren Schmidt 8619bcd7482SEdward Tomasz Napierala error = __elfN(load_sections)(imgp, hdr, phdr, rbase, &base_addr); 862292177e6SAlan Cox if (error != 0) 863e1743d02SSøren Schmidt goto fail; 8649bcd7482SEdward Tomasz Napierala 865eb32c1c7SAndrew Turner if (p->p_sysent->sv_protect != NULL) 866eb32c1c7SAndrew Turner p->p_sysent->sv_protect(imgp, SVP_INTERP); 867eb32c1c7SAndrew Turner 868c33fe779SJohn Polstra *addr = base_addr; 869c33fe779SJohn Polstra *entry = (unsigned long)hdr->e_entry + rbase; 870e1743d02SSøren Schmidt 871e1743d02SSøren Schmidt fail: 872c8a79999SPeter Wemm if (imgp->firstpage) 873c8a79999SPeter Wemm exec_unmap_first_page(imgp); 8740b2ed1aeSJeff Roberson 87578022527SKonstantin Belousov if (nd->ni_vp) { 87678022527SKonstantin Belousov if (imgp->textset) 87778022527SKonstantin Belousov VOP_UNSET_TEXT_CHECKED(nd->ni_vp); 878373d1a3fSAlan Cox vput(nd->ni_vp); 87978022527SKonstantin Belousov } 880911c2be0SMark Peek free(tempdata, M_TEMP); 881e1743d02SSøren Schmidt 882a7cddfedSJake Burkholder return (error); 883e1743d02SSøren Schmidt } 884e1743d02SSøren Schmidt 885a4007ae1SKonstantin Belousov /* 886a4007ae1SKonstantin Belousov * Select randomized valid address in the map map, between minv and 887a4007ae1SKonstantin Belousov * maxv, with specified alignment. The [minv, maxv) range must belong 888a4007ae1SKonstantin Belousov * to the map. Note that function only allocates the address, it is 889a4007ae1SKonstantin Belousov * up to caller to clamp maxv in a way that the final allocation 890a4007ae1SKonstantin Belousov * length fit into the map. 891a4007ae1SKonstantin Belousov * 892a4007ae1SKonstantin Belousov * Result is returned in *resp, error code indicates that arguments 893a4007ae1SKonstantin Belousov * did not pass sanity checks for overflow and range correctness. 894a4007ae1SKonstantin Belousov */ 8959cf78c1cSKonstantin Belousov static int 8969cf78c1cSKonstantin Belousov __CONCAT(rnd_, __elfN(base))(vm_map_t map, u_long minv, u_long maxv, 8979cf78c1cSKonstantin Belousov u_int align, u_long *resp) 898fa50a355SKonstantin Belousov { 899fa50a355SKonstantin Belousov u_long rbase, res; 900fa50a355SKonstantin Belousov 901fa50a355SKonstantin Belousov MPASS(vm_map_min(map) <= minv); 9029cf78c1cSKonstantin Belousov 9039cf78c1cSKonstantin Belousov if (minv >= maxv || minv + align >= maxv || maxv > vm_map_max(map)) { 9049cf78c1cSKonstantin Belousov uprintf("Invalid ELF segments layout\n"); 9059cf78c1cSKonstantin Belousov return (ENOEXEC); 9069cf78c1cSKonstantin Belousov } 9079cf78c1cSKonstantin Belousov 908fa50a355SKonstantin Belousov arc4rand(&rbase, sizeof(rbase), 0); 909fa50a355SKonstantin Belousov res = roundup(minv, (u_long)align) + rbase % (maxv - minv); 910fa50a355SKonstantin Belousov res &= ~((u_long)align - 1); 911fa50a355SKonstantin Belousov if (res >= maxv) 912fa50a355SKonstantin Belousov res -= align; 9139cf78c1cSKonstantin Belousov 914fa50a355SKonstantin Belousov KASSERT(res >= minv, 915fa50a355SKonstantin Belousov ("res %#lx < minv %#lx, maxv %#lx rbase %#lx", 916fa50a355SKonstantin Belousov res, minv, maxv, rbase)); 917fa50a355SKonstantin Belousov KASSERT(res < maxv, 918fa50a355SKonstantin Belousov ("res %#lx > maxv %#lx, minv %#lx rbase %#lx", 919fa50a355SKonstantin Belousov res, maxv, minv, rbase)); 9209cf78c1cSKonstantin Belousov 9219cf78c1cSKonstantin Belousov *resp = res; 9229cf78c1cSKonstantin Belousov return (0); 923fa50a355SKonstantin Belousov } 924fa50a355SKonstantin Belousov 92520e1174aSEdward Tomasz Napierala static int 92620e1174aSEdward Tomasz Napierala __elfN(enforce_limits)(struct image_params *imgp, const Elf_Ehdr *hdr, 927659a0041SJessica Clarke const Elf_Phdr *phdr) 92820e1174aSEdward Tomasz Napierala { 92920e1174aSEdward Tomasz Napierala struct vmspace *vmspace; 93020e1174aSEdward Tomasz Napierala const char *err_str; 93120e1174aSEdward Tomasz Napierala u_long text_size, data_size, total_size, text_addr, data_addr; 93220e1174aSEdward Tomasz Napierala u_long seg_size, seg_addr; 93320e1174aSEdward Tomasz Napierala int i; 93420e1174aSEdward Tomasz Napierala 93520e1174aSEdward Tomasz Napierala err_str = NULL; 93620e1174aSEdward Tomasz Napierala text_size = data_size = total_size = text_addr = data_addr = 0; 93720e1174aSEdward Tomasz Napierala 93820e1174aSEdward Tomasz Napierala for (i = 0; i < hdr->e_phnum; i++) { 93920e1174aSEdward Tomasz Napierala if (phdr[i].p_type != PT_LOAD || phdr[i].p_memsz == 0) 94020e1174aSEdward Tomasz Napierala continue; 94120e1174aSEdward Tomasz Napierala 942659a0041SJessica Clarke seg_addr = trunc_page(phdr[i].p_vaddr + imgp->et_dyn_addr); 94320e1174aSEdward Tomasz Napierala seg_size = round_page(phdr[i].p_memsz + 944659a0041SJessica Clarke phdr[i].p_vaddr + imgp->et_dyn_addr - seg_addr); 94520e1174aSEdward Tomasz Napierala 94620e1174aSEdward Tomasz Napierala /* 94720e1174aSEdward Tomasz Napierala * Make the largest executable segment the official 94820e1174aSEdward Tomasz Napierala * text segment and all others data. 94920e1174aSEdward Tomasz Napierala * 95020e1174aSEdward Tomasz Napierala * Note that obreak() assumes that data_addr + data_size == end 95120e1174aSEdward Tomasz Napierala * of data load area, and the ELF file format expects segments 95220e1174aSEdward Tomasz Napierala * to be sorted by address. If multiple data segments exist, 95320e1174aSEdward Tomasz Napierala * the last one will be used. 95420e1174aSEdward Tomasz Napierala */ 95520e1174aSEdward Tomasz Napierala 95620e1174aSEdward Tomasz Napierala if ((phdr[i].p_flags & PF_X) != 0 && text_size < seg_size) { 95720e1174aSEdward Tomasz Napierala text_size = seg_size; 95820e1174aSEdward Tomasz Napierala text_addr = seg_addr; 95920e1174aSEdward Tomasz Napierala } else { 96020e1174aSEdward Tomasz Napierala data_size = seg_size; 96120e1174aSEdward Tomasz Napierala data_addr = seg_addr; 96220e1174aSEdward Tomasz Napierala } 96320e1174aSEdward Tomasz Napierala total_size += seg_size; 96420e1174aSEdward Tomasz Napierala } 96520e1174aSEdward Tomasz Napierala 96620e1174aSEdward Tomasz Napierala if (data_addr == 0 && data_size == 0) { 96720e1174aSEdward Tomasz Napierala data_addr = text_addr; 96820e1174aSEdward Tomasz Napierala data_size = text_size; 96920e1174aSEdward Tomasz Napierala } 97020e1174aSEdward Tomasz Napierala 97120e1174aSEdward Tomasz Napierala /* 97220e1174aSEdward Tomasz Napierala * Check limits. It should be safe to check the 97320e1174aSEdward Tomasz Napierala * limits after loading the segments since we do 97420e1174aSEdward Tomasz Napierala * not actually fault in all the segments pages. 97520e1174aSEdward Tomasz Napierala */ 97620e1174aSEdward Tomasz Napierala PROC_LOCK(imgp->proc); 97720e1174aSEdward Tomasz Napierala if (data_size > lim_cur_proc(imgp->proc, RLIMIT_DATA)) 97820e1174aSEdward Tomasz Napierala err_str = "Data segment size exceeds process limit"; 97920e1174aSEdward Tomasz Napierala else if (text_size > maxtsiz) 98020e1174aSEdward Tomasz Napierala err_str = "Text segment size exceeds system limit"; 98120e1174aSEdward Tomasz Napierala else if (total_size > lim_cur_proc(imgp->proc, RLIMIT_VMEM)) 98220e1174aSEdward Tomasz Napierala err_str = "Total segment size exceeds process limit"; 98320e1174aSEdward Tomasz Napierala else if (racct_set(imgp->proc, RACCT_DATA, data_size) != 0) 98420e1174aSEdward Tomasz Napierala err_str = "Data segment size exceeds resource limit"; 98520e1174aSEdward Tomasz Napierala else if (racct_set(imgp->proc, RACCT_VMEM, total_size) != 0) 98620e1174aSEdward Tomasz Napierala err_str = "Total segment size exceeds resource limit"; 98720e1174aSEdward Tomasz Napierala PROC_UNLOCK(imgp->proc); 98820e1174aSEdward Tomasz Napierala if (err_str != NULL) { 98920e1174aSEdward Tomasz Napierala uprintf("%s\n", err_str); 99020e1174aSEdward Tomasz Napierala return (ENOMEM); 99120e1174aSEdward Tomasz Napierala } 99220e1174aSEdward Tomasz Napierala 99320e1174aSEdward Tomasz Napierala vmspace = imgp->proc->p_vmspace; 99420e1174aSEdward Tomasz Napierala vmspace->vm_tsize = text_size >> PAGE_SHIFT; 99520e1174aSEdward Tomasz Napierala vmspace->vm_taddr = (caddr_t)(uintptr_t)text_addr; 99620e1174aSEdward Tomasz Napierala vmspace->vm_dsize = data_size >> PAGE_SHIFT; 99720e1174aSEdward Tomasz Napierala vmspace->vm_daddr = (caddr_t)(uintptr_t)data_addr; 99820e1174aSEdward Tomasz Napierala 99920e1174aSEdward Tomasz Napierala return (0); 100020e1174aSEdward Tomasz Napierala } 100120e1174aSEdward Tomasz Napierala 100209c78d53SEdward Tomasz Napierala static int 100309c78d53SEdward Tomasz Napierala __elfN(get_interp)(struct image_params *imgp, const Elf_Phdr *phdr, 100409c78d53SEdward Tomasz Napierala char **interpp, bool *free_interpp) 100509c78d53SEdward Tomasz Napierala { 100609c78d53SEdward Tomasz Napierala struct thread *td; 100709c78d53SEdward Tomasz Napierala char *interp; 100809c78d53SEdward Tomasz Napierala int error, interp_name_len; 100909c78d53SEdward Tomasz Napierala 101009c78d53SEdward Tomasz Napierala KASSERT(phdr->p_type == PT_INTERP, 101109c78d53SEdward Tomasz Napierala ("%s: p_type %u != PT_INTERP", __func__, phdr->p_type)); 1012b65ca345SEdward Tomasz Napierala ASSERT_VOP_LOCKED(imgp->vp, __func__); 101309c78d53SEdward Tomasz Napierala 101409c78d53SEdward Tomasz Napierala td = curthread; 101509c78d53SEdward Tomasz Napierala 101609c78d53SEdward Tomasz Napierala /* Path to interpreter */ 101709c78d53SEdward Tomasz Napierala if (phdr->p_filesz < 2 || phdr->p_filesz > MAXPATHLEN) { 101809c78d53SEdward Tomasz Napierala uprintf("Invalid PT_INTERP\n"); 101909c78d53SEdward Tomasz Napierala return (ENOEXEC); 102009c78d53SEdward Tomasz Napierala } 102109c78d53SEdward Tomasz Napierala 102209c78d53SEdward Tomasz Napierala interp_name_len = phdr->p_filesz; 102309c78d53SEdward Tomasz Napierala if (phdr->p_offset > PAGE_SIZE || 102409c78d53SEdward Tomasz Napierala interp_name_len > PAGE_SIZE - phdr->p_offset) { 10250ddfdc60SKonstantin Belousov /* 1026f1f81d3bSKonstantin Belousov * The vnode lock might be needed by the pagedaemon to 10270ddfdc60SKonstantin Belousov * clean pages owned by the vnode. Do not allow sleep 10280ddfdc60SKonstantin Belousov * waiting for memory with the vnode locked, instead 10290ddfdc60SKonstantin Belousov * try non-sleepable allocation first, and if it 10300ddfdc60SKonstantin Belousov * fails, go to the slow path were we drop the lock 1031f1f81d3bSKonstantin Belousov * and do M_WAITOK. A text reference prevents 1032f1f81d3bSKonstantin Belousov * modifications to the vnode content. 10330ddfdc60SKonstantin Belousov */ 10342d6b8546SKonstantin Belousov interp = malloc(interp_name_len + 1, M_TEMP, M_NOWAIT); 10352d6b8546SKonstantin Belousov if (interp == NULL) { 1036b249ce48SMateusz Guzik VOP_UNLOCK(imgp->vp); 103709c78d53SEdward Tomasz Napierala interp = malloc(interp_name_len + 1, M_TEMP, M_WAITOK); 103878022527SKonstantin Belousov vn_lock(imgp->vp, LK_SHARED | LK_RETRY); 10392d6b8546SKonstantin Belousov } 10400ddfdc60SKonstantin Belousov 104109c78d53SEdward Tomasz Napierala error = vn_rdwr(UIO_READ, imgp->vp, interp, 104209c78d53SEdward Tomasz Napierala interp_name_len, phdr->p_offset, 104309c78d53SEdward Tomasz Napierala UIO_SYSSPACE, IO_NODELOCKED, td->td_ucred, 104409c78d53SEdward Tomasz Napierala NOCRED, NULL, td); 104509c78d53SEdward Tomasz Napierala if (error != 0) { 104609c78d53SEdward Tomasz Napierala free(interp, M_TEMP); 104709c78d53SEdward Tomasz Napierala uprintf("i/o error PT_INTERP %d\n", error); 104809c78d53SEdward Tomasz Napierala return (error); 104909c78d53SEdward Tomasz Napierala } 105009c78d53SEdward Tomasz Napierala interp[interp_name_len] = '\0'; 105109c78d53SEdward Tomasz Napierala 105209c78d53SEdward Tomasz Napierala *interpp = interp; 105309c78d53SEdward Tomasz Napierala *free_interpp = true; 105409c78d53SEdward Tomasz Napierala return (0); 105509c78d53SEdward Tomasz Napierala } 105609c78d53SEdward Tomasz Napierala 105709c78d53SEdward Tomasz Napierala interp = __DECONST(char *, imgp->image_header) + phdr->p_offset; 105809c78d53SEdward Tomasz Napierala if (interp[interp_name_len - 1] != '\0') { 105909c78d53SEdward Tomasz Napierala uprintf("Invalid PT_INTERP\n"); 106009c78d53SEdward Tomasz Napierala return (ENOEXEC); 106109c78d53SEdward Tomasz Napierala } 106209c78d53SEdward Tomasz Napierala 106309c78d53SEdward Tomasz Napierala *interpp = interp; 106409c78d53SEdward Tomasz Napierala *free_interpp = false; 106509c78d53SEdward Tomasz Napierala return (0); 106609c78d53SEdward Tomasz Napierala } 106709c78d53SEdward Tomasz Napierala 10689274fb35SEdward Tomasz Napierala static int 10699274fb35SEdward Tomasz Napierala __elfN(load_interp)(struct image_params *imgp, const Elf_Brandinfo *brand_info, 10709274fb35SEdward Tomasz Napierala const char *interp, u_long *addr, u_long *entry) 10719274fb35SEdward Tomasz Napierala { 10729274fb35SEdward Tomasz Napierala int error; 10739274fb35SEdward Tomasz Napierala 10749274fb35SEdward Tomasz Napierala if (brand_info->interp_newpath != NULL && 10759274fb35SEdward Tomasz Napierala (brand_info->interp_path == NULL || 10769274fb35SEdward Tomasz Napierala strcmp(interp, brand_info->interp_path) == 0)) { 10779274fb35SEdward Tomasz Napierala error = __elfN(load_file)(imgp->proc, 10789274fb35SEdward Tomasz Napierala brand_info->interp_newpath, addr, entry); 10799274fb35SEdward Tomasz Napierala if (error == 0) 10809274fb35SEdward Tomasz Napierala return (0); 10819274fb35SEdward Tomasz Napierala } 10829274fb35SEdward Tomasz Napierala 10839274fb35SEdward Tomasz Napierala error = __elfN(load_file)(imgp->proc, interp, addr, entry); 10849274fb35SEdward Tomasz Napierala if (error == 0) 10859274fb35SEdward Tomasz Napierala return (0); 10869274fb35SEdward Tomasz Napierala 10879274fb35SEdward Tomasz Napierala uprintf("ELF interpreter %s not found, error %d\n", interp, error); 10889274fb35SEdward Tomasz Napierala return (error); 10899274fb35SEdward Tomasz Napierala } 10909274fb35SEdward Tomasz Napierala 1091fa50a355SKonstantin Belousov /* 1092fa50a355SKonstantin Belousov * Impossible et_dyn_addr initial value indicating that the real base 1093fa50a355SKonstantin Belousov * must be calculated later with some randomization applied. 1094fa50a355SKonstantin Belousov */ 1095fa50a355SKonstantin Belousov #define ET_DYN_ADDR_RAND 1 1096fa50a355SKonstantin Belousov 1097303b270bSEivind Eklund static int 10983ebc1248SPeter Wemm __CONCAT(exec_, __elfN(imgact))(struct image_params *imgp) 1099e1743d02SSøren Schmidt { 11006c775eb6SKonstantin Belousov struct thread *td; 11016c775eb6SKonstantin Belousov const Elf_Ehdr *hdr; 110232c01de2SDmitry Chagin const Elf_Phdr *phdr; 1103e5e6093bSAlan Cox Elf_Auxargs *elf_auxargs; 11045856e12eSJohn Dyson struct vmspace *vmspace; 1105fa50a355SKonstantin Belousov vm_map_t map; 11069274fb35SEdward Tomasz Napierala char *interp; 1107d1dbc694SJohn Polstra Elf_Brandinfo *brand_info; 11085fe3ed62SJake Burkholder struct sysentvec *sv; 1109659a0041SJessica Clarke u_long addr, baddr, entry, proghdr; 1110e499988fSKonstantin Belousov u_long maxalign, maxsalign, mapsz, maxv, maxv1, anon_loc; 1111cefb93f2SKonstantin Belousov uint32_t fctl0; 11126c775eb6SKonstantin Belousov int32_t osrel; 111309c78d53SEdward Tomasz Napierala bool free_interp; 11149274fb35SEdward Tomasz Napierala int error, i, n; 11156c775eb6SKonstantin Belousov 11166c775eb6SKonstantin Belousov hdr = (const Elf_Ehdr *)imgp->image_header; 1117e1743d02SSøren Schmidt 1118e1743d02SSøren Schmidt /* 1119e1743d02SSøren Schmidt * Do we have a valid ELF header ? 1120900b28f9SMaxim Sobolev * 1121900b28f9SMaxim Sobolev * Only allow ET_EXEC & ET_DYN here, reject ET_DYN later 1122900b28f9SMaxim Sobolev * if particular brand doesn't support it. 1123e1743d02SSøren Schmidt */ 1124900b28f9SMaxim Sobolev if (__elfN(check_header)(hdr) != 0 || 1125900b28f9SMaxim Sobolev (hdr->e_type != ET_EXEC && hdr->e_type != ET_DYN)) 1126a7cddfedSJake Burkholder return (-1); 1127e1743d02SSøren Schmidt 1128e1743d02SSøren Schmidt /* 1129e1743d02SSøren Schmidt * From here on down, we return an errno, not -1, as we've 1130e1743d02SSøren Schmidt * detected an ELF file. 1131e1743d02SSøren Schmidt */ 1132e1743d02SSøren Schmidt 11337de1bc13SKonstantin Belousov if (!__elfN(phdr_in_zero_page)(hdr)) { 11346b16d664SEd Maste uprintf("Program headers not in the first page\n"); 1135a7cddfedSJake Burkholder return (ENOEXEC); 1136e1743d02SSøren Schmidt } 113752c24af7SPeter Wemm phdr = (const Elf_Phdr *)(imgp->image_header + hdr->e_phoff); 11386b16d664SEd Maste if (!aligned(phdr, Elf_Addr)) { 11396b16d664SEd Maste uprintf("Unaligned program headers\n"); 114093d1c728SKonstantin Belousov return (ENOEXEC); 11416b16d664SEd Maste } 11426c775eb6SKonstantin Belousov 11436c775eb6SKonstantin Belousov n = error = 0; 11447564c4adSKonstantin Belousov baddr = 0; 11456c775eb6SKonstantin Belousov osrel = 0; 1146cefb93f2SKonstantin Belousov fctl0 = 0; 11476c775eb6SKonstantin Belousov entry = proghdr = 0; 11489274fb35SEdward Tomasz Napierala interp = NULL; 114909c78d53SEdward Tomasz Napierala free_interp = false; 11506c775eb6SKonstantin Belousov td = curthread; 1151714d6d09SKonstantin Belousov 1152714d6d09SKonstantin Belousov /* 1153714d6d09SKonstantin Belousov * Somewhat arbitrary, limit accepted max alignment for the 1154714d6d09SKonstantin Belousov * loadable segment to the max supported superpage size. Too 1155714d6d09SKonstantin Belousov * large alignment requests are not useful and are indicators 1156714d6d09SKonstantin Belousov * of corrupted or outright malicious binary. 1157714d6d09SKonstantin Belousov */ 1158fa50a355SKonstantin Belousov maxalign = PAGE_SIZE; 1159714d6d09SKonstantin Belousov maxsalign = PAGE_SIZE * 1024; 1160714d6d09SKonstantin Belousov for (i = MAXPAGESIZES - 1; i > 0; i--) { 1161f076dd3eSAlan Cox if (pagesizes[i] > maxsalign) { 1162714d6d09SKonstantin Belousov maxsalign = pagesizes[i]; 1163f076dd3eSAlan Cox break; 1164f076dd3eSAlan Cox } 1165714d6d09SKonstantin Belousov } 1166714d6d09SKonstantin Belousov 1167fa50a355SKonstantin Belousov mapsz = 0; 11686c775eb6SKonstantin Belousov 11695fe3ed62SJake Burkholder for (i = 0; i < hdr->e_phnum; i++) { 1170291c06a1SKonstantin Belousov switch (phdr[i].p_type) { 1171291c06a1SKonstantin Belousov case PT_LOAD: 11727564c4adSKonstantin Belousov if (n == 0) 11737564c4adSKonstantin Belousov baddr = phdr[i].p_vaddr; 117436df8f54SKonstantin Belousov if (!powerof2(phdr[i].p_align) || 117536df8f54SKonstantin Belousov phdr[i].p_align > maxsalign) { 1176714d6d09SKonstantin Belousov uprintf("Invalid segment alignment\n"); 1177714d6d09SKonstantin Belousov error = ENOEXEC; 1178714d6d09SKonstantin Belousov goto ret; 1179714d6d09SKonstantin Belousov } 1180fa50a355SKonstantin Belousov if (phdr[i].p_align > maxalign) 1181fa50a355SKonstantin Belousov maxalign = phdr[i].p_align; 1182bf839416SKonstantin Belousov if (mapsz + phdr[i].p_memsz < mapsz) { 1183bf839416SKonstantin Belousov uprintf("Mapsize overflow\n"); 1184bf839416SKonstantin Belousov error = ENOEXEC; 1185bf839416SKonstantin Belousov goto ret; 1186bf839416SKonstantin Belousov } 1187fa50a355SKonstantin Belousov mapsz += phdr[i].p_memsz; 11887564c4adSKonstantin Belousov n++; 11899bcd7482SEdward Tomasz Napierala 11909bcd7482SEdward Tomasz Napierala /* 11919bcd7482SEdward Tomasz Napierala * If this segment contains the program headers, 11929bcd7482SEdward Tomasz Napierala * remember their virtual address for the AT_PHDR 11939bcd7482SEdward Tomasz Napierala * aux entry. Static binaries don't usually include 11949bcd7482SEdward Tomasz Napierala * a PT_PHDR entry. 11959bcd7482SEdward Tomasz Napierala */ 11969bcd7482SEdward Tomasz Napierala if (phdr[i].p_offset == 0 && 119788dd7a0aSKonstantin Belousov hdr->e_phoff + hdr->e_phnum * hdr->e_phentsize <= 119888dd7a0aSKonstantin Belousov phdr[i].p_filesz) 11999bcd7482SEdward Tomasz Napierala proghdr = phdr[i].p_vaddr + hdr->e_phoff; 1200291c06a1SKonstantin Belousov break; 1201291c06a1SKonstantin Belousov case PT_INTERP: 1202e5e6093bSAlan Cox /* Path to interpreter */ 1203d3ee0a15SJonathan T. Looney if (interp != NULL) { 1204d3ee0a15SJonathan T. Looney uprintf("Multiple PT_INTERP headers\n"); 1205d3ee0a15SJonathan T. Looney error = ENOEXEC; 1206d3ee0a15SJonathan T. Looney goto ret; 1207d3ee0a15SJonathan T. Looney } 120809c78d53SEdward Tomasz Napierala error = __elfN(get_interp)(imgp, &phdr[i], &interp, 120909c78d53SEdward Tomasz Napierala &free_interp); 121009c78d53SEdward Tomasz Napierala if (error != 0) 12116c775eb6SKonstantin Belousov goto ret; 1212291c06a1SKonstantin Belousov break; 1213291c06a1SKonstantin Belousov case PT_GNU_STACK: 1214fbafa98aSEd Maste if (__elfN(nxstack)) { 1215291c06a1SKonstantin Belousov imgp->stack_prot = 1216291c06a1SKonstantin Belousov __elfN(trans_prot)(phdr[i].p_flags); 1217fbafa98aSEd Maste if ((imgp->stack_prot & VM_PROT_RW) != 1218fbafa98aSEd Maste VM_PROT_RW) { 1219fbafa98aSEd Maste uprintf("Invalid PT_GNU_STACK\n"); 1220fbafa98aSEd Maste error = ENOEXEC; 1221fbafa98aSEd Maste goto ret; 1222fbafa98aSEd Maste } 1223fbafa98aSEd Maste } 1224316b3843SKonstantin Belousov imgp->stack_sz = phdr[i].p_memsz; 1225291c06a1SKonstantin Belousov break; 12269bcd7482SEdward Tomasz Napierala case PT_PHDR: /* Program header table info */ 12279bcd7482SEdward Tomasz Napierala proghdr = phdr[i].p_vaddr; 12289bcd7482SEdward Tomasz Napierala break; 12293ebc1248SPeter Wemm } 12303ebc1248SPeter Wemm } 12313ebc1248SPeter Wemm 123209c78d53SEdward Tomasz Napierala brand_info = __elfN(get_brandinfo)(imgp, interp, &osrel, &fctl0); 12335fe3ed62SJake Burkholder if (brand_info == NULL) { 12345fe3ed62SJake Burkholder uprintf("ELF binary type \"%u\" not known.\n", 12355fe3ed62SJake Burkholder hdr->e_ident[EI_OSABI]); 12366c775eb6SKonstantin Belousov error = ENOEXEC; 12376c775eb6SKonstantin Belousov goto ret; 12383ebc1248SPeter Wemm } 1239fa50a355SKonstantin Belousov sv = brand_info->sysvec; 1240ab02d85fSKonstantin Belousov if (hdr->e_type == ET_DYN) { 12416b16d664SEd Maste if ((brand_info->flags & BI_CAN_EXEC_DYN) == 0) { 12426b16d664SEd Maste uprintf("Cannot execute shared object\n"); 12436c775eb6SKonstantin Belousov error = ENOEXEC; 12446c775eb6SKonstantin Belousov goto ret; 12456b16d664SEd Maste } 12467564c4adSKonstantin Belousov /* 12477564c4adSKonstantin Belousov * Honour the base load address from the dso if it is 12487564c4adSKonstantin Belousov * non-zero for some reason. 12497564c4adSKonstantin Belousov */ 1250fa50a355SKonstantin Belousov if (baddr == 0) { 1251fa50a355SKonstantin Belousov if ((sv->sv_flags & SV_ASLR) == 0 || 1252fa50a355SKonstantin Belousov (fctl0 & NT_FREEBSD_FCTL_ASLR_DISABLE) != 0) 1253659a0041SJessica Clarke imgp->et_dyn_addr = __elfN(pie_base); 1254fa50a355SKonstantin Belousov else if ((__elfN(pie_aslr_enabled) && 1255fa50a355SKonstantin Belousov (imgp->proc->p_flag2 & P2_ASLR_DISABLE) == 0) || 1256fa50a355SKonstantin Belousov (imgp->proc->p_flag2 & P2_ASLR_ENABLE) != 0) 1257659a0041SJessica Clarke imgp->et_dyn_addr = ET_DYN_ADDR_RAND; 1258fa50a355SKonstantin Belousov else 1259659a0041SJessica Clarke imgp->et_dyn_addr = __elfN(pie_base); 126077ebe276SEd Maste } 1261fa50a355SKonstantin Belousov } 12623ebc1248SPeter Wemm 126360bb3943SAlan Cox /* 126460bb3943SAlan Cox * Avoid a possible deadlock if the current address space is destroyed 126560bb3943SAlan Cox * and that address space maps the locked vnode. In the common case, 126660bb3943SAlan Cox * the locked vnode's v_usecount is decremented but remains greater 126760bb3943SAlan Cox * than zero. Consequently, the vnode lock is not needed by vrele(). 126860bb3943SAlan Cox * However, in cases where the vnode lock is external, such as nullfs, 126960bb3943SAlan Cox * v_usecount may become zero. 12701dfab802SAlan Cox * 12711dfab802SAlan Cox * The VV_TEXT flag prevents modifications to the executable while 12721dfab802SAlan Cox * the vnode is unlocked. 127360bb3943SAlan Cox */ 1274b249ce48SMateusz Guzik VOP_UNLOCK(imgp->vp); 127560bb3943SAlan Cox 1276fa50a355SKonstantin Belousov /* 1277fa50a355SKonstantin Belousov * Decide whether to enable randomization of user mappings. 1278fa50a355SKonstantin Belousov * First, reset user preferences for the setid binaries. 1279fa50a355SKonstantin Belousov * Then, account for the support of the randomization by the 1280fa50a355SKonstantin Belousov * ABI, by user preferences, and make special treatment for 1281fa50a355SKonstantin Belousov * PIE binaries. 1282fa50a355SKonstantin Belousov */ 1283fa50a355SKonstantin Belousov if (imgp->credential_setid) { 1284fa50a355SKonstantin Belousov PROC_LOCK(imgp->proc); 1285796a8e1aSKonstantin Belousov imgp->proc->p_flag2 &= ~(P2_ASLR_ENABLE | P2_ASLR_DISABLE | 1286796a8e1aSKonstantin Belousov P2_WXORX_DISABLE | P2_WXORX_ENABLE_EXEC); 1287fa50a355SKonstantin Belousov PROC_UNLOCK(imgp->proc); 1288fa50a355SKonstantin Belousov } 1289fa50a355SKonstantin Belousov if ((sv->sv_flags & SV_ASLR) == 0 || 1290fa50a355SKonstantin Belousov (imgp->proc->p_flag2 & P2_ASLR_DISABLE) != 0 || 1291fa50a355SKonstantin Belousov (fctl0 & NT_FREEBSD_FCTL_ASLR_DISABLE) != 0) { 1292659a0041SJessica Clarke KASSERT(imgp->et_dyn_addr != ET_DYN_ADDR_RAND, 1293659a0041SJessica Clarke ("imgp->et_dyn_addr == RAND and !ASLR")); 1294fa50a355SKonstantin Belousov } else if ((imgp->proc->p_flag2 & P2_ASLR_ENABLE) != 0 || 1295fa50a355SKonstantin Belousov (__elfN(aslr_enabled) && hdr->e_type == ET_EXEC) || 1296659a0041SJessica Clarke imgp->et_dyn_addr == ET_DYN_ADDR_RAND) { 1297fa50a355SKonstantin Belousov imgp->map_flags |= MAP_ASLR; 1298fa50a355SKonstantin Belousov /* 1299fa50a355SKonstantin Belousov * If user does not care about sbrk, utilize the bss 1300fa50a355SKonstantin Belousov * grow region for mappings as well. We can select 1301fa50a355SKonstantin Belousov * the base for the image anywere and still not suffer 1302fa50a355SKonstantin Belousov * from the fragmentation. 1303fa50a355SKonstantin Belousov */ 1304fa50a355SKonstantin Belousov if (!__elfN(aslr_honor_sbrk) || 1305fa50a355SKonstantin Belousov (imgp->proc->p_flag2 & P2_ASLR_IGNSTART) != 0) 1306fa50a355SKonstantin Belousov imgp->map_flags |= MAP_ASLR_IGNSTART; 13071811c1e9SMark Johnston if (__elfN(aslr_stack)) 13081811c1e9SMark Johnston imgp->map_flags |= MAP_ASLR_STACK; 1309939f0b63SKornel Dulęba if (__elfN(aslr_shared_page)) 1310939f0b63SKornel Dulęba imgp->imgp_flags |= IMGP_ASLR_SHARED_PAGE; 1311fa50a355SKonstantin Belousov } 1312fa50a355SKonstantin Belousov 1313796a8e1aSKonstantin Belousov if ((!__elfN(allow_wx) && (fctl0 & NT_FREEBSD_FCTL_WXNEEDED) == 0 && 1314796a8e1aSKonstantin Belousov (imgp->proc->p_flag2 & P2_WXORX_DISABLE) == 0) || 1315796a8e1aSKonstantin Belousov (imgp->proc->p_flag2 & P2_WXORX_ENABLE_EXEC) != 0) 13162e1c94aaSKonstantin Belousov imgp->map_flags |= MAP_WXORX; 13172e1c94aaSKonstantin Belousov 131889b57fcfSKonstantin Belousov error = exec_new_vmspace(imgp, sv); 1319fa50a355SKonstantin Belousov 132019059a13SJohn Baldwin imgp->proc->p_sysent = sv; 1321615f22b2SDmitry Chagin imgp->proc->p_elf_brandinfo = brand_info; 1322e1743d02SSøren Schmidt 13231811c1e9SMark Johnston vmspace = imgp->proc->p_vmspace; 13241811c1e9SMark Johnston map = &vmspace->vm_map; 13251811c1e9SMark Johnston maxv = sv->sv_usrstack; 13261811c1e9SMark Johnston if ((imgp->map_flags & MAP_ASLR_STACK) == 0) 13271811c1e9SMark Johnston maxv -= lim_max(td, RLIMIT_STACK); 13281811c1e9SMark Johnston if (error == 0 && mapsz >= maxv - vm_map_min(map)) { 1329bf839416SKonstantin Belousov uprintf("Excessive mapping size\n"); 1330bf839416SKonstantin Belousov error = ENOEXEC; 1331bf839416SKonstantin Belousov } 1332bf839416SKonstantin Belousov 1333659a0041SJessica Clarke if (error == 0 && imgp->et_dyn_addr == ET_DYN_ADDR_RAND) { 1334fa50a355SKonstantin Belousov KASSERT((map->flags & MAP_ASLR) != 0, 1335fa50a355SKonstantin Belousov ("ET_DYN_ADDR_RAND but !MAP_ASLR")); 13369cf78c1cSKonstantin Belousov error = __CONCAT(rnd_, __elfN(base))(map, 1337fa50a355SKonstantin Belousov vm_map_min(map) + mapsz + lim_max(td, RLIMIT_DATA), 1338fa50a355SKonstantin Belousov /* reserve half of the address space to interpreter */ 1339659a0041SJessica Clarke maxv / 2, maxalign, &imgp->et_dyn_addr); 1340fa50a355SKonstantin Belousov } 1341fa50a355SKonstantin Belousov 134278022527SKonstantin Belousov vn_lock(imgp->vp, LK_SHARED | LK_RETRY); 13436c775eb6SKonstantin Belousov if (error != 0) 13446c775eb6SKonstantin Belousov goto ret; 134560bb3943SAlan Cox 1346659a0041SJessica Clarke error = __elfN(load_sections)(imgp, hdr, phdr, imgp->et_dyn_addr, NULL); 1347292177e6SAlan Cox if (error != 0) 13486c775eb6SKonstantin Belousov goto ret; 1349e1743d02SSøren Schmidt 1350659a0041SJessica Clarke error = __elfN(enforce_limits)(imgp, hdr, phdr); 135120e1174aSEdward Tomasz Napierala if (error != 0) 135220e1174aSEdward Tomasz Napierala goto ret; 1353cac45152SMatthew Dillon 1354cac45152SMatthew Dillon /* 1355c460ac3aSPeter Wemm * We load the dynamic linker where a userland call 1356c460ac3aSPeter Wemm * to mmap(0, ...) would put it. The rationale behind this 1357c460ac3aSPeter Wemm * calculation is that it leaves room for the heap to grow to 1358c460ac3aSPeter Wemm * its maximum allowed size. 1359c460ac3aSPeter Wemm */ 13606c775eb6SKonstantin Belousov addr = round_page((vm_offset_t)vmspace->vm_daddr + lim_max(td, 1361292177e6SAlan Cox RLIMIT_DATA)); 1362fa50a355SKonstantin Belousov if ((map->flags & MAP_ASLR) != 0) { 1363fa50a355SKonstantin Belousov maxv1 = maxv / 2 + addr / 2; 13649cf78c1cSKonstantin Belousov error = __CONCAT(rnd_, __elfN(base))(map, addr, maxv1, 13653e00c11aSAlan Cox #if VM_NRESERVLEVEL > 0 13663e00c11aSAlan Cox pagesizes[VM_NRESERVLEVEL] != 0 ? 13673e00c11aSAlan Cox /* Align anon_loc to the largest superpage size. */ 13683e00c11aSAlan Cox pagesizes[VM_NRESERVLEVEL] : 13693e00c11aSAlan Cox #endif 13703e00c11aSAlan Cox pagesizes[0], &anon_loc); 13719cf78c1cSKonstantin Belousov if (error != 0) 13729cf78c1cSKonstantin Belousov goto ret; 1373e499988fSKonstantin Belousov map->anon_loc = anon_loc; 1374fa50a355SKonstantin Belousov } else { 1375fa50a355SKonstantin Belousov map->anon_loc = addr; 1376fa50a355SKonstantin Belousov } 1377e1743d02SSøren Schmidt 1378659a0041SJessica Clarke entry = (u_long)hdr->e_entry + imgp->et_dyn_addr; 1379ea5a2b2eSSøren Schmidt imgp->entry_addr = entry; 1380ea5a2b2eSSøren Schmidt 1381eb32c1c7SAndrew Turner if (sv->sv_protect != NULL) 1382eb32c1c7SAndrew Turner sv->sv_protect(imgp, SVP_IMAGE); 1383eb32c1c7SAndrew Turner 138460bb3943SAlan Cox if (interp != NULL) { 1385b249ce48SMateusz Guzik VOP_UNLOCK(imgp->vp); 1386fa50a355SKonstantin Belousov if ((map->flags & MAP_ASLR) != 0) { 138741032835SJason A. Harmening /* Assume that interpreter fits into 1/4 of AS */ 1388fa50a355SKonstantin Belousov maxv1 = maxv / 2 + addr / 2; 13899cf78c1cSKonstantin Belousov error = __CONCAT(rnd_, __elfN(base))(map, addr, 13909cf78c1cSKonstantin Belousov maxv1, PAGE_SIZE, &addr); 1391fa50a355SKonstantin Belousov } 13929cf78c1cSKonstantin Belousov if (error == 0) { 13939cf78c1cSKonstantin Belousov error = __elfN(load_interp)(imgp, brand_info, interp, 13949cf78c1cSKonstantin Belousov &addr, &imgp->entry_addr); 13959cf78c1cSKonstantin Belousov } 139678022527SKonstantin Belousov vn_lock(imgp->vp, LK_SHARED | LK_RETRY); 13979274fb35SEdward Tomasz Napierala if (error != 0) 13986c775eb6SKonstantin Belousov goto ret; 139995c807cfSRobert Watson } else 1400659a0041SJessica Clarke addr = imgp->et_dyn_addr; 1401ea5a2b2eSSøren Schmidt 14021811c1e9SMark Johnston error = exec_map_stack(imgp); 14031811c1e9SMark Johnston if (error != 0) 14041811c1e9SMark Johnston goto ret; 14051811c1e9SMark Johnston 1406e1743d02SSøren Schmidt /* 1407e3532331SJohn Baldwin * Construct auxargs table (used by the copyout_auxargs routine) 1408e1743d02SSøren Schmidt */ 14092d6b8546SKonstantin Belousov elf_auxargs = malloc(sizeof(Elf_Auxargs), M_TEMP, M_NOWAIT); 14102d6b8546SKonstantin Belousov if (elf_auxargs == NULL) { 1411b249ce48SMateusz Guzik VOP_UNLOCK(imgp->vp); 1412a163d034SWarner Losh elf_auxargs = malloc(sizeof(Elf_Auxargs), M_TEMP, M_WAITOK); 141378022527SKonstantin Belousov vn_lock(imgp->vp, LK_SHARED | LK_RETRY); 14142d6b8546SKonstantin Belousov } 1415e1743d02SSøren Schmidt elf_auxargs->execfd = -1; 1416659a0041SJessica Clarke elf_auxargs->phdr = proghdr + imgp->et_dyn_addr; 1417e1743d02SSøren Schmidt elf_auxargs->phent = hdr->e_phentsize; 1418e1743d02SSøren Schmidt elf_auxargs->phnum = hdr->e_phnum; 1419e1743d02SSøren Schmidt elf_auxargs->pagesz = PAGE_SIZE; 1420e1743d02SSøren Schmidt elf_auxargs->base = addr; 1421e1743d02SSøren Schmidt elf_auxargs->flags = 0; 1422e1743d02SSøren Schmidt elf_auxargs->entry = entry; 1423d36eec69SWarner Losh elf_auxargs->hdr_eflags = hdr->e_flags; 1424e1743d02SSøren Schmidt 1425e1743d02SSøren Schmidt imgp->auxargs = elf_auxargs; 1426e1743d02SSøren Schmidt imgp->interpreted = 0; 1427a0ea661fSNathan Whitehorn imgp->reloc_base = addr; 142832c01de2SDmitry Chagin imgp->proc->p_osrel = osrel; 1429cefb93f2SKonstantin Belousov imgp->proc->p_fctl0 = fctl0; 1430885f13dcSJohn Baldwin imgp->proc->p_elf_flags = hdr->e_flags; 1431f231de47SKonstantin Belousov 14326c775eb6SKonstantin Belousov ret: 1433b4b20492SKonstantin Belousov ASSERT_VOP_LOCKED(imgp->vp, "skipped relock"); 143409c78d53SEdward Tomasz Napierala if (free_interp) 143509c78d53SEdward Tomasz Napierala free(interp, M_TEMP); 1436a7cddfedSJake Burkholder return (error); 1437e1743d02SSøren Schmidt } 1438e1743d02SSøren Schmidt 1439ebf98866SMark Johnston #define elf_suword __CONCAT(suword, __ELF_WORD_SIZE) 14403ebc1248SPeter Wemm 144103b0d68cSJohn Baldwin int 1442d8010b11SJohn Baldwin __elfN(freebsd_copyout_auxargs)(struct image_params *imgp, uintptr_t base) 1443e1743d02SSøren Schmidt { 1444ecbb00a2SDoug Rabson Elf_Auxargs *args = (Elf_Auxargs *)imgp->auxargs; 14455f77b8a8SBrooks Davis Elf_Auxinfo *argarray, *pos; 1446361971fbSKornel Dulęba struct vmspace *vmspace; 1447ff41239fSKonstantin Belousov rlim_t stacksz; 1448326bf508SBrooks Davis int error, oc; 1449326bf508SBrooks Davis uint32_t bsdflags; 1450e1743d02SSøren Schmidt 14515f77b8a8SBrooks Davis argarray = pos = malloc(AT_COUNT * sizeof(*pos), M_TEMP, 14525f77b8a8SBrooks Davis M_WAITOK | M_ZERO); 1453e1743d02SSøren Schmidt 1454361971fbSKornel Dulęba vmspace = imgp->proc->p_vmspace; 1455361971fbSKornel Dulęba 145635c2a5a8SWarner Losh if (args->execfd != -1) 1457e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_EXECFD, args->execfd); 1458e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PHDR, args->phdr); 1459e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PHENT, args->phent); 1460e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PHNUM, args->phnum); 1461e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PAGESZ, args->pagesz); 1462e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_FLAGS, args->flags); 1463e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_ENTRY, args->entry); 1464e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_BASE, args->base); 1465d36eec69SWarner Losh AUXARGS_ENTRY(pos, AT_EHDRFLAGS, args->hdr_eflags); 14663ff06357SKonstantin Belousov if (imgp->execpathp != 0) 1467b24e6ac8SBrooks Davis AUXARGS_ENTRY_PTR(pos, AT_EXECPATH, imgp->execpathp); 1468b96bd95bSIan Lepore AUXARGS_ENTRY(pos, AT_OSRELDATE, 1469b96bd95bSIan Lepore imgp->proc->p_ucred->cr_prison->pr_osreldate); 1470ee235befSKonstantin Belousov if (imgp->canary != 0) { 1471b24e6ac8SBrooks Davis AUXARGS_ENTRY_PTR(pos, AT_CANARY, imgp->canary); 1472ee235befSKonstantin Belousov AUXARGS_ENTRY(pos, AT_CANARYLEN, imgp->canarylen); 1473ee235befSKonstantin Belousov } 1474ee235befSKonstantin Belousov AUXARGS_ENTRY(pos, AT_NCPUS, mp_ncpus); 1475ee235befSKonstantin Belousov if (imgp->pagesizes != 0) { 1476b24e6ac8SBrooks Davis AUXARGS_ENTRY_PTR(pos, AT_PAGESIZES, imgp->pagesizes); 1477ee235befSKonstantin Belousov AUXARGS_ENTRY(pos, AT_PAGESIZESLEN, imgp->pagesizeslen); 1478ee235befSKonstantin Belousov } 1479361971fbSKornel Dulęba if ((imgp->sysent->sv_flags & SV_TIMEKEEP) != 0) { 1480aea81038SKonstantin Belousov AUXARGS_ENTRY(pos, AT_TIMEKEEP, 1481361971fbSKornel Dulęba vmspace->vm_shp_base + imgp->sysent->sv_timekeep_offset); 1482aea81038SKonstantin Belousov } 148326d8f3e1SKonstantin Belousov AUXARGS_ENTRY(pos, AT_STACKPROT, imgp->sysent->sv_shared_page_obj 148426d8f3e1SKonstantin Belousov != NULL && imgp->stack_prot != 0 ? imgp->stack_prot : 148526d8f3e1SKonstantin Belousov imgp->sysent->sv_stackprot); 1486c2f37b92SJohn Baldwin if (imgp->sysent->sv_hwcap != NULL) 1487c2f37b92SJohn Baldwin AUXARGS_ENTRY(pos, AT_HWCAP, *imgp->sysent->sv_hwcap); 1488904d8c49SMichal Meloun if (imgp->sysent->sv_hwcap2 != NULL) 1489904d8c49SMichal Meloun AUXARGS_ENTRY(pos, AT_HWCAP2, *imgp->sysent->sv_hwcap2); 1490ff41239fSKonstantin Belousov bsdflags = 0; 1491ff41239fSKonstantin Belousov bsdflags |= __elfN(sigfastblock) ? ELF_BSDF_SIGFASTBLK : 0; 1492ff41239fSKonstantin Belousov oc = atomic_load_int(&vm_overcommit); 1493ff41239fSKonstantin Belousov bsdflags |= (oc & (SWAP_RESERVE_FORCE_ON | SWAP_RESERVE_RLIMIT_ON)) != 1494ff41239fSKonstantin Belousov 0 ? ELF_BSDF_VMNOOVERCOMMIT : 0; 1495ff41239fSKonstantin Belousov AUXARGS_ENTRY(pos, AT_BSDFLAGS, bsdflags); 14969df1c38bSBrooks Davis AUXARGS_ENTRY(pos, AT_ARGC, imgp->args->argc); 14979df1c38bSBrooks Davis AUXARGS_ENTRY_PTR(pos, AT_ARGV, imgp->argv); 14989df1c38bSBrooks Davis AUXARGS_ENTRY(pos, AT_ENVC, imgp->args->envc); 14999df1c38bSBrooks Davis AUXARGS_ENTRY_PTR(pos, AT_ENVV, imgp->envv); 15009df1c38bSBrooks Davis AUXARGS_ENTRY_PTR(pos, AT_PS_STRINGS, imgp->ps_strings); 1501361971fbSKornel Dulęba #ifdef RANDOM_FENESTRASX 1502361971fbSKornel Dulęba if ((imgp->sysent->sv_flags & SV_RNG_SEED_VER) != 0) { 1503361971fbSKornel Dulęba AUXARGS_ENTRY(pos, AT_FXRNG, 1504361971fbSKornel Dulęba vmspace->vm_shp_base + imgp->sysent->sv_fxrng_gen_offset); 1505361971fbSKornel Dulęba } 1506361971fbSKornel Dulęba #endif 1507361971fbSKornel Dulęba if ((imgp->sysent->sv_flags & SV_DSO_SIG) != 0 && __elfN(vdso) != 0) { 1508361971fbSKornel Dulęba AUXARGS_ENTRY(pos, AT_KPRELOAD, 1509361971fbSKornel Dulęba vmspace->vm_shp_base + imgp->sysent->sv_vdso_offset); 1510361971fbSKornel Dulęba } 1511ff41239fSKonstantin Belousov AUXARGS_ENTRY(pos, AT_USRSTACKBASE, round_page(vmspace->vm_stacktop)); 1512ff41239fSKonstantin Belousov stacksz = imgp->proc->p_limit->pl_rlimit[RLIMIT_STACK].rlim_cur; 1513ff41239fSKonstantin Belousov AUXARGS_ENTRY(pos, AT_USRSTACKLIM, stacksz); 1514e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_NULL, 0); 1515e1743d02SSøren Schmidt 1516e1743d02SSøren Schmidt free(imgp->auxargs, M_TEMP); 1517e1743d02SSøren Schmidt imgp->auxargs = NULL; 1518d8b2f079SBrooks Davis KASSERT(pos - argarray <= AT_COUNT, ("Too many auxargs")); 15195f77b8a8SBrooks Davis 1520d8010b11SJohn Baldwin error = copyout(argarray, (void *)base, sizeof(*argarray) * AT_COUNT); 15215f77b8a8SBrooks Davis free(argarray, M_TEMP); 152203b0d68cSJohn Baldwin return (error); 1523e3532331SJohn Baldwin } 1524e1743d02SSøren Schmidt 1525e3532331SJohn Baldwin int 152631174518SJohn Baldwin __elfN(freebsd_fixup)(uintptr_t *stack_base, struct image_params *imgp) 1527e3532331SJohn Baldwin { 1528e3532331SJohn Baldwin Elf_Addr *base; 1529e3532331SJohn Baldwin 1530e3532331SJohn Baldwin base = (Elf_Addr *)*stack_base; 15313ebc1248SPeter Wemm base--; 1532ebf98866SMark Johnston if (elf_suword(base, imgp->args->argc) == -1) 15335f77b8a8SBrooks Davis return (EFAULT); 153431174518SJohn Baldwin *stack_base = (uintptr_t)base; 1535a7cddfedSJake Burkholder return (0); 1536e1743d02SSøren Schmidt } 1537e1743d02SSøren Schmidt 1538e1743d02SSøren Schmidt /* 15398c64af4fSJohn Polstra * Code for generating ELF core dumps. 15408c64af4fSJohn Polstra */ 15418c64af4fSJohn Polstra 15424d77a549SAlfred Perlstein typedef void (*segment_callback)(vm_map_entry_t, void *); 15430ff27d31SJohn Polstra 15440ff27d31SJohn Polstra /* Closure for cb_put_phdr(). */ 15450ff27d31SJohn Polstra struct phdr_closure { 15460ff27d31SJohn Polstra Elf_Phdr *phdr; /* Program header to fill in */ 15470ff27d31SJohn Polstra Elf_Off offset; /* Offset of segment in core file */ 15480ff27d31SJohn Polstra }; 15490ff27d31SJohn Polstra 1550bd390213SMikolaj Golub struct note_info { 1551bd390213SMikolaj Golub int type; /* Note type. */ 15526b71405bSJohn Baldwin struct regset *regset; /* Register set. */ 1553bd390213SMikolaj Golub outfunc_t outfunc; /* Output function. */ 1554bd390213SMikolaj Golub void *outarg; /* Argument for the output function. */ 1555bd390213SMikolaj Golub size_t outsize; /* Output size. */ 1556bd390213SMikolaj Golub TAILQ_ENTRY(note_info) link; /* Link to the next note info. */ 1557bd390213SMikolaj Golub }; 1558bd390213SMikolaj Golub 1559bd390213SMikolaj Golub TAILQ_HEAD(note_info_list, note_info); 1560bd390213SMikolaj Golub 156178f57a9cSMark Johnston extern int compress_user_cores; 156278f57a9cSMark Johnston extern int compress_user_cores_level; 156378f57a9cSMark Johnston 15644d77a549SAlfred Perlstein static void cb_put_phdr(vm_map_entry_t, void *); 15654d77a549SAlfred Perlstein static void cb_size_segment(vm_map_entry_t, void *); 15665bc3c617SKonstantin Belousov static void each_dumpable_segment(struct thread *, segment_callback, void *, 15675bc3c617SKonstantin Belousov int); 1568aa14e9b7SMark Johnston static int __elfN(corehdr)(struct coredump_params *, int, void *, size_t, 15695bc3c617SKonstantin Belousov struct note_info_list *, size_t, int); 1570435754a5SEdward Tomasz Napierala static void __elfN(putnote)(struct thread *td, struct note_info *, struct sbuf *); 1571bd390213SMikolaj Golub 1572bd390213SMikolaj Golub static void __elfN(note_prpsinfo)(void *, struct sbuf *, size_t *); 1573bd390213SMikolaj Golub static void __elfN(note_threadmd)(void *, struct sbuf *, size_t *); 1574f1fca82eSMikolaj Golub static void __elfN(note_procstat_auxv)(void *, struct sbuf *, size_t *); 1575f1fca82eSMikolaj Golub static void __elfN(note_procstat_proc)(void *, struct sbuf *, size_t *); 1576f1fca82eSMikolaj Golub static void __elfN(note_procstat_psstrings)(void *, struct sbuf *, size_t *); 1577f1fca82eSMikolaj Golub static void note_procstat_files(void *, struct sbuf *, size_t *); 1578f1fca82eSMikolaj Golub static void note_procstat_groups(void *, struct sbuf *, size_t *); 1579f1fca82eSMikolaj Golub static void note_procstat_osrel(void *, struct sbuf *, size_t *); 1580f1fca82eSMikolaj Golub static void note_procstat_rlimit(void *, struct sbuf *, size_t *); 1581f1fca82eSMikolaj Golub static void note_procstat_umask(void *, struct sbuf *, size_t *); 1582f1fca82eSMikolaj Golub static void note_procstat_vmmap(void *, struct sbuf *, size_t *); 15838c64af4fSJohn Polstra 1584aa14e9b7SMark Johnston static int 158578f57a9cSMark Johnston core_compressed_write(void *base, size_t len, off_t offset, void *arg) 1586aa14e9b7SMark Johnston { 1587aa14e9b7SMark Johnston 1588aa14e9b7SMark Johnston return (core_write((struct coredump_params *)arg, base, len, offset, 1589f31695ccSMark Johnston UIO_SYSSPACE, NULL)); 1590aa14e9b7SMark Johnston } 1591aa14e9b7SMark Johnston 15928c64af4fSJohn Polstra int 1593e7228204SAlfred Perlstein __elfN(coredump)(struct thread *td, struct vnode *vp, off_t limit, int flags) 1594fca666a1SJulian Elischer { 1595247aba24SMarcel Moolenaar struct ucred *cred = td->td_ucred; 159686ffb3d1SKonstantin Belousov int compm, error = 0; 15970ff27d31SJohn Polstra struct sseg_closure seginfo; 1598bd390213SMikolaj Golub struct note_info_list notelst; 1599aa14e9b7SMark Johnston struct coredump_params params; 1600bd390213SMikolaj Golub struct note_info *ninfo; 1601aa14e9b7SMark Johnston void *hdr, *tmpbuf; 1602bd390213SMikolaj Golub size_t hdrsize, notesz, coresize; 16038c64af4fSJohn Polstra 1604e7228204SAlfred Perlstein hdr = NULL; 160502d131adSMark Johnston tmpbuf = NULL; 1606bd390213SMikolaj Golub TAILQ_INIT(¬elst); 1607e7228204SAlfred Perlstein 16080ff27d31SJohn Polstra /* Size the program segments. */ 1609905d192dSEdward Tomasz Napierala __elfN(size_segments)(td, &seginfo, flags); 16100ff27d31SJohn Polstra 16110ff27d31SJohn Polstra /* 1612bd390213SMikolaj Golub * Collect info about the core file header area. 16130ff27d31SJohn Polstra */ 1614bd390213SMikolaj Golub hdrsize = sizeof(Elf_Ehdr) + sizeof(Elf_Phdr) * (1 + seginfo.count); 1615c17b0bd2SConrad Meyer if (seginfo.count + 1 >= PN_XNUM) 1616c17b0bd2SConrad Meyer hdrsize += sizeof(Elf_Shdr); 1617435754a5SEdward Tomasz Napierala td->td_proc->p_sysent->sv_elf_core_prepare_notes(td, ¬elst, ¬esz); 1618bd390213SMikolaj Golub coresize = round_page(hdrsize + notesz) + seginfo.size; 16190ff27d31SJohn Polstra 162002d131adSMark Johnston /* Set up core dump parameters. */ 162102d131adSMark Johnston params.offset = 0; 162202d131adSMark Johnston params.active_cred = cred; 162302d131adSMark Johnston params.file_cred = NOCRED; 162402d131adSMark Johnston params.td = td; 162502d131adSMark Johnston params.vp = vp; 162678f57a9cSMark Johnston params.comp = NULL; 162702d131adSMark Johnston 1628afcc55f3SEdward Tomasz Napierala #ifdef RACCT 16294b5c9cf6SEdward Tomasz Napierala if (racct_enable) { 16301ba5ad42SEdward Tomasz Napierala PROC_LOCK(td->td_proc); 1631bd390213SMikolaj Golub error = racct_add(td->td_proc, RACCT_CORE, coresize); 16321ba5ad42SEdward Tomasz Napierala PROC_UNLOCK(td->td_proc); 16331ba5ad42SEdward Tomasz Napierala if (error != 0) { 16341ba5ad42SEdward Tomasz Napierala error = EFAULT; 16351ba5ad42SEdward Tomasz Napierala goto done; 16361ba5ad42SEdward Tomasz Napierala } 16374b5c9cf6SEdward Tomasz Napierala } 1638afcc55f3SEdward Tomasz Napierala #endif 1639bd390213SMikolaj Golub if (coresize >= limit) { 1640fba6b1afSAlfred Perlstein error = EFAULT; 1641fba6b1afSAlfred Perlstein goto done; 1642fba6b1afSAlfred Perlstein } 16430ff27d31SJohn Polstra 1644aa14e9b7SMark Johnston /* Create a compression stream if necessary. */ 164586ffb3d1SKonstantin Belousov compm = compress_user_cores; 164686ffb3d1SKonstantin Belousov if ((flags & (SVC_PT_COREDUMP | SVC_NOCOMPRESS)) == SVC_PT_COREDUMP && 164786ffb3d1SKonstantin Belousov compm == 0) 164886ffb3d1SKonstantin Belousov compm = COMPRESS_GZIP; 164986ffb3d1SKonstantin Belousov if (compm != 0) { 165078f57a9cSMark Johnston params.comp = compressor_init(core_compressed_write, 165186ffb3d1SKonstantin Belousov compm, CORE_BUF_SIZE, 165278f57a9cSMark Johnston compress_user_cores_level, ¶ms); 165378f57a9cSMark Johnston if (params.comp == NULL) { 1654aa14e9b7SMark Johnston error = EFAULT; 1655aa14e9b7SMark Johnston goto done; 1656aa14e9b7SMark Johnston } 1657aa14e9b7SMark Johnston tmpbuf = malloc(CORE_BUF_SIZE, M_TEMP, M_WAITOK | M_ZERO); 1658aa14e9b7SMark Johnston } 1659aa14e9b7SMark Johnston 16600ff27d31SJohn Polstra /* 16610ff27d31SJohn Polstra * Allocate memory for building the header, fill it up, 1662bd390213SMikolaj Golub * and write it out following the notes. 16630ff27d31SJohn Polstra */ 1664a163d034SWarner Losh hdr = malloc(hdrsize, M_TEMP, M_WAITOK); 1665aa14e9b7SMark Johnston error = __elfN(corehdr)(¶ms, seginfo.count, hdr, hdrsize, ¬elst, 16665bc3c617SKonstantin Belousov notesz, flags); 16670ff27d31SJohn Polstra 16680ff27d31SJohn Polstra /* Write the contents of all of the writable segments. */ 16690ff27d31SJohn Polstra if (error == 0) { 16700ff27d31SJohn Polstra Elf_Phdr *php; 16712b471bc6STim J. Robbins off_t offset; 16720ff27d31SJohn Polstra int i; 16730ff27d31SJohn Polstra 16740ff27d31SJohn Polstra php = (Elf_Phdr *)((char *)hdr + sizeof(Elf_Ehdr)) + 1; 1675bd390213SMikolaj Golub offset = round_page(hdrsize + notesz); 16760ff27d31SJohn Polstra for (i = 0; i < seginfo.count; i++) { 1677f31695ccSMark Johnston error = core_output((char *)(uintptr_t)php->p_vaddr, 1678aa14e9b7SMark Johnston php->p_filesz, offset, ¶ms, tmpbuf); 16790ff27d31SJohn Polstra if (error != 0) 16802b471bc6STim J. Robbins break; 16810ff27d31SJohn Polstra offset += php->p_filesz; 16820ff27d31SJohn Polstra php++; 16830ff27d31SJohn Polstra } 168478f57a9cSMark Johnston if (error == 0 && params.comp != NULL) 168578f57a9cSMark Johnston error = compressor_flush(params.comp); 16860ff27d31SJohn Polstra } 1687e7228204SAlfred Perlstein if (error) { 1688e7228204SAlfred Perlstein log(LOG_WARNING, 1689e7228204SAlfred Perlstein "Failed to write core file for process %s (error %d)\n", 1690e7228204SAlfred Perlstein curproc->p_comm, error); 1691e7228204SAlfred Perlstein } 1692e7228204SAlfred Perlstein 1693e7228204SAlfred Perlstein done: 1694aa14e9b7SMark Johnston free(tmpbuf, M_TEMP); 169578f57a9cSMark Johnston if (params.comp != NULL) 169678f57a9cSMark Johnston compressor_fini(params.comp); 1697bd390213SMikolaj Golub while ((ninfo = TAILQ_FIRST(¬elst)) != NULL) { 1698bd390213SMikolaj Golub TAILQ_REMOVE(¬elst, ninfo, link); 1699bd390213SMikolaj Golub free(ninfo, M_TEMP); 1700bd390213SMikolaj Golub } 1701bd390213SMikolaj Golub if (hdr != NULL) 17020ff27d31SJohn Polstra free(hdr, M_TEMP); 17030ff27d31SJohn Polstra 1704a7cddfedSJake Burkholder return (error); 17058c64af4fSJohn Polstra } 17068c64af4fSJohn Polstra 17070ff27d31SJohn Polstra /* 17081005d8afSConrad Meyer * A callback for each_dumpable_segment() to write out the segment's 17090ff27d31SJohn Polstra * program header entry. 17100ff27d31SJohn Polstra */ 17110ff27d31SJohn Polstra static void 17125cc6d253SEd Maste cb_put_phdr(vm_map_entry_t entry, void *closure) 17130ff27d31SJohn Polstra { 17140ff27d31SJohn Polstra struct phdr_closure *phc = (struct phdr_closure *)closure; 17150ff27d31SJohn Polstra Elf_Phdr *phdr = phc->phdr; 17160ff27d31SJohn Polstra 17170ff27d31SJohn Polstra phc->offset = round_page(phc->offset); 17180ff27d31SJohn Polstra 17190ff27d31SJohn Polstra phdr->p_type = PT_LOAD; 17200ff27d31SJohn Polstra phdr->p_offset = phc->offset; 17210ff27d31SJohn Polstra phdr->p_vaddr = entry->start; 17220ff27d31SJohn Polstra phdr->p_paddr = 0; 17230ff27d31SJohn Polstra phdr->p_filesz = phdr->p_memsz = entry->end - entry->start; 17240ff27d31SJohn Polstra phdr->p_align = PAGE_SIZE; 1725ed167eaaSKonstantin Belousov phdr->p_flags = __elfN(untrans_prot)(entry->protection); 17260ff27d31SJohn Polstra 17270ff27d31SJohn Polstra phc->offset += phdr->p_filesz; 17280ff27d31SJohn Polstra phc->phdr++; 17290ff27d31SJohn Polstra } 17300ff27d31SJohn Polstra 17310ff27d31SJohn Polstra /* 17321005d8afSConrad Meyer * A callback for each_dumpable_segment() to gather information about 17330ff27d31SJohn Polstra * the number of segments and their total size. 17340ff27d31SJohn Polstra */ 17350ff27d31SJohn Polstra static void 1736f3325003SConrad Meyer cb_size_segment(vm_map_entry_t entry, void *closure) 17370ff27d31SJohn Polstra { 17380ff27d31SJohn Polstra struct sseg_closure *ssc = (struct sseg_closure *)closure; 17390ff27d31SJohn Polstra 17400ff27d31SJohn Polstra ssc->count++; 17410ff27d31SJohn Polstra ssc->size += entry->end - entry->start; 17420ff27d31SJohn Polstra } 17430ff27d31SJohn Polstra 1744905d192dSEdward Tomasz Napierala void 1745905d192dSEdward Tomasz Napierala __elfN(size_segments)(struct thread *td, struct sseg_closure *seginfo, 1746905d192dSEdward Tomasz Napierala int flags) 1747905d192dSEdward Tomasz Napierala { 1748905d192dSEdward Tomasz Napierala seginfo->count = 0; 1749905d192dSEdward Tomasz Napierala seginfo->size = 0; 1750905d192dSEdward Tomasz Napierala 1751905d192dSEdward Tomasz Napierala each_dumpable_segment(td, cb_size_segment, seginfo, flags); 1752905d192dSEdward Tomasz Napierala } 1753905d192dSEdward Tomasz Napierala 17540ff27d31SJohn Polstra /* 17550ff27d31SJohn Polstra * For each writable segment in the process's memory map, call the given 17560ff27d31SJohn Polstra * function with a pointer to the map entry and some arbitrary 17570ff27d31SJohn Polstra * caller-supplied data. 17580ff27d31SJohn Polstra */ 17590ff27d31SJohn Polstra static void 17605bc3c617SKonstantin Belousov each_dumpable_segment(struct thread *td, segment_callback func, void *closure, 17615bc3c617SKonstantin Belousov int flags) 17620ff27d31SJohn Polstra { 1763247aba24SMarcel Moolenaar struct proc *p = td->td_proc; 17640ff27d31SJohn Polstra vm_map_t map = &p->p_vmspace->vm_map; 17650ff27d31SJohn Polstra vm_map_entry_t entry; 1766976a87a2SAlan Cox vm_object_t backing_object, object; 1767fec41f07SMark Johnston bool ignore_entry; 17680ff27d31SJohn Polstra 1769976a87a2SAlan Cox vm_map_lock_read(map); 17702288078cSDoug Moore VM_MAP_ENTRY_FOREACH(entry, map) { 1771fa7dd9c5SMatthew Dillon /* 1772fa7dd9c5SMatthew Dillon * Don't dump inaccessible mappings, deal with legacy 1773fa7dd9c5SMatthew Dillon * coredump mode. 1774fa7dd9c5SMatthew Dillon * 1775fa7dd9c5SMatthew Dillon * Note that read-only segments related to the elf binary 1776fa7dd9c5SMatthew Dillon * are marked MAP_ENTRY_NOCOREDUMP now so we no longer 1777fa7dd9c5SMatthew Dillon * need to arbitrarily ignore such segments. 1778fa7dd9c5SMatthew Dillon */ 177986ffb3d1SKonstantin Belousov if ((flags & SVC_ALL) == 0) { 1780fa7dd9c5SMatthew Dillon if (elf_legacy_coredump) { 178186ffb3d1SKonstantin Belousov if ((entry->protection & VM_PROT_RW) != 178286ffb3d1SKonstantin Belousov VM_PROT_RW) 17830ff27d31SJohn Polstra continue; 1784fa7dd9c5SMatthew Dillon } else { 1785fa7dd9c5SMatthew Dillon if ((entry->protection & VM_PROT_ALL) == 0) 1786fa7dd9c5SMatthew Dillon continue; 1787fa7dd9c5SMatthew Dillon } 178886ffb3d1SKonstantin Belousov } 17890ff27d31SJohn Polstra 17909730a5daSPaul Saab /* 1791fa7dd9c5SMatthew Dillon * Dont include memory segment in the coredump if 1792fa7dd9c5SMatthew Dillon * MAP_NOCORE is set in mmap(2) or MADV_NOCORE in 1793fa7dd9c5SMatthew Dillon * madvise(2). Do not dump submaps (i.e. parts of the 1794fa7dd9c5SMatthew Dillon * kernel map). 17959730a5daSPaul Saab */ 179686ffb3d1SKonstantin Belousov if ((entry->eflags & MAP_ENTRY_IS_SUB_MAP) != 0) 17979730a5daSPaul Saab continue; 179886ffb3d1SKonstantin Belousov if ((entry->eflags & MAP_ENTRY_NOCOREDUMP) != 0 && 179986ffb3d1SKonstantin Belousov (flags & SVC_ALL) == 0) 180086ffb3d1SKonstantin Belousov continue; 1801976a87a2SAlan Cox if ((object = entry->object.vm_object) == NULL) 18020ff27d31SJohn Polstra continue; 18030ff27d31SJohn Polstra 18040ff27d31SJohn Polstra /* Ignore memory-mapped devices and such things. */ 1805bc403f03SAttilio Rao VM_OBJECT_RLOCK(object); 1806976a87a2SAlan Cox while ((backing_object = object->backing_object) != NULL) { 1807bc403f03SAttilio Rao VM_OBJECT_RLOCK(backing_object); 1808bc403f03SAttilio Rao VM_OBJECT_RUNLOCK(object); 1809976a87a2SAlan Cox object = backing_object; 1810976a87a2SAlan Cox } 1811fec41f07SMark Johnston ignore_entry = (object->flags & OBJ_FICTITIOUS) != 0; 1812bc403f03SAttilio Rao VM_OBJECT_RUNLOCK(object); 1813976a87a2SAlan Cox if (ignore_entry) 18140ff27d31SJohn Polstra continue; 18150ff27d31SJohn Polstra 18160ff27d31SJohn Polstra (*func)(entry, closure); 18170ff27d31SJohn Polstra } 1818976a87a2SAlan Cox vm_map_unlock_read(map); 18190ff27d31SJohn Polstra } 18200ff27d31SJohn Polstra 18210ff27d31SJohn Polstra /* 18220ff27d31SJohn Polstra * Write the core file header to the file, including padding up to 18230ff27d31SJohn Polstra * the page boundary. 18240ff27d31SJohn Polstra */ 18258c64af4fSJohn Polstra static int 1826aa14e9b7SMark Johnston __elfN(corehdr)(struct coredump_params *p, int numsegs, void *hdr, 18275bc3c617SKonstantin Belousov size_t hdrsize, struct note_info_list *notelst, size_t notesz, 18285bc3c617SKonstantin Belousov int flags) 18298c64af4fSJohn Polstra { 1830bd390213SMikolaj Golub struct note_info *ninfo; 1831bd390213SMikolaj Golub struct sbuf *sb; 1832bd390213SMikolaj Golub int error; 18338c64af4fSJohn Polstra 18348c64af4fSJohn Polstra /* Fill in the header. */ 18350ff27d31SJohn Polstra bzero(hdr, hdrsize); 18365bc3c617SKonstantin Belousov __elfN(puthdr)(p->td, hdr, hdrsize, numsegs, notesz, flags); 18378c64af4fSJohn Polstra 1838bd390213SMikolaj Golub sb = sbuf_new(NULL, NULL, CORE_BUF_SIZE, SBUF_FIXEDLEN); 1839aa14e9b7SMark Johnston sbuf_set_drain(sb, sbuf_drain_core_output, p); 1840bd390213SMikolaj Golub sbuf_start_section(sb, NULL); 1841bd390213SMikolaj Golub sbuf_bcat(sb, hdr, hdrsize); 1842bd390213SMikolaj Golub TAILQ_FOREACH(ninfo, notelst, link) 1843435754a5SEdward Tomasz Napierala __elfN(putnote)(p->td, ninfo, sb); 1844bd390213SMikolaj Golub /* Align up to a page boundary for the program segments. */ 1845bd390213SMikolaj Golub sbuf_end_section(sb, -1, PAGE_SIZE, 0); 1846bd390213SMikolaj Golub error = sbuf_finish(sb); 1847bd390213SMikolaj Golub sbuf_delete(sb); 1848bd390213SMikolaj Golub 1849bd390213SMikolaj Golub return (error); 1850e7228204SAlfred Perlstein } 1851bd390213SMikolaj Golub 1852435754a5SEdward Tomasz Napierala void 1853bd390213SMikolaj Golub __elfN(prepare_notes)(struct thread *td, struct note_info_list *list, 1854bd390213SMikolaj Golub size_t *sizep) 1855bd390213SMikolaj Golub { 1856bd390213SMikolaj Golub struct proc *p; 1857bd390213SMikolaj Golub struct thread *thr; 1858bd390213SMikolaj Golub size_t size; 1859bd390213SMikolaj Golub 1860bd390213SMikolaj Golub p = td->td_proc; 1861bd390213SMikolaj Golub size = 0; 1862bd390213SMikolaj Golub 18636b71405bSJohn Baldwin size += __elfN(register_note)(td, list, NT_PRPSINFO, 18646b71405bSJohn Baldwin __elfN(note_prpsinfo), p); 1865bd390213SMikolaj Golub 1866bd390213SMikolaj Golub /* 1867bd390213SMikolaj Golub * To have the debugger select the right thread (LWP) as the initial 1868bd390213SMikolaj Golub * thread, we dump the state of the thread passed to us in td first. 1869bd390213SMikolaj Golub * This is the thread that causes the core dump and thus likely to 1870bd390213SMikolaj Golub * be the right thread one wants to have selected in the debugger. 1871bd390213SMikolaj Golub */ 1872bd390213SMikolaj Golub thr = td; 1873bd390213SMikolaj Golub while (thr != NULL) { 18746b71405bSJohn Baldwin size += __elfN(prepare_register_notes)(td, list, thr); 1875435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, -1, 1876bd390213SMikolaj Golub __elfN(note_threadmd), thr); 1877bd390213SMikolaj Golub 187861b4c627SEdward Tomasz Napierala thr = thr == td ? TAILQ_FIRST(&p->p_threads) : 1879bd390213SMikolaj Golub TAILQ_NEXT(thr, td_plist); 1880bd390213SMikolaj Golub if (thr == td) 1881bd390213SMikolaj Golub thr = TAILQ_NEXT(thr, td_plist); 1882dada0278SJohn Polstra } 1883dada0278SJohn Polstra 1884435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_PROC, 1885f1fca82eSMikolaj Golub __elfN(note_procstat_proc), p); 1886435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_FILES, 1887f1fca82eSMikolaj Golub note_procstat_files, p); 1888435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_VMMAP, 1889f1fca82eSMikolaj Golub note_procstat_vmmap, p); 1890435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_GROUPS, 1891f1fca82eSMikolaj Golub note_procstat_groups, p); 1892435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_UMASK, 1893f1fca82eSMikolaj Golub note_procstat_umask, p); 1894435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_RLIMIT, 1895f1fca82eSMikolaj Golub note_procstat_rlimit, p); 1896435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_OSREL, 1897f1fca82eSMikolaj Golub note_procstat_osrel, p); 1898435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_PSSTRINGS, 1899f1fca82eSMikolaj Golub __elfN(note_procstat_psstrings), p); 1900435754a5SEdward Tomasz Napierala size += __elfN(register_note)(td, list, NT_PROCSTAT_AUXV, 1901f1fca82eSMikolaj Golub __elfN(note_procstat_auxv), p); 1902f1fca82eSMikolaj Golub 1903bd390213SMikolaj Golub *sizep = size; 1904bd390213SMikolaj Golub } 1905bd390213SMikolaj Golub 1906905d192dSEdward Tomasz Napierala void 1907bd390213SMikolaj Golub __elfN(puthdr)(struct thread *td, void *hdr, size_t hdrsize, int numsegs, 19085bc3c617SKonstantin Belousov size_t notesz, int flags) 1909bd390213SMikolaj Golub { 1910bd390213SMikolaj Golub Elf_Ehdr *ehdr; 1911bd390213SMikolaj Golub Elf_Phdr *phdr; 1912c17b0bd2SConrad Meyer Elf_Shdr *shdr; 1913bd390213SMikolaj Golub struct phdr_closure phc; 19145d9f7901SDmitry Chagin Elf_Brandinfo *bi; 1915bd390213SMikolaj Golub 1916bd390213SMikolaj Golub ehdr = (Elf_Ehdr *)hdr; 19175d9f7901SDmitry Chagin bi = td->td_proc->p_elf_brandinfo; 1918bd390213SMikolaj Golub 1919bd390213SMikolaj Golub ehdr->e_ident[EI_MAG0] = ELFMAG0; 1920bd390213SMikolaj Golub ehdr->e_ident[EI_MAG1] = ELFMAG1; 1921bd390213SMikolaj Golub ehdr->e_ident[EI_MAG2] = ELFMAG2; 1922bd390213SMikolaj Golub ehdr->e_ident[EI_MAG3] = ELFMAG3; 1923bd390213SMikolaj Golub ehdr->e_ident[EI_CLASS] = ELF_CLASS; 1924bd390213SMikolaj Golub ehdr->e_ident[EI_DATA] = ELF_DATA; 1925bd390213SMikolaj Golub ehdr->e_ident[EI_VERSION] = EV_CURRENT; 1926435754a5SEdward Tomasz Napierala ehdr->e_ident[EI_OSABI] = td->td_proc->p_sysent->sv_elf_core_osabi; 1927bd390213SMikolaj Golub ehdr->e_ident[EI_ABIVERSION] = 0; 1928bd390213SMikolaj Golub ehdr->e_ident[EI_PAD] = 0; 1929bd390213SMikolaj Golub ehdr->e_type = ET_CORE; 19305d9f7901SDmitry Chagin ehdr->e_machine = bi->machine; 1931bd390213SMikolaj Golub ehdr->e_version = EV_CURRENT; 1932bd390213SMikolaj Golub ehdr->e_entry = 0; 1933bd390213SMikolaj Golub ehdr->e_phoff = sizeof(Elf_Ehdr); 1934885f13dcSJohn Baldwin ehdr->e_flags = td->td_proc->p_elf_flags; 1935bd390213SMikolaj Golub ehdr->e_ehsize = sizeof(Elf_Ehdr); 1936bd390213SMikolaj Golub ehdr->e_phentsize = sizeof(Elf_Phdr); 1937bd390213SMikolaj Golub ehdr->e_shentsize = sizeof(Elf_Shdr); 1938bd390213SMikolaj Golub ehdr->e_shstrndx = SHN_UNDEF; 1939c17b0bd2SConrad Meyer if (numsegs + 1 < PN_XNUM) { 1940c17b0bd2SConrad Meyer ehdr->e_phnum = numsegs + 1; 1941c17b0bd2SConrad Meyer ehdr->e_shnum = 0; 1942c17b0bd2SConrad Meyer } else { 1943c17b0bd2SConrad Meyer ehdr->e_phnum = PN_XNUM; 1944c17b0bd2SConrad Meyer ehdr->e_shnum = 1; 1945c17b0bd2SConrad Meyer 1946c17b0bd2SConrad Meyer ehdr->e_shoff = ehdr->e_phoff + 1947c17b0bd2SConrad Meyer (numsegs + 1) * ehdr->e_phentsize; 1948c17b0bd2SConrad Meyer KASSERT(ehdr->e_shoff == hdrsize - sizeof(Elf_Shdr), 1949c17b0bd2SConrad Meyer ("e_shoff: %zu, hdrsize - shdr: %zu", 195007f825e8SConrad Meyer (size_t)ehdr->e_shoff, hdrsize - sizeof(Elf_Shdr))); 1951c17b0bd2SConrad Meyer 1952c17b0bd2SConrad Meyer shdr = (Elf_Shdr *)((char *)hdr + ehdr->e_shoff); 1953c17b0bd2SConrad Meyer memset(shdr, 0, sizeof(*shdr)); 1954c17b0bd2SConrad Meyer /* 1955c17b0bd2SConrad Meyer * A special first section is used to hold large segment and 1956c17b0bd2SConrad Meyer * section counts. This was proposed by Sun Microsystems in 1957c17b0bd2SConrad Meyer * Solaris and has been adopted by Linux; the standard ELF 1958c17b0bd2SConrad Meyer * tools are already familiar with the technique. 1959c17b0bd2SConrad Meyer * 1960c17b0bd2SConrad Meyer * See table 7-7 of the Solaris "Linker and Libraries Guide" 1961c17b0bd2SConrad Meyer * (or 12-7 depending on the version of the document) for more 1962c17b0bd2SConrad Meyer * details. 1963c17b0bd2SConrad Meyer */ 1964c17b0bd2SConrad Meyer shdr->sh_type = SHT_NULL; 1965c17b0bd2SConrad Meyer shdr->sh_size = ehdr->e_shnum; 1966c17b0bd2SConrad Meyer shdr->sh_link = ehdr->e_shstrndx; 1967c17b0bd2SConrad Meyer shdr->sh_info = numsegs + 1; 1968c17b0bd2SConrad Meyer } 1969bd390213SMikolaj Golub 1970bd390213SMikolaj Golub /* 1971bd390213SMikolaj Golub * Fill in the program header entries. 1972bd390213SMikolaj Golub */ 1973c17b0bd2SConrad Meyer phdr = (Elf_Phdr *)((char *)hdr + ehdr->e_phoff); 1974bd390213SMikolaj Golub 1975bd390213SMikolaj Golub /* The note segement. */ 1976bd390213SMikolaj Golub phdr->p_type = PT_NOTE; 1977bd390213SMikolaj Golub phdr->p_offset = hdrsize; 1978bd390213SMikolaj Golub phdr->p_vaddr = 0; 1979bd390213SMikolaj Golub phdr->p_paddr = 0; 1980bd390213SMikolaj Golub phdr->p_filesz = notesz; 1981bd390213SMikolaj Golub phdr->p_memsz = 0; 1982bd390213SMikolaj Golub phdr->p_flags = PF_R; 19831b8388cdSMikolaj Golub phdr->p_align = ELF_NOTE_ROUNDSIZE; 1984bd390213SMikolaj Golub phdr++; 1985bd390213SMikolaj Golub 1986bd390213SMikolaj Golub /* All the writable segments from the program. */ 1987bd390213SMikolaj Golub phc.phdr = phdr; 1988bd390213SMikolaj Golub phc.offset = round_page(hdrsize + notesz); 19895bc3c617SKonstantin Belousov each_dumpable_segment(td, cb_put_phdr, &phc, flags); 1990bd390213SMikolaj Golub } 1991bd390213SMikolaj Golub 19926b71405bSJohn Baldwin static size_t 19936b71405bSJohn Baldwin __elfN(register_regset_note)(struct thread *td, struct note_info_list *list, 19946b71405bSJohn Baldwin struct regset *regset, struct thread *target_td) 19956b71405bSJohn Baldwin { 19966b71405bSJohn Baldwin const struct sysentvec *sv; 19976b71405bSJohn Baldwin struct note_info *ninfo; 19986b71405bSJohn Baldwin size_t size, notesize; 19996b71405bSJohn Baldwin 20006b71405bSJohn Baldwin size = 0; 20016b71405bSJohn Baldwin if (!regset->get(regset, target_td, NULL, &size) || size == 0) 20026b71405bSJohn Baldwin return (0); 20036b71405bSJohn Baldwin 20046b71405bSJohn Baldwin ninfo = malloc(sizeof(*ninfo), M_TEMP, M_ZERO | M_WAITOK); 20056b71405bSJohn Baldwin ninfo->type = regset->note; 20066b71405bSJohn Baldwin ninfo->regset = regset; 20076b71405bSJohn Baldwin ninfo->outarg = target_td; 20086b71405bSJohn Baldwin ninfo->outsize = size; 20096b71405bSJohn Baldwin TAILQ_INSERT_TAIL(list, ninfo, link); 20106b71405bSJohn Baldwin 20116b71405bSJohn Baldwin sv = td->td_proc->p_sysent; 20126b71405bSJohn Baldwin notesize = sizeof(Elf_Note) + /* note header */ 20136b71405bSJohn Baldwin roundup2(strlen(sv->sv_elf_core_abi_vendor) + 1, ELF_NOTE_ROUNDSIZE) + 20146b71405bSJohn Baldwin /* note name */ 20156b71405bSJohn Baldwin roundup2(size, ELF_NOTE_ROUNDSIZE); /* note description */ 20166b71405bSJohn Baldwin 20176b71405bSJohn Baldwin return (notesize); 20186b71405bSJohn Baldwin } 20196b71405bSJohn Baldwin 2020435754a5SEdward Tomasz Napierala size_t 2021435754a5SEdward Tomasz Napierala __elfN(register_note)(struct thread *td, struct note_info_list *list, 2022435754a5SEdward Tomasz Napierala int type, outfunc_t out, void *arg) 2023bd390213SMikolaj Golub { 2024435754a5SEdward Tomasz Napierala const struct sysentvec *sv; 2025bd390213SMikolaj Golub struct note_info *ninfo; 2026bd390213SMikolaj Golub size_t size, notesize; 2027bd390213SMikolaj Golub 2028435754a5SEdward Tomasz Napierala sv = td->td_proc->p_sysent; 2029bd390213SMikolaj Golub size = 0; 2030bd390213SMikolaj Golub out(arg, NULL, &size); 2031bd390213SMikolaj Golub ninfo = malloc(sizeof(*ninfo), M_TEMP, M_ZERO | M_WAITOK); 2032bd390213SMikolaj Golub ninfo->type = type; 2033bd390213SMikolaj Golub ninfo->outfunc = out; 2034bd390213SMikolaj Golub ninfo->outarg = arg; 2035bd390213SMikolaj Golub ninfo->outsize = size; 2036bd390213SMikolaj Golub TAILQ_INSERT_TAIL(list, ninfo, link); 2037bd390213SMikolaj Golub 2038bd390213SMikolaj Golub if (type == -1) 2039bd390213SMikolaj Golub return (size); 2040bd390213SMikolaj Golub 2041bd390213SMikolaj Golub notesize = sizeof(Elf_Note) + /* note header */ 2042435754a5SEdward Tomasz Napierala roundup2(strlen(sv->sv_elf_core_abi_vendor) + 1, ELF_NOTE_ROUNDSIZE) + 2043180e57e5SJohn Baldwin /* note name */ 2044180e57e5SJohn Baldwin roundup2(size, ELF_NOTE_ROUNDSIZE); /* note description */ 2045180e57e5SJohn Baldwin 2046180e57e5SJohn Baldwin return (notesize); 2047180e57e5SJohn Baldwin } 2048180e57e5SJohn Baldwin 2049180e57e5SJohn Baldwin static size_t 2050180e57e5SJohn Baldwin append_note_data(const void *src, void *dst, size_t len) 2051180e57e5SJohn Baldwin { 2052180e57e5SJohn Baldwin size_t padded_len; 2053180e57e5SJohn Baldwin 2054180e57e5SJohn Baldwin padded_len = roundup2(len, ELF_NOTE_ROUNDSIZE); 2055180e57e5SJohn Baldwin if (dst != NULL) { 2056180e57e5SJohn Baldwin bcopy(src, dst, len); 2057180e57e5SJohn Baldwin bzero((char *)dst + len, padded_len - len); 2058180e57e5SJohn Baldwin } 2059180e57e5SJohn Baldwin return (padded_len); 2060180e57e5SJohn Baldwin } 2061180e57e5SJohn Baldwin 2062180e57e5SJohn Baldwin size_t 2063180e57e5SJohn Baldwin __elfN(populate_note)(int type, void *src, void *dst, size_t size, void **descp) 2064180e57e5SJohn Baldwin { 2065180e57e5SJohn Baldwin Elf_Note *note; 2066180e57e5SJohn Baldwin char *buf; 2067180e57e5SJohn Baldwin size_t notesize; 2068180e57e5SJohn Baldwin 2069180e57e5SJohn Baldwin buf = dst; 2070180e57e5SJohn Baldwin if (buf != NULL) { 2071180e57e5SJohn Baldwin note = (Elf_Note *)buf; 2072180e57e5SJohn Baldwin note->n_namesz = sizeof(FREEBSD_ABI_VENDOR); 2073180e57e5SJohn Baldwin note->n_descsz = size; 2074180e57e5SJohn Baldwin note->n_type = type; 2075180e57e5SJohn Baldwin buf += sizeof(*note); 2076180e57e5SJohn Baldwin buf += append_note_data(FREEBSD_ABI_VENDOR, buf, 2077180e57e5SJohn Baldwin sizeof(FREEBSD_ABI_VENDOR)); 2078180e57e5SJohn Baldwin append_note_data(src, buf, size); 2079180e57e5SJohn Baldwin if (descp != NULL) 2080180e57e5SJohn Baldwin *descp = buf; 2081180e57e5SJohn Baldwin } 2082180e57e5SJohn Baldwin 2083180e57e5SJohn Baldwin notesize = sizeof(Elf_Note) + /* note header */ 2084180e57e5SJohn Baldwin roundup2(sizeof(FREEBSD_ABI_VENDOR), ELF_NOTE_ROUNDSIZE) + 2085180e57e5SJohn Baldwin /* note name */ 20861b8388cdSMikolaj Golub roundup2(size, ELF_NOTE_ROUNDSIZE); /* note description */ 2087bd390213SMikolaj Golub 2088bd390213SMikolaj Golub return (notesize); 2089bd390213SMikolaj Golub } 2090bd390213SMikolaj Golub 2091bd390213SMikolaj Golub static void 2092435754a5SEdward Tomasz Napierala __elfN(putnote)(struct thread *td, struct note_info *ninfo, struct sbuf *sb) 2093bd390213SMikolaj Golub { 2094bd390213SMikolaj Golub Elf_Note note; 2095435754a5SEdward Tomasz Napierala const struct sysentvec *sv; 209614bdbaf2SConrad Meyer ssize_t old_len, sect_len; 209714bdbaf2SConrad Meyer size_t new_len, descsz, i; 2098bd390213SMikolaj Golub 2099bd390213SMikolaj Golub if (ninfo->type == -1) { 2100bd390213SMikolaj Golub ninfo->outfunc(ninfo->outarg, sb, &ninfo->outsize); 2101bd390213SMikolaj Golub return; 2102bd390213SMikolaj Golub } 2103bd390213SMikolaj Golub 2104435754a5SEdward Tomasz Napierala sv = td->td_proc->p_sysent; 2105435754a5SEdward Tomasz Napierala 2106435754a5SEdward Tomasz Napierala note.n_namesz = strlen(sv->sv_elf_core_abi_vendor) + 1; 2107bd390213SMikolaj Golub note.n_descsz = ninfo->outsize; 2108bd390213SMikolaj Golub note.n_type = ninfo->type; 2109bd390213SMikolaj Golub 2110bd390213SMikolaj Golub sbuf_bcat(sb, ¬e, sizeof(note)); 2111bd390213SMikolaj Golub sbuf_start_section(sb, &old_len); 2112435754a5SEdward Tomasz Napierala sbuf_bcat(sb, sv->sv_elf_core_abi_vendor, 2113435754a5SEdward Tomasz Napierala strlen(sv->sv_elf_core_abi_vendor) + 1); 21141b8388cdSMikolaj Golub sbuf_end_section(sb, old_len, ELF_NOTE_ROUNDSIZE, 0); 2115bd390213SMikolaj Golub if (note.n_descsz == 0) 2116bd390213SMikolaj Golub return; 2117bd390213SMikolaj Golub sbuf_start_section(sb, &old_len); 21186b71405bSJohn Baldwin if (ninfo->regset != NULL) { 21196b71405bSJohn Baldwin struct regset *regset = ninfo->regset; 21206b71405bSJohn Baldwin void *buf; 21216b71405bSJohn Baldwin 21226b71405bSJohn Baldwin buf = malloc(ninfo->outsize, M_TEMP, M_ZERO | M_WAITOK); 21236b71405bSJohn Baldwin (void)regset->get(regset, ninfo->outarg, buf, &ninfo->outsize); 21246b71405bSJohn Baldwin sbuf_bcat(sb, buf, ninfo->outsize); 21256b71405bSJohn Baldwin free(buf, M_TEMP); 21266b71405bSJohn Baldwin } else 2127bd390213SMikolaj Golub ninfo->outfunc(ninfo->outarg, sb, &ninfo->outsize); 212814bdbaf2SConrad Meyer sect_len = sbuf_end_section(sb, old_len, ELF_NOTE_ROUNDSIZE, 0); 212914bdbaf2SConrad Meyer if (sect_len < 0) 213014bdbaf2SConrad Meyer return; 213114bdbaf2SConrad Meyer 213214bdbaf2SConrad Meyer new_len = (size_t)sect_len; 213314bdbaf2SConrad Meyer descsz = roundup(note.n_descsz, ELF_NOTE_ROUNDSIZE); 213414bdbaf2SConrad Meyer if (new_len < descsz) { 213514bdbaf2SConrad Meyer /* 213614bdbaf2SConrad Meyer * It is expected that individual note emitters will correctly 213714bdbaf2SConrad Meyer * predict their expected output size and fill up to that size 213814bdbaf2SConrad Meyer * themselves, padding in a format-specific way if needed. 213914bdbaf2SConrad Meyer * However, in case they don't, just do it here with zeros. 214014bdbaf2SConrad Meyer */ 214114bdbaf2SConrad Meyer for (i = 0; i < descsz - new_len; i++) 214214bdbaf2SConrad Meyer sbuf_putc(sb, 0); 214314bdbaf2SConrad Meyer } else if (new_len > descsz) { 214414bdbaf2SConrad Meyer /* 214514bdbaf2SConrad Meyer * We can't always truncate sb -- we may have drained some 214614bdbaf2SConrad Meyer * of it already. 214714bdbaf2SConrad Meyer */ 214814bdbaf2SConrad Meyer KASSERT(new_len == descsz, ("%s: Note type %u changed as we " 214914bdbaf2SConrad Meyer "read it (%zu > %zu). Since it is longer than " 215014bdbaf2SConrad Meyer "expected, this coredump's notes are corrupt. THIS " 215114bdbaf2SConrad Meyer "IS A BUG in the note_procstat routine for type %u.\n", 215214bdbaf2SConrad Meyer __func__, (unsigned)note.n_type, new_len, descsz, 215314bdbaf2SConrad Meyer (unsigned)note.n_type)); 215414bdbaf2SConrad Meyer } 2155bd390213SMikolaj Golub } 2156bd390213SMikolaj Golub 2157bd390213SMikolaj Golub /* 2158bd390213SMikolaj Golub * Miscellaneous note out functions. 2159bd390213SMikolaj Golub */ 2160bd390213SMikolaj Golub 2161841c0c7eSNathan Whitehorn #if defined(COMPAT_FREEBSD32) && __ELF_WORD_SIZE == 32 2162841c0c7eSNathan Whitehorn #include <compat/freebsd32/freebsd32.h> 216351645e83SJohn Baldwin #include <compat/freebsd32/freebsd32_signal.h> 2164841c0c7eSNathan Whitehorn 216562919d78SPeter Wemm typedef struct prstatus32 elf_prstatus_t; 216662919d78SPeter Wemm typedef struct prpsinfo32 elf_prpsinfo_t; 216762919d78SPeter Wemm typedef struct fpreg32 elf_prfpregset_t; 216862919d78SPeter Wemm typedef struct fpreg32 elf_fpregset_t; 216962919d78SPeter Wemm typedef struct reg32 elf_gregset_t; 21707f08176eSAttilio Rao typedef struct thrmisc32 elf_thrmisc_t; 21710288d427SJohn Baldwin typedef struct ptrace_lwpinfo32 elf_lwpinfo_t; 2172f1fca82eSMikolaj Golub #define ELF_KERN_PROC_MASK KERN_PROC_MASK32 2173f1fca82eSMikolaj Golub typedef struct kinfo_proc32 elf_kinfo_proc_t; 2174f1fca82eSMikolaj Golub typedef uint32_t elf_ps_strings_t; 217562919d78SPeter Wemm #else 217662919d78SPeter Wemm typedef prstatus_t elf_prstatus_t; 217762919d78SPeter Wemm typedef prpsinfo_t elf_prpsinfo_t; 217862919d78SPeter Wemm typedef prfpregset_t elf_prfpregset_t; 217962919d78SPeter Wemm typedef prfpregset_t elf_fpregset_t; 218062919d78SPeter Wemm typedef gregset_t elf_gregset_t; 21817f08176eSAttilio Rao typedef thrmisc_t elf_thrmisc_t; 21820288d427SJohn Baldwin typedef struct ptrace_lwpinfo elf_lwpinfo_t; 2183f1fca82eSMikolaj Golub #define ELF_KERN_PROC_MASK 0 2184f1fca82eSMikolaj Golub typedef struct kinfo_proc elf_kinfo_proc_t; 2185f1fca82eSMikolaj Golub typedef vm_offset_t elf_ps_strings_t; 218662919d78SPeter Wemm #endif 218762919d78SPeter Wemm 21888c64af4fSJohn Polstra static void 2189bd390213SMikolaj Golub __elfN(note_prpsinfo)(void *arg, struct sbuf *sb, size_t *sizep) 21908c64af4fSJohn Polstra { 2191c77547d2SJohn Baldwin struct sbuf sbarg; 2192c77547d2SJohn Baldwin size_t len; 2193c77547d2SJohn Baldwin char *cp, *end; 2194247aba24SMarcel Moolenaar struct proc *p; 2195bd390213SMikolaj Golub elf_prpsinfo_t *psinfo; 2196c77547d2SJohn Baldwin int error; 21978c64af4fSJohn Polstra 219861b4c627SEdward Tomasz Napierala p = arg; 2199bd390213SMikolaj Golub if (sb != NULL) { 2200bd390213SMikolaj Golub KASSERT(*sizep == sizeof(*psinfo), ("invalid size")); 2201bd390213SMikolaj Golub psinfo = malloc(sizeof(*psinfo), M_TEMP, M_ZERO | M_WAITOK); 22028c9b7b2cSMarcel Moolenaar psinfo->pr_version = PRPSINFO_VERSION; 220362919d78SPeter Wemm psinfo->pr_psinfosz = sizeof(elf_prpsinfo_t); 2204ccd3953eSJohn Baldwin strlcpy(psinfo->pr_fname, p->p_comm, sizeof(psinfo->pr_fname)); 2205c77547d2SJohn Baldwin PROC_LOCK(p); 2206c77547d2SJohn Baldwin if (p->p_args != NULL) { 2207c77547d2SJohn Baldwin len = sizeof(psinfo->pr_psargs) - 1; 2208c77547d2SJohn Baldwin if (len > p->p_args->ar_length) 2209c77547d2SJohn Baldwin len = p->p_args->ar_length; 2210c77547d2SJohn Baldwin memcpy(psinfo->pr_psargs, p->p_args->ar_args, len); 2211c77547d2SJohn Baldwin PROC_UNLOCK(p); 2212c77547d2SJohn Baldwin error = 0; 2213c77547d2SJohn Baldwin } else { 2214c77547d2SJohn Baldwin _PHOLD(p); 2215c77547d2SJohn Baldwin PROC_UNLOCK(p); 2216c77547d2SJohn Baldwin sbuf_new(&sbarg, psinfo->pr_psargs, 2217c77547d2SJohn Baldwin sizeof(psinfo->pr_psargs), SBUF_FIXEDLEN); 2218c77547d2SJohn Baldwin error = proc_getargv(curthread, p, &sbarg); 2219c77547d2SJohn Baldwin PRELE(p); 222000d17cf3SKonstantin Belousov if (sbuf_finish(&sbarg) == 0) { 222100d17cf3SKonstantin Belousov len = sbuf_len(&sbarg); 222200d17cf3SKonstantin Belousov if (len > 0) 222300d17cf3SKonstantin Belousov len--; 222400d17cf3SKonstantin Belousov } else { 2225c77547d2SJohn Baldwin len = sizeof(psinfo->pr_psargs) - 1; 222600d17cf3SKonstantin Belousov } 2227c77547d2SJohn Baldwin sbuf_delete(&sbarg); 2228c77547d2SJohn Baldwin } 222900d17cf3SKonstantin Belousov if (error != 0 || len == 0 || (ssize_t)len == -1) 2230ccd3953eSJohn Baldwin strlcpy(psinfo->pr_psargs, p->p_comm, 22318c9b7b2cSMarcel Moolenaar sizeof(psinfo->pr_psargs)); 2232c77547d2SJohn Baldwin else { 2233c77547d2SJohn Baldwin KASSERT(len < sizeof(psinfo->pr_psargs), 2234c77547d2SJohn Baldwin ("len is too long: %zu vs %zu", len, 2235c77547d2SJohn Baldwin sizeof(psinfo->pr_psargs))); 2236c77547d2SJohn Baldwin cp = psinfo->pr_psargs; 2237c77547d2SJohn Baldwin end = cp + len - 1; 2238c77547d2SJohn Baldwin for (;;) { 2239c77547d2SJohn Baldwin cp = memchr(cp, '\0', end - cp); 2240c77547d2SJohn Baldwin if (cp == NULL) 2241c77547d2SJohn Baldwin break; 2242c77547d2SJohn Baldwin *cp = ' '; 2243c77547d2SJohn Baldwin } 2244c77547d2SJohn Baldwin } 2245ccb83afdSJohn Baldwin psinfo->pr_pid = p->p_pid; 2246bd390213SMikolaj Golub sbuf_bcat(sb, psinfo, sizeof(*psinfo)); 2247bd390213SMikolaj Golub free(psinfo, M_TEMP); 2248bd390213SMikolaj Golub } 2249bd390213SMikolaj Golub *sizep = sizeof(*psinfo); 2250bd390213SMikolaj Golub } 2251bd390213SMikolaj Golub 2252548a2ec4SAndrew Turner static bool 2253548a2ec4SAndrew Turner __elfN(get_prstatus)(struct regset *rs, struct thread *td, void *buf, 2254548a2ec4SAndrew Turner size_t *sizep) 2255bd390213SMikolaj Golub { 2256bd390213SMikolaj Golub elf_prstatus_t *status; 2257bd390213SMikolaj Golub 2258548a2ec4SAndrew Turner if (buf != NULL) { 2259548a2ec4SAndrew Turner KASSERT(*sizep == sizeof(*status), ("%s: invalid size", 2260548a2ec4SAndrew Turner __func__)); 2261548a2ec4SAndrew Turner status = buf; 22621babcad6SMark Johnston memset(status, 0, *sizep); 22638c9b7b2cSMarcel Moolenaar status->pr_version = PRSTATUS_VERSION; 226462919d78SPeter Wemm status->pr_statussz = sizeof(elf_prstatus_t); 226562919d78SPeter Wemm status->pr_gregsetsz = sizeof(elf_gregset_t); 226662919d78SPeter Wemm status->pr_fpregsetsz = sizeof(elf_fpregset_t); 22678c9b7b2cSMarcel Moolenaar status->pr_osreldate = osreldate; 2268bd390213SMikolaj Golub status->pr_cursig = td->td_proc->p_sig; 2269bd390213SMikolaj Golub status->pr_pid = td->td_tid; 2270841c0c7eSNathan Whitehorn #if defined(COMPAT_FREEBSD32) && __ELF_WORD_SIZE == 32 2271bd390213SMikolaj Golub fill_regs32(td, &status->pr_reg); 227262919d78SPeter Wemm #else 2273bd390213SMikolaj Golub fill_regs(td, &status->pr_reg); 227462919d78SPeter Wemm #endif 2275548a2ec4SAndrew Turner } 2276548a2ec4SAndrew Turner *sizep = sizeof(*status); 2277548a2ec4SAndrew Turner return (true); 2278548a2ec4SAndrew Turner } 2279548a2ec4SAndrew Turner 2280548a2ec4SAndrew Turner static bool 2281548a2ec4SAndrew Turner __elfN(set_prstatus)(struct regset *rs, struct thread *td, void *buf, 2282548a2ec4SAndrew Turner size_t size) 2283548a2ec4SAndrew Turner { 2284548a2ec4SAndrew Turner elf_prstatus_t *status; 2285548a2ec4SAndrew Turner 2286548a2ec4SAndrew Turner KASSERT(size == sizeof(*status), ("%s: invalid size", __func__)); 2287548a2ec4SAndrew Turner status = buf; 2288548a2ec4SAndrew Turner #if defined(COMPAT_FREEBSD32) && __ELF_WORD_SIZE == 32 2289548a2ec4SAndrew Turner set_regs32(td, &status->pr_reg); 2290548a2ec4SAndrew Turner #else 2291548a2ec4SAndrew Turner set_regs(td, &status->pr_reg); 2292548a2ec4SAndrew Turner #endif 2293548a2ec4SAndrew Turner return (true); 2294548a2ec4SAndrew Turner } 2295548a2ec4SAndrew Turner 2296548a2ec4SAndrew Turner static struct regset __elfN(regset_prstatus) = { 2297548a2ec4SAndrew Turner .note = NT_PRSTATUS, 2298548a2ec4SAndrew Turner .size = sizeof(elf_prstatus_t), 2299548a2ec4SAndrew Turner .get = __elfN(get_prstatus), 2300548a2ec4SAndrew Turner .set = __elfN(set_prstatus), 2301548a2ec4SAndrew Turner }; 2302548a2ec4SAndrew Turner ELF_REGSET(__elfN(regset_prstatus)); 2303548a2ec4SAndrew Turner 2304548a2ec4SAndrew Turner static bool 2305548a2ec4SAndrew Turner __elfN(get_fpregset)(struct regset *rs, struct thread *td, void *buf, 2306548a2ec4SAndrew Turner size_t *sizep) 2307548a2ec4SAndrew Turner { 2308548a2ec4SAndrew Turner elf_prfpregset_t *fpregset; 2309548a2ec4SAndrew Turner 2310548a2ec4SAndrew Turner if (buf != NULL) { 2311548a2ec4SAndrew Turner KASSERT(*sizep == sizeof(*fpregset), ("%s: invalid size", 2312548a2ec4SAndrew Turner __func__)); 2313548a2ec4SAndrew Turner fpregset = buf; 2314548a2ec4SAndrew Turner #if defined(COMPAT_FREEBSD32) && __ELF_WORD_SIZE == 32 2315548a2ec4SAndrew Turner fill_fpregs32(td, fpregset); 2316548a2ec4SAndrew Turner #else 2317548a2ec4SAndrew Turner fill_fpregs(td, fpregset); 2318548a2ec4SAndrew Turner #endif 2319548a2ec4SAndrew Turner } 23200b25cbc7SJohn Baldwin *sizep = sizeof(*fpregset); 2321548a2ec4SAndrew Turner return (true); 2322548a2ec4SAndrew Turner } 2323548a2ec4SAndrew Turner 2324548a2ec4SAndrew Turner static bool 2325548a2ec4SAndrew Turner __elfN(set_fpregset)(struct regset *rs, struct thread *td, void *buf, 2326548a2ec4SAndrew Turner size_t size) 2327548a2ec4SAndrew Turner { 2328548a2ec4SAndrew Turner elf_prfpregset_t *fpregset; 2329548a2ec4SAndrew Turner 2330548a2ec4SAndrew Turner fpregset = buf; 2331548a2ec4SAndrew Turner KASSERT(size == sizeof(*fpregset), ("%s: invalid size", __func__)); 2332548a2ec4SAndrew Turner #if defined(COMPAT_FREEBSD32) && __ELF_WORD_SIZE == 32 2333548a2ec4SAndrew Turner set_fpregs32(td, fpregset); 2334548a2ec4SAndrew Turner #else 2335548a2ec4SAndrew Turner set_fpregs(td, fpregset); 2336548a2ec4SAndrew Turner #endif 2337548a2ec4SAndrew Turner return (true); 2338548a2ec4SAndrew Turner } 2339548a2ec4SAndrew Turner 2340548a2ec4SAndrew Turner static struct regset __elfN(regset_fpregset) = { 2341548a2ec4SAndrew Turner .note = NT_FPREGSET, 2342548a2ec4SAndrew Turner .size = sizeof(elf_prfpregset_t), 2343548a2ec4SAndrew Turner .get = __elfN(get_fpregset), 2344548a2ec4SAndrew Turner .set = __elfN(set_fpregset), 2345548a2ec4SAndrew Turner }; 2346548a2ec4SAndrew Turner ELF_REGSET(__elfN(regset_fpregset)); 2347548a2ec4SAndrew Turner 23480288d427SJohn Baldwin static bool 23490288d427SJohn Baldwin __elfN(get_thrmisc)(struct regset *rs, struct thread *td, void *buf, 23500288d427SJohn Baldwin size_t *sizep) 23510288d427SJohn Baldwin { 23520288d427SJohn Baldwin elf_thrmisc_t *thrmisc; 23530288d427SJohn Baldwin 23540288d427SJohn Baldwin if (buf != NULL) { 23550288d427SJohn Baldwin KASSERT(*sizep == sizeof(*thrmisc), 23560288d427SJohn Baldwin ("%s: invalid size", __func__)); 23570288d427SJohn Baldwin thrmisc = buf; 23580288d427SJohn Baldwin bzero(thrmisc, sizeof(*thrmisc)); 23590288d427SJohn Baldwin strcpy(thrmisc->pr_tname, td->td_name); 23600288d427SJohn Baldwin } 23610288d427SJohn Baldwin *sizep = sizeof(*thrmisc); 23620288d427SJohn Baldwin return (true); 23630288d427SJohn Baldwin } 23640288d427SJohn Baldwin 23650288d427SJohn Baldwin static struct regset __elfN(regset_thrmisc) = { 23660288d427SJohn Baldwin .note = NT_THRMISC, 23670288d427SJohn Baldwin .size = sizeof(elf_thrmisc_t), 23680288d427SJohn Baldwin .get = __elfN(get_thrmisc), 23690288d427SJohn Baldwin }; 23700288d427SJohn Baldwin ELF_REGSET(__elfN(regset_thrmisc)); 23710288d427SJohn Baldwin 23720288d427SJohn Baldwin static bool 23730288d427SJohn Baldwin __elfN(get_lwpinfo)(struct regset *rs, struct thread *td, void *buf, 23740288d427SJohn Baldwin size_t *sizep) 23750288d427SJohn Baldwin { 23760288d427SJohn Baldwin elf_lwpinfo_t pl; 23770288d427SJohn Baldwin size_t size; 23780288d427SJohn Baldwin int structsize; 23790288d427SJohn Baldwin 23800288d427SJohn Baldwin size = sizeof(structsize) + sizeof(pl); 23810288d427SJohn Baldwin if (buf != NULL) { 23820288d427SJohn Baldwin KASSERT(*sizep == size, ("%s: invalid size", __func__)); 23830288d427SJohn Baldwin structsize = sizeof(pl); 23840288d427SJohn Baldwin memcpy(buf, &structsize, sizeof(structsize)); 23850288d427SJohn Baldwin bzero(&pl, sizeof(pl)); 23860288d427SJohn Baldwin pl.pl_lwpid = td->td_tid; 23870288d427SJohn Baldwin pl.pl_event = PL_EVENT_NONE; 23880288d427SJohn Baldwin pl.pl_sigmask = td->td_sigmask; 23890288d427SJohn Baldwin pl.pl_siglist = td->td_siglist; 23900288d427SJohn Baldwin if (td->td_si.si_signo != 0) { 23910288d427SJohn Baldwin pl.pl_event = PL_EVENT_SIGNAL; 23920288d427SJohn Baldwin pl.pl_flags |= PL_FLAG_SI; 23930288d427SJohn Baldwin #if defined(COMPAT_FREEBSD32) && __ELF_WORD_SIZE == 32 23940288d427SJohn Baldwin siginfo_to_siginfo32(&td->td_si, &pl.pl_siginfo); 23950288d427SJohn Baldwin #else 23960288d427SJohn Baldwin pl.pl_siginfo = td->td_si; 23970288d427SJohn Baldwin #endif 23980288d427SJohn Baldwin } 23990288d427SJohn Baldwin strcpy(pl.pl_tdname, td->td_name); 24000288d427SJohn Baldwin /* XXX TODO: supply more information in struct ptrace_lwpinfo*/ 24010288d427SJohn Baldwin memcpy((int *)buf + 1, &pl, sizeof(pl)); 24020288d427SJohn Baldwin } 24030288d427SJohn Baldwin *sizep = size; 24040288d427SJohn Baldwin return (true); 24050288d427SJohn Baldwin } 24060288d427SJohn Baldwin 24070288d427SJohn Baldwin static struct regset __elfN(regset_lwpinfo) = { 24080288d427SJohn Baldwin .note = NT_PTLWPINFO, 24090288d427SJohn Baldwin .size = sizeof(int) + sizeof(elf_lwpinfo_t), 24100288d427SJohn Baldwin .get = __elfN(get_lwpinfo), 24110288d427SJohn Baldwin }; 24120288d427SJohn Baldwin ELF_REGSET(__elfN(regset_lwpinfo)); 24130288d427SJohn Baldwin 24146b71405bSJohn Baldwin static size_t 24156b71405bSJohn Baldwin __elfN(prepare_register_notes)(struct thread *td, struct note_info_list *list, 24166b71405bSJohn Baldwin struct thread *target_td) 2417bd390213SMikolaj Golub { 24186b71405bSJohn Baldwin struct sysentvec *sv = td->td_proc->p_sysent; 24196b71405bSJohn Baldwin struct regset **regsetp, **regset_end, *regset; 24206b71405bSJohn Baldwin size_t size; 2421bd390213SMikolaj Golub 24226b71405bSJohn Baldwin size = 0; 24236b71405bSJohn Baldwin 24246b71405bSJohn Baldwin /* NT_PRSTATUS must be the first register set note. */ 24256b71405bSJohn Baldwin size += __elfN(register_regset_note)(td, list, &__elfN(regset_prstatus), 24266b71405bSJohn Baldwin target_td); 24276b71405bSJohn Baldwin 24286b71405bSJohn Baldwin regsetp = sv->sv_regset_begin; 24296b71405bSJohn Baldwin if (regsetp == NULL) { 24306b71405bSJohn Baldwin /* XXX: This shouldn't be true for any FreeBSD ABIs. */ 24316b71405bSJohn Baldwin size += __elfN(register_regset_note)(td, list, 24326b71405bSJohn Baldwin &__elfN(regset_fpregset), target_td); 24336b71405bSJohn Baldwin return (size); 2434bd390213SMikolaj Golub } 24356b71405bSJohn Baldwin regset_end = sv->sv_regset_end; 24366b71405bSJohn Baldwin MPASS(regset_end != NULL); 24376b71405bSJohn Baldwin for (; regsetp < regset_end; regsetp++) { 24386b71405bSJohn Baldwin regset = *regsetp; 24396b71405bSJohn Baldwin if (regset->note == NT_PRSTATUS) 24406b71405bSJohn Baldwin continue; 24416b71405bSJohn Baldwin size += __elfN(register_regset_note)(td, list, regset, 24426b71405bSJohn Baldwin target_td); 24436b71405bSJohn Baldwin } 24446b71405bSJohn Baldwin return (size); 2445bd390213SMikolaj Golub } 2446bd390213SMikolaj Golub 24474da47b2fSMarcel Moolenaar /* 24484da47b2fSMarcel Moolenaar * Allow for MD specific notes, as well as any MD 24494da47b2fSMarcel Moolenaar * specific preparations for writing MI notes. 24504da47b2fSMarcel Moolenaar */ 24518c64af4fSJohn Polstra static void 2452bd390213SMikolaj Golub __elfN(note_threadmd)(void *arg, struct sbuf *sb, size_t *sizep) 24538c64af4fSJohn Polstra { 2454bd390213SMikolaj Golub struct thread *td; 2455bd390213SMikolaj Golub void *buf; 2456bd390213SMikolaj Golub size_t size; 24578c64af4fSJohn Polstra 2458bd390213SMikolaj Golub td = (struct thread *)arg; 2459bd390213SMikolaj Golub size = *sizep; 2460bd390213SMikolaj Golub if (size != 0 && sb != NULL) 2461bd390213SMikolaj Golub buf = malloc(size, M_TEMP, M_ZERO | M_WAITOK); 246283a396ceSChristian Brueffer else 246383a396ceSChristian Brueffer buf = NULL; 2464bd390213SMikolaj Golub size = 0; 2465bd390213SMikolaj Golub __elfN(dump_thread)(td, buf, &size); 246664779280SKonstantin Belousov KASSERT(sb == NULL || *sizep == size, ("invalid size")); 246783a396ceSChristian Brueffer if (size != 0 && sb != NULL) 2468bd390213SMikolaj Golub sbuf_bcat(sb, buf, size); 2469a1761d73SChristian Brueffer free(buf, M_TEMP); 2470bd390213SMikolaj Golub *sizep = size; 24718c64af4fSJohn Polstra } 24728c64af4fSJohn Polstra 2473f1fca82eSMikolaj Golub #ifdef KINFO_PROC_SIZE 2474f1fca82eSMikolaj Golub CTASSERT(sizeof(struct kinfo_proc) == KINFO_PROC_SIZE); 2475f1fca82eSMikolaj Golub #endif 2476f1fca82eSMikolaj Golub 2477f1fca82eSMikolaj Golub static void 2478f1fca82eSMikolaj Golub __elfN(note_procstat_proc)(void *arg, struct sbuf *sb, size_t *sizep) 2479f1fca82eSMikolaj Golub { 2480f1fca82eSMikolaj Golub struct proc *p; 2481f1fca82eSMikolaj Golub size_t size; 2482f1fca82eSMikolaj Golub int structsize; 2483f1fca82eSMikolaj Golub 248461b4c627SEdward Tomasz Napierala p = arg; 2485f1fca82eSMikolaj Golub size = sizeof(structsize) + p->p_numthreads * 2486f1fca82eSMikolaj Golub sizeof(elf_kinfo_proc_t); 2487f1fca82eSMikolaj Golub 2488f1fca82eSMikolaj Golub if (sb != NULL) { 2489f1fca82eSMikolaj Golub KASSERT(*sizep == size, ("invalid size")); 2490f1fca82eSMikolaj Golub structsize = sizeof(elf_kinfo_proc_t); 2491f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 24924daea938SKonstantin Belousov sx_slock(&proctree_lock); 2493f1fca82eSMikolaj Golub PROC_LOCK(p); 2494f1fca82eSMikolaj Golub kern_proc_out(p, sb, ELF_KERN_PROC_MASK); 24954daea938SKonstantin Belousov sx_sunlock(&proctree_lock); 2496f1fca82eSMikolaj Golub } 2497f1fca82eSMikolaj Golub *sizep = size; 2498f1fca82eSMikolaj Golub } 2499f1fca82eSMikolaj Golub 2500f1fca82eSMikolaj Golub #ifdef KINFO_FILE_SIZE 2501f1fca82eSMikolaj Golub CTASSERT(sizeof(struct kinfo_file) == KINFO_FILE_SIZE); 2502f1fca82eSMikolaj Golub #endif 2503f1fca82eSMikolaj Golub 2504f1fca82eSMikolaj Golub static void 2505f1fca82eSMikolaj Golub note_procstat_files(void *arg, struct sbuf *sb, size_t *sizep) 2506f1fca82eSMikolaj Golub { 2507f1fca82eSMikolaj Golub struct proc *p; 250814bdbaf2SConrad Meyer size_t size, sect_sz, i; 250914bdbaf2SConrad Meyer ssize_t start_len, sect_len; 251014bdbaf2SConrad Meyer int structsize, filedesc_flags; 251114bdbaf2SConrad Meyer 2512bcb60d52SConrad Meyer if (coredump_pack_fileinfo) 251314bdbaf2SConrad Meyer filedesc_flags = KERN_FILEDESC_PACK_KINFO; 251414bdbaf2SConrad Meyer else 251514bdbaf2SConrad Meyer filedesc_flags = 0; 2516f1fca82eSMikolaj Golub 251761b4c627SEdward Tomasz Napierala p = arg; 251814bdbaf2SConrad Meyer structsize = sizeof(struct kinfo_file); 2519f1fca82eSMikolaj Golub if (sb == NULL) { 2520f1fca82eSMikolaj Golub size = 0; 2521f1fca82eSMikolaj Golub sb = sbuf_new(NULL, NULL, 128, SBUF_FIXEDLEN); 25225c32e9fcSAlexander Motin sbuf_set_drain(sb, sbuf_count_drain, &size); 2523f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2524f1fca82eSMikolaj Golub PROC_LOCK(p); 252514bdbaf2SConrad Meyer kern_proc_filedesc_out(p, sb, -1, filedesc_flags); 2526f1fca82eSMikolaj Golub sbuf_finish(sb); 2527f1fca82eSMikolaj Golub sbuf_delete(sb); 2528f1fca82eSMikolaj Golub *sizep = size; 2529f1fca82eSMikolaj Golub } else { 253014bdbaf2SConrad Meyer sbuf_start_section(sb, &start_len); 253114bdbaf2SConrad Meyer 2532f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2533f1fca82eSMikolaj Golub PROC_LOCK(p); 253414bdbaf2SConrad Meyer kern_proc_filedesc_out(p, sb, *sizep - sizeof(structsize), 253514bdbaf2SConrad Meyer filedesc_flags); 253614bdbaf2SConrad Meyer 253714bdbaf2SConrad Meyer sect_len = sbuf_end_section(sb, start_len, 0, 0); 253814bdbaf2SConrad Meyer if (sect_len < 0) 253914bdbaf2SConrad Meyer return; 254014bdbaf2SConrad Meyer sect_sz = sect_len; 254114bdbaf2SConrad Meyer 254214bdbaf2SConrad Meyer KASSERT(sect_sz <= *sizep, 254314bdbaf2SConrad Meyer ("kern_proc_filedesc_out did not respect maxlen; " 254414bdbaf2SConrad Meyer "requested %zu, got %zu", *sizep - sizeof(structsize), 254514bdbaf2SConrad Meyer sect_sz - sizeof(structsize))); 254614bdbaf2SConrad Meyer 254714bdbaf2SConrad Meyer for (i = 0; i < *sizep - sect_sz && sb->s_error == 0; i++) 254814bdbaf2SConrad Meyer sbuf_putc(sb, 0); 2549f1fca82eSMikolaj Golub } 2550f1fca82eSMikolaj Golub } 2551f1fca82eSMikolaj Golub 2552f1fca82eSMikolaj Golub #ifdef KINFO_VMENTRY_SIZE 2553f1fca82eSMikolaj Golub CTASSERT(sizeof(struct kinfo_vmentry) == KINFO_VMENTRY_SIZE); 2554f1fca82eSMikolaj Golub #endif 2555f1fca82eSMikolaj Golub 2556f1fca82eSMikolaj Golub static void 2557f1fca82eSMikolaj Golub note_procstat_vmmap(void *arg, struct sbuf *sb, size_t *sizep) 2558f1fca82eSMikolaj Golub { 2559f1fca82eSMikolaj Golub struct proc *p; 2560f1fca82eSMikolaj Golub size_t size; 2561e6b95927SConrad Meyer int structsize, vmmap_flags; 2562e6b95927SConrad Meyer 2563e6b95927SConrad Meyer if (coredump_pack_vmmapinfo) 2564e6b95927SConrad Meyer vmmap_flags = KERN_VMMAP_PACK_KINFO; 2565e6b95927SConrad Meyer else 2566e6b95927SConrad Meyer vmmap_flags = 0; 2567f1fca82eSMikolaj Golub 256861b4c627SEdward Tomasz Napierala p = arg; 2569e6b95927SConrad Meyer structsize = sizeof(struct kinfo_vmentry); 2570f1fca82eSMikolaj Golub if (sb == NULL) { 2571f1fca82eSMikolaj Golub size = 0; 2572f1fca82eSMikolaj Golub sb = sbuf_new(NULL, NULL, 128, SBUF_FIXEDLEN); 25735c32e9fcSAlexander Motin sbuf_set_drain(sb, sbuf_count_drain, &size); 2574f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2575f1fca82eSMikolaj Golub PROC_LOCK(p); 2576e6b95927SConrad Meyer kern_proc_vmmap_out(p, sb, -1, vmmap_flags); 2577f1fca82eSMikolaj Golub sbuf_finish(sb); 2578f1fca82eSMikolaj Golub sbuf_delete(sb); 2579f1fca82eSMikolaj Golub *sizep = size; 2580f1fca82eSMikolaj Golub } else { 2581f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2582f1fca82eSMikolaj Golub PROC_LOCK(p); 2583e6b95927SConrad Meyer kern_proc_vmmap_out(p, sb, *sizep - sizeof(structsize), 2584e6b95927SConrad Meyer vmmap_flags); 2585f1fca82eSMikolaj Golub } 2586f1fca82eSMikolaj Golub } 2587f1fca82eSMikolaj Golub 2588f1fca82eSMikolaj Golub static void 2589f1fca82eSMikolaj Golub note_procstat_groups(void *arg, struct sbuf *sb, size_t *sizep) 2590f1fca82eSMikolaj Golub { 2591f1fca82eSMikolaj Golub struct proc *p; 2592f1fca82eSMikolaj Golub size_t size; 2593f1fca82eSMikolaj Golub int structsize; 2594f1fca82eSMikolaj Golub 259561b4c627SEdward Tomasz Napierala p = arg; 2596f1fca82eSMikolaj Golub size = sizeof(structsize) + p->p_ucred->cr_ngroups * sizeof(gid_t); 2597f1fca82eSMikolaj Golub if (sb != NULL) { 2598f1fca82eSMikolaj Golub KASSERT(*sizep == size, ("invalid size")); 2599f1fca82eSMikolaj Golub structsize = sizeof(gid_t); 2600f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2601f1fca82eSMikolaj Golub sbuf_bcat(sb, p->p_ucred->cr_groups, p->p_ucred->cr_ngroups * 2602f1fca82eSMikolaj Golub sizeof(gid_t)); 2603f1fca82eSMikolaj Golub } 2604f1fca82eSMikolaj Golub *sizep = size; 2605f1fca82eSMikolaj Golub } 2606f1fca82eSMikolaj Golub 2607f1fca82eSMikolaj Golub static void 2608f1fca82eSMikolaj Golub note_procstat_umask(void *arg, struct sbuf *sb, size_t *sizep) 2609f1fca82eSMikolaj Golub { 2610f1fca82eSMikolaj Golub struct proc *p; 2611f1fca82eSMikolaj Golub size_t size; 2612f1fca82eSMikolaj Golub int structsize; 2613f1fca82eSMikolaj Golub 261461b4c627SEdward Tomasz Napierala p = arg; 261585078b85SConrad Meyer size = sizeof(structsize) + sizeof(p->p_pd->pd_cmask); 2616f1fca82eSMikolaj Golub if (sb != NULL) { 2617f1fca82eSMikolaj Golub KASSERT(*sizep == size, ("invalid size")); 261885078b85SConrad Meyer structsize = sizeof(p->p_pd->pd_cmask); 2619f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 262085078b85SConrad Meyer sbuf_bcat(sb, &p->p_pd->pd_cmask, sizeof(p->p_pd->pd_cmask)); 2621f1fca82eSMikolaj Golub } 2622f1fca82eSMikolaj Golub *sizep = size; 2623f1fca82eSMikolaj Golub } 2624f1fca82eSMikolaj Golub 2625f1fca82eSMikolaj Golub static void 2626f1fca82eSMikolaj Golub note_procstat_rlimit(void *arg, struct sbuf *sb, size_t *sizep) 2627f1fca82eSMikolaj Golub { 2628f1fca82eSMikolaj Golub struct proc *p; 2629f1fca82eSMikolaj Golub struct rlimit rlim[RLIM_NLIMITS]; 2630f1fca82eSMikolaj Golub size_t size; 2631f1fca82eSMikolaj Golub int structsize, i; 2632f1fca82eSMikolaj Golub 263361b4c627SEdward Tomasz Napierala p = arg; 2634f1fca82eSMikolaj Golub size = sizeof(structsize) + sizeof(rlim); 2635f1fca82eSMikolaj Golub if (sb != NULL) { 2636f1fca82eSMikolaj Golub KASSERT(*sizep == size, ("invalid size")); 2637f1fca82eSMikolaj Golub structsize = sizeof(rlim); 2638f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2639f1fca82eSMikolaj Golub PROC_LOCK(p); 2640f1fca82eSMikolaj Golub for (i = 0; i < RLIM_NLIMITS; i++) 2641f6f6d240SMateusz Guzik lim_rlimit_proc(p, i, &rlim[i]); 2642f1fca82eSMikolaj Golub PROC_UNLOCK(p); 2643f1fca82eSMikolaj Golub sbuf_bcat(sb, rlim, sizeof(rlim)); 2644f1fca82eSMikolaj Golub } 2645f1fca82eSMikolaj Golub *sizep = size; 2646f1fca82eSMikolaj Golub } 2647f1fca82eSMikolaj Golub 2648f1fca82eSMikolaj Golub static void 2649f1fca82eSMikolaj Golub note_procstat_osrel(void *arg, struct sbuf *sb, size_t *sizep) 2650f1fca82eSMikolaj Golub { 2651f1fca82eSMikolaj Golub struct proc *p; 2652f1fca82eSMikolaj Golub size_t size; 2653f1fca82eSMikolaj Golub int structsize; 2654f1fca82eSMikolaj Golub 265561b4c627SEdward Tomasz Napierala p = arg; 2656f1fca82eSMikolaj Golub size = sizeof(structsize) + sizeof(p->p_osrel); 2657f1fca82eSMikolaj Golub if (sb != NULL) { 2658f1fca82eSMikolaj Golub KASSERT(*sizep == size, ("invalid size")); 2659f1fca82eSMikolaj Golub structsize = sizeof(p->p_osrel); 2660f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2661f1fca82eSMikolaj Golub sbuf_bcat(sb, &p->p_osrel, sizeof(p->p_osrel)); 2662f1fca82eSMikolaj Golub } 2663f1fca82eSMikolaj Golub *sizep = size; 2664f1fca82eSMikolaj Golub } 2665f1fca82eSMikolaj Golub 2666f1fca82eSMikolaj Golub static void 2667f1fca82eSMikolaj Golub __elfN(note_procstat_psstrings)(void *arg, struct sbuf *sb, size_t *sizep) 2668f1fca82eSMikolaj Golub { 2669f1fca82eSMikolaj Golub struct proc *p; 2670f1fca82eSMikolaj Golub elf_ps_strings_t ps_strings; 2671f1fca82eSMikolaj Golub size_t size; 2672f1fca82eSMikolaj Golub int structsize; 2673f1fca82eSMikolaj Golub 267461b4c627SEdward Tomasz Napierala p = arg; 2675f1fca82eSMikolaj Golub size = sizeof(structsize) + sizeof(ps_strings); 2676f1fca82eSMikolaj Golub if (sb != NULL) { 2677f1fca82eSMikolaj Golub KASSERT(*sizep == size, ("invalid size")); 2678f1fca82eSMikolaj Golub structsize = sizeof(ps_strings); 2679f1fca82eSMikolaj Golub #if defined(COMPAT_FREEBSD32) && __ELF_WORD_SIZE == 32 2680706f4a81SMark Johnston ps_strings = PTROUT(PROC_PS_STRINGS(p)); 2681f1fca82eSMikolaj Golub #else 2682706f4a81SMark Johnston ps_strings = PROC_PS_STRINGS(p); 2683f1fca82eSMikolaj Golub #endif 2684f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2685f1fca82eSMikolaj Golub sbuf_bcat(sb, &ps_strings, sizeof(ps_strings)); 2686f1fca82eSMikolaj Golub } 2687f1fca82eSMikolaj Golub *sizep = size; 2688f1fca82eSMikolaj Golub } 2689f1fca82eSMikolaj Golub 2690f1fca82eSMikolaj Golub static void 2691f1fca82eSMikolaj Golub __elfN(note_procstat_auxv)(void *arg, struct sbuf *sb, size_t *sizep) 2692f1fca82eSMikolaj Golub { 2693f1fca82eSMikolaj Golub struct proc *p; 2694f1fca82eSMikolaj Golub size_t size; 2695f1fca82eSMikolaj Golub int structsize; 2696f1fca82eSMikolaj Golub 269761b4c627SEdward Tomasz Napierala p = arg; 2698f1fca82eSMikolaj Golub if (sb == NULL) { 2699f1fca82eSMikolaj Golub size = 0; 270006250515SEdward Tomasz Napierala sb = sbuf_new(NULL, NULL, AT_COUNT * sizeof(Elf_Auxinfo), 270106250515SEdward Tomasz Napierala SBUF_FIXEDLEN); 27025c32e9fcSAlexander Motin sbuf_set_drain(sb, sbuf_count_drain, &size); 2703f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2704f1fca82eSMikolaj Golub PHOLD(p); 2705f1fca82eSMikolaj Golub proc_getauxv(curthread, p, sb); 2706f1fca82eSMikolaj Golub PRELE(p); 2707f1fca82eSMikolaj Golub sbuf_finish(sb); 2708f1fca82eSMikolaj Golub sbuf_delete(sb); 2709f1fca82eSMikolaj Golub *sizep = size; 2710f1fca82eSMikolaj Golub } else { 2711f1fca82eSMikolaj Golub structsize = sizeof(Elf_Auxinfo); 2712f1fca82eSMikolaj Golub sbuf_bcat(sb, &structsize, sizeof(structsize)); 2713f1fca82eSMikolaj Golub PHOLD(p); 2714f1fca82eSMikolaj Golub proc_getauxv(curthread, p, sb); 2715f1fca82eSMikolaj Golub PRELE(p); 2716f1fca82eSMikolaj Golub } 2717f1fca82eSMikolaj Golub } 2718f1fca82eSMikolaj Golub 271929d4f8bfSKonstantin Belousov #define MAX_NOTES_LOOP 4096 2720a04633ceSAndrew Turner bool 2721364d1b2fSJohn Baldwin __elfN(parse_notes)(const struct image_params *imgp, const Elf_Note *checknote, 272292328a32SKonstantin Belousov const char *note_vendor, const Elf_Phdr *pnote, 272319621645SAlex Richardson bool (*cb)(const Elf_Note *, void *, bool *), void *cb_arg) 272432c01de2SDmitry Chagin { 2725267c52fcSKonstantin Belousov const Elf_Note *note, *note0, *note_end; 272632c01de2SDmitry Chagin const char *note_name; 27276c775eb6SKonstantin Belousov char *buf; 27286c775eb6SKonstantin Belousov int i, error; 272919621645SAlex Richardson bool res; 273032c01de2SDmitry Chagin 27316c775eb6SKonstantin Belousov /* We need some limit, might as well use PAGE_SIZE. */ 27326c775eb6SKonstantin Belousov if (pnote == NULL || pnote->p_filesz > PAGE_SIZE) 273319621645SAlex Richardson return (false); 27346c775eb6SKonstantin Belousov ASSERT_VOP_LOCKED(imgp->vp, "parse_notes"); 27356c775eb6SKonstantin Belousov if (pnote->p_offset > PAGE_SIZE || 27366c775eb6SKonstantin Belousov pnote->p_filesz > PAGE_SIZE - pnote->p_offset) { 27372d6b8546SKonstantin Belousov buf = malloc(pnote->p_filesz, M_TEMP, M_NOWAIT); 27382d6b8546SKonstantin Belousov if (buf == NULL) { 2739b249ce48SMateusz Guzik VOP_UNLOCK(imgp->vp); 27406c775eb6SKonstantin Belousov buf = malloc(pnote->p_filesz, M_TEMP, M_WAITOK); 274178022527SKonstantin Belousov vn_lock(imgp->vp, LK_SHARED | LK_RETRY); 27422d6b8546SKonstantin Belousov } 27436c775eb6SKonstantin Belousov error = vn_rdwr(UIO_READ, imgp->vp, buf, pnote->p_filesz, 27446c775eb6SKonstantin Belousov pnote->p_offset, UIO_SYSSPACE, IO_NODELOCKED, 27456c775eb6SKonstantin Belousov curthread->td_ucred, NOCRED, NULL, curthread); 27466c775eb6SKonstantin Belousov if (error != 0) { 27476c775eb6SKonstantin Belousov uprintf("i/o error PT_NOTE\n"); 2748eda8fe63SKonstantin Belousov goto retf; 27496c775eb6SKonstantin Belousov } 27506c775eb6SKonstantin Belousov note = note0 = (const Elf_Note *)buf; 27516c775eb6SKonstantin Belousov note_end = (const Elf_Note *)(buf + pnote->p_filesz); 27526c775eb6SKonstantin Belousov } else { 27536c775eb6SKonstantin Belousov note = note0 = (const Elf_Note *)(imgp->image_header + 27546c775eb6SKonstantin Belousov pnote->p_offset); 275532c01de2SDmitry Chagin note_end = (const Elf_Note *)(imgp->image_header + 275632c01de2SDmitry Chagin pnote->p_offset + pnote->p_filesz); 27576c775eb6SKonstantin Belousov buf = NULL; 27586c775eb6SKonstantin Belousov } 275929d4f8bfSKonstantin Belousov for (i = 0; i < MAX_NOTES_LOOP && note >= note0 && note < note_end; 276029d4f8bfSKonstantin Belousov i++) { 276129d4f8bfSKonstantin Belousov if (!aligned(note, Elf32_Addr)) { 276229d4f8bfSKonstantin Belousov uprintf("Unaligned ELF note\n"); 276329d4f8bfSKonstantin Belousov goto retf; 276429d4f8bfSKonstantin Belousov } 276529d4f8bfSKonstantin Belousov if ((const char *)note_end - (const char *)note < 276629d4f8bfSKonstantin Belousov sizeof(Elf_Note)) { 276729d4f8bfSKonstantin Belousov uprintf("ELF note to short\n"); 2768eda8fe63SKonstantin Belousov goto retf; 27696c775eb6SKonstantin Belousov } 277092328a32SKonstantin Belousov if (note->n_namesz != checknote->n_namesz || 277192328a32SKonstantin Belousov note->n_descsz != checknote->n_descsz || 277292328a32SKonstantin Belousov note->n_type != checknote->n_type) 277332c01de2SDmitry Chagin goto nextnote; 277432c01de2SDmitry Chagin note_name = (const char *)(note + 1); 277592328a32SKonstantin Belousov if (note_name + checknote->n_namesz >= 277692328a32SKonstantin Belousov (const char *)note_end || strncmp(note_vendor, 277792328a32SKonstantin Belousov note_name, checknote->n_namesz) != 0) 277832c01de2SDmitry Chagin goto nextnote; 277932c01de2SDmitry Chagin 278092328a32SKonstantin Belousov if (cb(note, cb_arg, &res)) 27816c775eb6SKonstantin Belousov goto ret; 278232c01de2SDmitry Chagin nextnote: 278332c01de2SDmitry Chagin note = (const Elf_Note *)((const char *)(note + 1) + 27841b8388cdSMikolaj Golub roundup2(note->n_namesz, ELF_NOTE_ROUNDSIZE) + 27851b8388cdSMikolaj Golub roundup2(note->n_descsz, ELF_NOTE_ROUNDSIZE)); 278632c01de2SDmitry Chagin } 278729d4f8bfSKonstantin Belousov if (i >= MAX_NOTES_LOOP) 278829d4f8bfSKonstantin Belousov uprintf("ELF note parser reached %d notes\n", i); 2789eda8fe63SKonstantin Belousov retf: 279019621645SAlex Richardson res = false; 27916c775eb6SKonstantin Belousov ret: 27926c775eb6SKonstantin Belousov free(buf, M_TEMP); 27936c775eb6SKonstantin Belousov return (res); 279432c01de2SDmitry Chagin } 279532c01de2SDmitry Chagin 279692328a32SKonstantin Belousov struct brandnote_cb_arg { 279792328a32SKonstantin Belousov Elf_Brandnote *brandnote; 279892328a32SKonstantin Belousov int32_t *osrel; 279992328a32SKonstantin Belousov }; 280092328a32SKonstantin Belousov 280119621645SAlex Richardson static bool 280219621645SAlex Richardson brandnote_cb(const Elf_Note *note, void *arg0, bool *res) 280392328a32SKonstantin Belousov { 280492328a32SKonstantin Belousov struct brandnote_cb_arg *arg; 280592328a32SKonstantin Belousov 280692328a32SKonstantin Belousov arg = arg0; 280792328a32SKonstantin Belousov 280892328a32SKonstantin Belousov /* 280992328a32SKonstantin Belousov * Fetch the osreldate for binary from the ELF OSABI-note if 281092328a32SKonstantin Belousov * necessary. 281192328a32SKonstantin Belousov */ 281292328a32SKonstantin Belousov *res = (arg->brandnote->flags & BN_TRANSLATE_OSREL) != 0 && 281392328a32SKonstantin Belousov arg->brandnote->trans_osrel != NULL ? 281419621645SAlex Richardson arg->brandnote->trans_osrel(note, arg->osrel) : true; 281592328a32SKonstantin Belousov 281619621645SAlex Richardson return (true); 281792328a32SKonstantin Belousov } 281892328a32SKonstantin Belousov 2819cefb93f2SKonstantin Belousov static Elf_Note fctl_note = { 2820cefb93f2SKonstantin Belousov .n_namesz = sizeof(FREEBSD_ABI_VENDOR), 2821cefb93f2SKonstantin Belousov .n_descsz = sizeof(uint32_t), 2822cefb93f2SKonstantin Belousov .n_type = NT_FREEBSD_FEATURE_CTL, 2823cefb93f2SKonstantin Belousov }; 2824cefb93f2SKonstantin Belousov 2825cefb93f2SKonstantin Belousov struct fctl_cb_arg { 282619621645SAlex Richardson bool *has_fctl0; 2827cefb93f2SKonstantin Belousov uint32_t *fctl0; 2828cefb93f2SKonstantin Belousov }; 2829cefb93f2SKonstantin Belousov 283019621645SAlex Richardson static bool 283119621645SAlex Richardson note_fctl_cb(const Elf_Note *note, void *arg0, bool *res) 2832cefb93f2SKonstantin Belousov { 2833cefb93f2SKonstantin Belousov struct fctl_cb_arg *arg; 2834cefb93f2SKonstantin Belousov const Elf32_Word *desc; 2835cefb93f2SKonstantin Belousov uintptr_t p; 2836cefb93f2SKonstantin Belousov 2837cefb93f2SKonstantin Belousov arg = arg0; 2838cefb93f2SKonstantin Belousov p = (uintptr_t)(note + 1); 2839cefb93f2SKonstantin Belousov p += roundup2(note->n_namesz, ELF_NOTE_ROUNDSIZE); 2840cefb93f2SKonstantin Belousov desc = (const Elf32_Word *)p; 284119621645SAlex Richardson *arg->has_fctl0 = true; 2842cefb93f2SKonstantin Belousov *arg->fctl0 = desc[0]; 284319621645SAlex Richardson *res = true; 284419621645SAlex Richardson return (true); 2845cefb93f2SKonstantin Belousov } 2846cefb93f2SKonstantin Belousov 284732c01de2SDmitry Chagin /* 2848cefb93f2SKonstantin Belousov * Try to find the appropriate ABI-note section for checknote, fetch 2849cefb93f2SKonstantin Belousov * the osreldate and feature control flags for binary from the ELF 2850cefb93f2SKonstantin Belousov * OSABI-note. Only the first page of the image is searched, the same 2851cefb93f2SKonstantin Belousov * as for headers. 28521a9c7decSKonstantin Belousov */ 285319621645SAlex Richardson static bool 285492328a32SKonstantin Belousov __elfN(check_note)(struct image_params *imgp, Elf_Brandnote *brandnote, 285519621645SAlex Richardson int32_t *osrel, bool *has_fctl0, uint32_t *fctl0) 28561a9c7decSKonstantin Belousov { 28571a9c7decSKonstantin Belousov const Elf_Phdr *phdr; 28581a9c7decSKonstantin Belousov const Elf_Ehdr *hdr; 285992328a32SKonstantin Belousov struct brandnote_cb_arg b_arg; 2860cefb93f2SKonstantin Belousov struct fctl_cb_arg f_arg; 2861cefb93f2SKonstantin Belousov int i, j; 28621a9c7decSKonstantin Belousov 28631a9c7decSKonstantin Belousov hdr = (const Elf_Ehdr *)imgp->image_header; 28641a9c7decSKonstantin Belousov phdr = (const Elf_Phdr *)(imgp->image_header + hdr->e_phoff); 286592328a32SKonstantin Belousov b_arg.brandnote = brandnote; 286692328a32SKonstantin Belousov b_arg.osrel = osrel; 28670cad2aa2SKonstantin Belousov f_arg.has_fctl0 = has_fctl0; 2868cefb93f2SKonstantin Belousov f_arg.fctl0 = fctl0; 28691a9c7decSKonstantin Belousov 28701a9c7decSKonstantin Belousov for (i = 0; i < hdr->e_phnum; i++) { 287192328a32SKonstantin Belousov if (phdr[i].p_type == PT_NOTE && __elfN(parse_notes)(imgp, 287292328a32SKonstantin Belousov &brandnote->hdr, brandnote->vendor, &phdr[i], brandnote_cb, 287392328a32SKonstantin Belousov &b_arg)) { 2874cefb93f2SKonstantin Belousov for (j = 0; j < hdr->e_phnum; j++) { 2875cefb93f2SKonstantin Belousov if (phdr[j].p_type == PT_NOTE && 2876cefb93f2SKonstantin Belousov __elfN(parse_notes)(imgp, &fctl_note, 2877cefb93f2SKonstantin Belousov FREEBSD_ABI_VENDOR, &phdr[j], 2878cefb93f2SKonstantin Belousov note_fctl_cb, &f_arg)) 2879cefb93f2SKonstantin Belousov break; 2880cefb93f2SKonstantin Belousov } 288119621645SAlex Richardson return (true); 28821a9c7decSKonstantin Belousov } 288392328a32SKonstantin Belousov } 288419621645SAlex Richardson return (false); 28851a9c7decSKonstantin Belousov 28861a9c7decSKonstantin Belousov } 28871a9c7decSKonstantin Belousov 28881a9c7decSKonstantin Belousov /* 2889e1743d02SSøren Schmidt * Tell kern_execve.c about it, with a little help from the linker. 2890e1743d02SSøren Schmidt */ 2891a360a43dSJake Burkholder static struct execsw __elfN(execsw) = { 2892b7feabf9SEd Maste .ex_imgact = __CONCAT(exec_, __elfN(imgact)), 2893b7feabf9SEd Maste .ex_name = __XSTRING(__CONCAT(ELF, __ELF_WORD_SIZE)) 2894a360a43dSJake Burkholder }; 2895a360a43dSJake Burkholder EXEC_SET(__CONCAT(elf, __ELF_WORD_SIZE), __elfN(execsw)); 2896e7228204SAlfred Perlstein 2897ed167eaaSKonstantin Belousov static vm_prot_t 2898ed167eaaSKonstantin Belousov __elfN(trans_prot)(Elf_Word flags) 2899ed167eaaSKonstantin Belousov { 2900ed167eaaSKonstantin Belousov vm_prot_t prot; 2901ed167eaaSKonstantin Belousov 2902ed167eaaSKonstantin Belousov prot = 0; 2903ed167eaaSKonstantin Belousov if (flags & PF_X) 2904ed167eaaSKonstantin Belousov prot |= VM_PROT_EXECUTE; 2905ed167eaaSKonstantin Belousov if (flags & PF_W) 2906ed167eaaSKonstantin Belousov prot |= VM_PROT_WRITE; 2907ed167eaaSKonstantin Belousov if (flags & PF_R) 2908ed167eaaSKonstantin Belousov prot |= VM_PROT_READ; 2909eb785fabSKonstantin Belousov #if __ELF_WORD_SIZE == 32 && (defined(__amd64__) || defined(__i386__)) 2910126b36a2SKonstantin Belousov if (i386_read_exec && (flags & PF_R)) 2911676eda08SMarcel Moolenaar prot |= VM_PROT_EXECUTE; 2912676eda08SMarcel Moolenaar #endif 2913ed167eaaSKonstantin Belousov return (prot); 2914ed167eaaSKonstantin Belousov } 2915ed167eaaSKonstantin Belousov 2916ed167eaaSKonstantin Belousov static Elf_Word 2917ed167eaaSKonstantin Belousov __elfN(untrans_prot)(vm_prot_t prot) 2918ed167eaaSKonstantin Belousov { 2919ed167eaaSKonstantin Belousov Elf_Word flags; 2920ed167eaaSKonstantin Belousov 2921ed167eaaSKonstantin Belousov flags = 0; 2922ed167eaaSKonstantin Belousov if (prot & VM_PROT_EXECUTE) 2923ed167eaaSKonstantin Belousov flags |= PF_X; 2924ed167eaaSKonstantin Belousov if (prot & VM_PROT_READ) 2925ed167eaaSKonstantin Belousov flags |= PF_R; 2926ed167eaaSKonstantin Belousov if (prot & VM_PROT_WRITE) 2927ed167eaaSKonstantin Belousov flags |= PF_W; 2928ed167eaaSKonstantin Belousov return (flags); 2929ed167eaaSKonstantin Belousov } 2930