1e1743d02SSøren Schmidt /*- 221a3ee0eSDavid E. O'Brien * Copyright (c) 2000 David O'Brien 3e1743d02SSøren Schmidt * Copyright (c) 1995-1996 S�ren Schmidt 4e1743d02SSøren Schmidt * Copyright (c) 1996 Peter Wemm 5e1743d02SSøren Schmidt * All rights reserved. 6e1743d02SSøren Schmidt * 7e1743d02SSøren Schmidt * Redistribution and use in source and binary forms, with or without 8e1743d02SSøren Schmidt * modification, are permitted provided that the following conditions 9e1743d02SSøren Schmidt * are met: 10e1743d02SSøren Schmidt * 1. Redistributions of source code must retain the above copyright 11e1743d02SSøren Schmidt * notice, this list of conditions and the following disclaimer 12e1743d02SSøren Schmidt * in this position and unchanged. 13e1743d02SSøren Schmidt * 2. Redistributions in binary form must reproduce the above copyright 14e1743d02SSøren Schmidt * notice, this list of conditions and the following disclaimer in the 15e1743d02SSøren Schmidt * documentation and/or other materials provided with the distribution. 16e1743d02SSøren Schmidt * 3. The name of the author may not be used to endorse or promote products 1721dc7d4fSJens Schweikhardt * derived from this software without specific prior written permission 18e1743d02SSøren Schmidt * 19e1743d02SSøren Schmidt * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 20e1743d02SSøren Schmidt * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 21e1743d02SSøren Schmidt * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 22e1743d02SSøren Schmidt * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 23e1743d02SSøren Schmidt * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 24e1743d02SSøren Schmidt * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 25e1743d02SSøren Schmidt * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 26e1743d02SSøren Schmidt * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 27e1743d02SSøren Schmidt * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 28e1743d02SSøren Schmidt * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 29e1743d02SSøren Schmidt * 30c3aac50fSPeter Wemm * $FreeBSD$ 31e1743d02SSøren Schmidt */ 32e1743d02SSøren Schmidt 33e1743d02SSøren Schmidt #include <sys/param.h> 34e1743d02SSøren Schmidt #include <sys/exec.h> 358c64af4fSJohn Polstra #include <sys/fcntl.h> 36e1743d02SSøren Schmidt #include <sys/imgact.h> 37e1743d02SSøren Schmidt #include <sys/imgact_elf.h> 38e1743d02SSøren Schmidt #include <sys/kernel.h> 39f34fa851SJohn Baldwin #include <sys/lock.h> 40e1743d02SSøren Schmidt #include <sys/malloc.h> 4135e0e5b3SJohn Baldwin #include <sys/mutex.h> 428c64af4fSJohn Polstra #include <sys/mman.h> 43a794e791SBruce Evans #include <sys/namei.h> 448c64af4fSJohn Polstra #include <sys/pioctl.h> 45a794e791SBruce Evans #include <sys/proc.h> 468c64af4fSJohn Polstra #include <sys/procfs.h> 478c64af4fSJohn Polstra #include <sys/resourcevar.h> 4836240ea5SDoug Rabson #include <sys/systm.h> 49e1743d02SSøren Schmidt #include <sys/signalvar.h> 508c64af4fSJohn Polstra #include <sys/stat.h> 511005a129SJohn Baldwin #include <sys/sx.h> 528c64af4fSJohn Polstra #include <sys/syscall.h> 53e1743d02SSøren Schmidt #include <sys/sysctl.h> 548c64af4fSJohn Polstra #include <sys/sysent.h> 55a794e791SBruce Evans #include <sys/vnode.h> 56e1743d02SSøren Schmidt 57e1743d02SSøren Schmidt #include <vm/vm.h> 58e1743d02SSøren Schmidt #include <vm/vm_kern.h> 59e1743d02SSøren Schmidt #include <vm/vm_param.h> 60e1743d02SSøren Schmidt #include <vm/pmap.h> 61e1743d02SSøren Schmidt #include <vm/vm_map.h> 620ff27d31SJohn Polstra #include <vm/vm_object.h> 63e1743d02SSøren Schmidt #include <vm/vm_extern.h> 64e1743d02SSøren Schmidt 6552c24af7SPeter Wemm #include <machine/elf.h> 66e1743d02SSøren Schmidt #include <machine/md_var.h> 67e1743d02SSøren Schmidt 68c815a20cSDavid E. O'Brien #define OLD_EI_BRAND 8 69c815a20cSDavid E. O'Brien 7052c24af7SPeter Wemm __ElfType(Brandinfo); 7152c24af7SPeter Wemm __ElfType(Auxargs); 72ecbb00a2SDoug Rabson 733ebc1248SPeter Wemm static int __elfN(check_header)(const Elf_Ehdr *hdr); 745fe3ed62SJake Burkholder static Elf_Brandinfo *__elfN(get_brandinfo)(const Elf_Ehdr *hdr, 755fe3ed62SJake Burkholder const char *interp); 763ebc1248SPeter Wemm static int __elfN(load_file)(struct proc *p, const char *file, u_long *addr, 773ebc1248SPeter Wemm u_long *entry, size_t pagesize); 783ebc1248SPeter Wemm static int __elfN(load_section)(struct proc *p, 790b2ed1aeSJeff Roberson struct vmspace *vmspace, struct vnode *vp, vm_object_t object, 808c64af4fSJohn Polstra vm_offset_t offset, caddr_t vmaddr, size_t memsz, size_t filsz, 813ebc1248SPeter Wemm vm_prot_t prot, size_t pagesize); 823ebc1248SPeter Wemm static int __CONCAT(exec_, __elfN(imgact))(struct image_params *imgp); 83e1743d02SSøren Schmidt 84d8a4f230SBruce Evans static int elf_trace = 0; 85551d79e1SMarcel Moolenaar static int elf_legacy_coredump = 0; 863ebc1248SPeter Wemm #if __ELF_WORD_SIZE == 32 873ebc1248SPeter Wemm SYSCTL_INT(_debug, OID_AUTO, elf32_trace, CTLFLAG_RW, &elf_trace, 0, ""); 88551d79e1SMarcel Moolenaar SYSCTL_INT(_debug, OID_AUTO, elf32_legacy_coredump, CTLFLAG_RW, 89551d79e1SMarcel Moolenaar &elf_legacy_coredump, 0, ""); 903ebc1248SPeter Wemm #else 913ebc1248SPeter Wemm SYSCTL_INT(_debug, OID_AUTO, elf64_trace, CTLFLAG_RW, &elf_trace, 0, ""); 92551d79e1SMarcel Moolenaar SYSCTL_INT(_debug, OID_AUTO, elf64_legacy_coredump, CTLFLAG_RW, 93fa7dd9c5SMatthew Dillon &elf_legacy_coredump, 0, ""); 94551d79e1SMarcel Moolenaar #endif 95e1743d02SSøren Schmidt 963ebc1248SPeter Wemm static Elf_Brandinfo *elf_brand_list[MAX_BRANDS]; 975fe3ed62SJake Burkholder extern int fallback_elf_brand; 98e1743d02SSøren Schmidt 99e1743d02SSøren Schmidt int 1003ebc1248SPeter Wemm __elfN(insert_brand_entry)(Elf_Brandinfo *entry) 101e1743d02SSøren Schmidt { 102e1743d02SSøren Schmidt int i; 103e1743d02SSøren Schmidt 1043ebc1248SPeter Wemm for (i = 0; i < MAX_BRANDS; i++) { 105ea5a2b2eSSøren Schmidt if (elf_brand_list[i] == NULL) { 106ea5a2b2eSSøren Schmidt elf_brand_list[i] = entry; 107e1743d02SSøren Schmidt break; 108e1743d02SSøren Schmidt } 109e1743d02SSøren Schmidt } 110ea5a2b2eSSøren Schmidt if (i == MAX_BRANDS) 111a7cddfedSJake Burkholder return (-1); 112a7cddfedSJake Burkholder return (0); 113e1743d02SSøren Schmidt } 114e1743d02SSøren Schmidt 115e1743d02SSøren Schmidt int 1163ebc1248SPeter Wemm __elfN(remove_brand_entry)(Elf_Brandinfo *entry) 117e1743d02SSøren Schmidt { 118e1743d02SSøren Schmidt int i; 119e1743d02SSøren Schmidt 1203ebc1248SPeter Wemm for (i = 0; i < MAX_BRANDS; i++) { 121ea5a2b2eSSøren Schmidt if (elf_brand_list[i] == entry) { 122ea5a2b2eSSøren Schmidt elf_brand_list[i] = NULL; 123e1743d02SSøren Schmidt break; 124e1743d02SSøren Schmidt } 125e1743d02SSøren Schmidt } 126ea5a2b2eSSøren Schmidt if (i == MAX_BRANDS) 127a7cddfedSJake Burkholder return (-1); 128a7cddfedSJake Burkholder return (0); 129e1743d02SSøren Schmidt } 130e1743d02SSøren Schmidt 131096977faSMark Newton int 1323ebc1248SPeter Wemm __elfN(brand_inuse)(Elf_Brandinfo *entry) 133096977faSMark Newton { 134096977faSMark Newton struct proc *p; 135553629ebSJake Burkholder int rval = FALSE; 136096977faSMark Newton 1371005a129SJohn Baldwin sx_slock(&allproc_lock); 1382e3c8fcbSPoul-Henning Kamp LIST_FOREACH(p, &allproc, p_list) { 139553629ebSJake Burkholder if (p->p_sysent == entry->sysvec) { 140553629ebSJake Burkholder rval = TRUE; 141553629ebSJake Burkholder break; 142096977faSMark Newton } 143553629ebSJake Burkholder } 1441005a129SJohn Baldwin sx_sunlock(&allproc_lock); 145096977faSMark Newton 146553629ebSJake Burkholder return (rval); 147096977faSMark Newton } 148096977faSMark Newton 1495fe3ed62SJake Burkholder static Elf_Brandinfo * 1505fe3ed62SJake Burkholder __elfN(get_brandinfo)(const Elf_Ehdr *hdr, const char *interp) 1515fe3ed62SJake Burkholder { 1525fe3ed62SJake Burkholder Elf_Brandinfo *bi; 1535fe3ed62SJake Burkholder int i; 1545fe3ed62SJake Burkholder 1555fe3ed62SJake Burkholder /* 1565fe3ed62SJake Burkholder * We support three types of branding -- (1) the ELF EI_OSABI field 1575fe3ed62SJake Burkholder * that SCO added to the ELF spec, (2) FreeBSD 3.x's traditional string 1585fe3ed62SJake Burkholder * branding w/in the ELF header, and (3) path of the `interp_path' 1595fe3ed62SJake Burkholder * field. We should also look for an ".note.ABI-tag" ELF section now 1605fe3ed62SJake Burkholder * in all Linux ELF binaries, FreeBSD 4.1+, and some NetBSD ones. 1615fe3ed62SJake Burkholder */ 1625fe3ed62SJake Burkholder 1635fe3ed62SJake Burkholder /* If the executable has a brand, search for it in the brand list. */ 1645fe3ed62SJake Burkholder for (i = 0; i < MAX_BRANDS; i++) { 1655fe3ed62SJake Burkholder bi = elf_brand_list[i]; 1665fe3ed62SJake Burkholder if (bi != NULL && hdr->e_machine == bi->machine && 1675fe3ed62SJake Burkholder (hdr->e_ident[EI_OSABI] == bi->brand || 1685fe3ed62SJake Burkholder strncmp((const char *)&hdr->e_ident[OLD_EI_BRAND], 1695fe3ed62SJake Burkholder bi->compat_3_brand, strlen(bi->compat_3_brand)) == 0)) 1705fe3ed62SJake Burkholder return (bi); 1715fe3ed62SJake Burkholder } 1725fe3ed62SJake Burkholder 1735fe3ed62SJake Burkholder /* Lacking a known brand, search for a recognized interpreter. */ 1745fe3ed62SJake Burkholder if (interp != NULL) { 1755fe3ed62SJake Burkholder for (i = 0; i < MAX_BRANDS; i++) { 1765fe3ed62SJake Burkholder bi = elf_brand_list[i]; 1775fe3ed62SJake Burkholder if (bi != NULL && hdr->e_machine == bi->machine && 1785fe3ed62SJake Burkholder strcmp(interp, bi->interp_path) == 0) 1795fe3ed62SJake Burkholder return (bi); 1805fe3ed62SJake Burkholder } 1815fe3ed62SJake Burkholder } 1825fe3ed62SJake Burkholder 1835fe3ed62SJake Burkholder /* Lacking a recognized interpreter, try the default brand */ 1845fe3ed62SJake Burkholder for (i = 0; i < MAX_BRANDS; i++) { 1855fe3ed62SJake Burkholder bi = elf_brand_list[i]; 1865fe3ed62SJake Burkholder if (bi != NULL && hdr->e_machine == bi->machine && 1875fe3ed62SJake Burkholder fallback_elf_brand == bi->brand) 1885fe3ed62SJake Burkholder return (bi); 1895fe3ed62SJake Burkholder } 1905fe3ed62SJake Burkholder return (NULL); 1915fe3ed62SJake Burkholder } 1925fe3ed62SJake Burkholder 193e1743d02SSøren Schmidt static int 1943ebc1248SPeter Wemm __elfN(check_header)(const Elf_Ehdr *hdr) 195e1743d02SSøren Schmidt { 196d0ca7c29SPeter Wemm Elf_Brandinfo *bi; 1973ebc1248SPeter Wemm int i; 1983ebc1248SPeter Wemm 19952c24af7SPeter Wemm if (!IS_ELF(*hdr) || 20052c24af7SPeter Wemm hdr->e_ident[EI_CLASS] != ELF_TARG_CLASS || 20152c24af7SPeter Wemm hdr->e_ident[EI_DATA] != ELF_TARG_DATA || 20252c24af7SPeter Wemm hdr->e_ident[EI_VERSION] != EV_CURRENT) 203a7cddfedSJake Burkholder return (ENOEXEC); 204e1743d02SSøren Schmidt 2053ebc1248SPeter Wemm /* 2063ebc1248SPeter Wemm * Make sure we have at least one brand for this machine. 2073ebc1248SPeter Wemm */ 2083ebc1248SPeter Wemm 2093ebc1248SPeter Wemm for (i = 0; i < MAX_BRANDS; i++) { 210d0ca7c29SPeter Wemm bi = elf_brand_list[i]; 211d0ca7c29SPeter Wemm if (bi != NULL && bi->machine == hdr->e_machine) 2123ebc1248SPeter Wemm break; 2133ebc1248SPeter Wemm } 2143ebc1248SPeter Wemm if (i == MAX_BRANDS) 215a7cddfedSJake Burkholder return (ENOEXEC); 216e1743d02SSøren Schmidt 217c33fe779SJohn Polstra if (hdr->e_version != ELF_TARG_VER) 218a7cddfedSJake Burkholder return (ENOEXEC); 219e1743d02SSøren Schmidt 220a7cddfedSJake Burkholder return (0); 221e1743d02SSøren Schmidt } 222e1743d02SSøren Schmidt 223e1743d02SSøren Schmidt static int 2243ebc1248SPeter Wemm __elfN(map_partial)(vm_map_t map, vm_object_t object, vm_ooffset_t offset, 2253ebc1248SPeter Wemm vm_offset_t start, vm_offset_t end, vm_prot_t prot, 2263ebc1248SPeter Wemm vm_prot_t max) 2273ebc1248SPeter Wemm { 2283ebc1248SPeter Wemm int error, rv; 2293ebc1248SPeter Wemm vm_offset_t off; 2303ebc1248SPeter Wemm vm_offset_t data_buf = 0; 2313ebc1248SPeter Wemm 2323ebc1248SPeter Wemm /* 2333ebc1248SPeter Wemm * Create the page if it doesn't exist yet. Ignore errors. 2343ebc1248SPeter Wemm */ 2353ebc1248SPeter Wemm vm_map_lock(map); 23681f223caSJake Burkholder vm_map_insert(map, NULL, 0, trunc_page(start), round_page(end), max, 23781f223caSJake Burkholder max, 0); 2383ebc1248SPeter Wemm vm_map_unlock(map); 2393ebc1248SPeter Wemm 2403ebc1248SPeter Wemm /* 2413ebc1248SPeter Wemm * Find the page from the underlying object. 2423ebc1248SPeter Wemm */ 2433ebc1248SPeter Wemm if (object) { 2443ebc1248SPeter Wemm vm_object_reference(object); 2453ebc1248SPeter Wemm rv = vm_map_find(exec_map, 2463ebc1248SPeter Wemm object, 2473ebc1248SPeter Wemm trunc_page(offset), 2483ebc1248SPeter Wemm &data_buf, 2493ebc1248SPeter Wemm PAGE_SIZE, 2503ebc1248SPeter Wemm TRUE, 2513ebc1248SPeter Wemm VM_PROT_READ, 2523ebc1248SPeter Wemm VM_PROT_ALL, 2533ebc1248SPeter Wemm MAP_COPY_ON_WRITE | MAP_PREFAULT_PARTIAL); 2543ebc1248SPeter Wemm if (rv != KERN_SUCCESS) { 2553ebc1248SPeter Wemm vm_object_deallocate(object); 256a7cddfedSJake Burkholder return (rv); 2573ebc1248SPeter Wemm } 2583ebc1248SPeter Wemm 2593ebc1248SPeter Wemm off = offset - trunc_page(offset); 260ca0387efSJake Burkholder error = copyout((caddr_t)data_buf + off, (caddr_t)start, 261ca0387efSJake Burkholder end - start); 2623ebc1248SPeter Wemm vm_map_remove(exec_map, data_buf, data_buf + PAGE_SIZE); 2633ebc1248SPeter Wemm if (error) { 264a7cddfedSJake Burkholder return (KERN_FAILURE); 2653ebc1248SPeter Wemm } 2663ebc1248SPeter Wemm } 2673ebc1248SPeter Wemm 268a7cddfedSJake Burkholder return (KERN_SUCCESS); 2693ebc1248SPeter Wemm } 2703ebc1248SPeter Wemm 2713ebc1248SPeter Wemm static int 2723ebc1248SPeter Wemm __elfN(map_insert)(vm_map_t map, vm_object_t object, vm_ooffset_t offset, 2733ebc1248SPeter Wemm vm_offset_t start, vm_offset_t end, vm_prot_t prot, 2743ebc1248SPeter Wemm vm_prot_t max, int cow) 2753ebc1248SPeter Wemm { 2763ebc1248SPeter Wemm int rv; 2773ebc1248SPeter Wemm 2783ebc1248SPeter Wemm if (start != trunc_page(start)) { 27981f223caSJake Burkholder rv = __elfN(map_partial)(map, object, offset, start, 28081f223caSJake Burkholder round_page(start), prot, max); 2813ebc1248SPeter Wemm if (rv) 282a7cddfedSJake Burkholder return (rv); 2833ebc1248SPeter Wemm offset += round_page(start) - start; 2843ebc1248SPeter Wemm start = round_page(start); 2853ebc1248SPeter Wemm } 2863ebc1248SPeter Wemm if (end != round_page(end)) { 28781f223caSJake Burkholder rv = __elfN(map_partial)(map, object, offset + 28881f223caSJake Burkholder trunc_page(end) - start, trunc_page(end), end, prot, max); 2893ebc1248SPeter Wemm if (rv) 290a7cddfedSJake Burkholder return (rv); 2913ebc1248SPeter Wemm end = trunc_page(end); 2923ebc1248SPeter Wemm } 2933ebc1248SPeter Wemm if (end > start) { 2943ebc1248SPeter Wemm if (offset & PAGE_MASK) { 2953ebc1248SPeter Wemm vm_offset_t data_buf, off; 2963ebc1248SPeter Wemm vm_size_t sz; 2973ebc1248SPeter Wemm int error; 2983ebc1248SPeter Wemm 2993ebc1248SPeter Wemm /* 3003ebc1248SPeter Wemm * The mapping is not page aligned. This means we have 3013ebc1248SPeter Wemm * to copy the data. Sigh. 3023ebc1248SPeter Wemm */ 30381f223caSJake Burkholder rv = vm_map_find(map, 0, 0, &start, end - start, 3043ebc1248SPeter Wemm FALSE, prot, max, 0); 3053ebc1248SPeter Wemm if (rv) 306a7cddfedSJake Burkholder return (rv); 3073ebc1248SPeter Wemm while (start < end) { 3083ebc1248SPeter Wemm vm_object_reference(object); 3093ebc1248SPeter Wemm rv = vm_map_find(exec_map, 3103ebc1248SPeter Wemm object, 3113ebc1248SPeter Wemm trunc_page(offset), 3123ebc1248SPeter Wemm &data_buf, 3133ebc1248SPeter Wemm 2 * PAGE_SIZE, 3143ebc1248SPeter Wemm TRUE, 3153ebc1248SPeter Wemm VM_PROT_READ, 3163ebc1248SPeter Wemm VM_PROT_ALL, 3173ebc1248SPeter Wemm (MAP_COPY_ON_WRITE 3183ebc1248SPeter Wemm | MAP_PREFAULT_PARTIAL)); 3193ebc1248SPeter Wemm if (rv != KERN_SUCCESS) { 3203ebc1248SPeter Wemm vm_object_deallocate(object); 321a7cddfedSJake Burkholder return (rv); 3223ebc1248SPeter Wemm } 3233ebc1248SPeter Wemm off = offset - trunc_page(offset); 3243ebc1248SPeter Wemm sz = end - start; 3253ebc1248SPeter Wemm if (sz > PAGE_SIZE) 3263ebc1248SPeter Wemm sz = PAGE_SIZE; 3273ebc1248SPeter Wemm error = copyout((caddr_t)data_buf + off, 3283ebc1248SPeter Wemm (caddr_t)start, sz); 3293ebc1248SPeter Wemm vm_map_remove(exec_map, data_buf, 3303ebc1248SPeter Wemm data_buf + 2 * PAGE_SIZE); 3313ebc1248SPeter Wemm if (error) { 332a7cddfedSJake Burkholder return (KERN_FAILURE); 3333ebc1248SPeter Wemm } 3343ebc1248SPeter Wemm start += sz; 3353ebc1248SPeter Wemm } 3363ebc1248SPeter Wemm rv = KERN_SUCCESS; 3373ebc1248SPeter Wemm } else { 3383ebc1248SPeter Wemm vm_map_lock(map); 3393ebc1248SPeter Wemm rv = vm_map_insert(map, object, offset, start, end, 3403ebc1248SPeter Wemm prot, max, cow); 3413ebc1248SPeter Wemm vm_map_unlock(map); 3423ebc1248SPeter Wemm } 343a7cddfedSJake Burkholder return (rv); 3443ebc1248SPeter Wemm } else { 345a7cddfedSJake Burkholder return (KERN_SUCCESS); 3463ebc1248SPeter Wemm } 3473ebc1248SPeter Wemm } 3483ebc1248SPeter Wemm 3493ebc1248SPeter Wemm static int 3503ebc1248SPeter Wemm __elfN(load_section)(struct proc *p, struct vmspace *vmspace, 3513ebc1248SPeter Wemm struct vnode *vp, vm_object_t object, vm_offset_t offset, 3523ebc1248SPeter Wemm caddr_t vmaddr, size_t memsz, size_t filsz, vm_prot_t prot, 3533ebc1248SPeter Wemm size_t pagesize) 354e1743d02SSøren Schmidt { 355e1743d02SSøren Schmidt size_t map_len; 356e1743d02SSøren Schmidt vm_offset_t map_addr; 357fa7dd9c5SMatthew Dillon int error, rv, cow; 358e1743d02SSøren Schmidt size_t copy_len; 35952c24af7SPeter Wemm vm_offset_t file_addr; 36052c24af7SPeter Wemm vm_offset_t data_buf = 0; 36152c24af7SPeter Wemm 3620cddd8f0SMatthew Dillon GIANT_REQUIRED; 3630cddd8f0SMatthew Dillon 36452c24af7SPeter Wemm error = 0; 365e1743d02SSøren Schmidt 36625ead034SBrian Feldman /* 36725ead034SBrian Feldman * It's necessary to fail if the filsz + offset taken from the 36825ead034SBrian Feldman * header is greater than the actual file pager object's size. 36925ead034SBrian Feldman * If we were to allow this, then the vm_map_find() below would 37025ead034SBrian Feldman * walk right off the end of the file object and into the ether. 37125ead034SBrian Feldman * 37225ead034SBrian Feldman * While I'm here, might as well check for something else that 37325ead034SBrian Feldman * is invalid: filsz cannot be greater than memsz. 37425ead034SBrian Feldman */ 37525ead034SBrian Feldman if ((off_t)filsz + offset > object->un_pager.vnp.vnp_size || 37625ead034SBrian Feldman filsz > memsz) { 37725ead034SBrian Feldman uprintf("elf_load_section: truncated ELF file\n"); 37825ead034SBrian Feldman return (ENOEXEC); 37925ead034SBrian Feldman } 38025ead034SBrian Feldman 3813ebc1248SPeter Wemm #define trunc_page_ps(va, ps) ((va) & ~(ps - 1)) 3823ebc1248SPeter Wemm #define round_page_ps(va, ps) (((va) + (ps - 1)) & ~(ps - 1)) 3833ebc1248SPeter Wemm 3843ebc1248SPeter Wemm map_addr = trunc_page_ps((vm_offset_t)vmaddr, pagesize); 3853ebc1248SPeter Wemm file_addr = trunc_page_ps(offset, pagesize); 386e1743d02SSøren Schmidt 387e1743d02SSøren Schmidt /* 38852c24af7SPeter Wemm * We have two choices. We can either clear the data in the last page 38952c24af7SPeter Wemm * of an oversized mapping, or we can start the anon mapping a page 39052c24af7SPeter Wemm * early and copy the initialized data into that first page. We 39152c24af7SPeter Wemm * choose the second.. 39252c24af7SPeter Wemm */ 39352c24af7SPeter Wemm if (memsz > filsz) 3943ebc1248SPeter Wemm map_len = trunc_page_ps(offset + filsz, pagesize) - file_addr; 39552c24af7SPeter Wemm else 3963ebc1248SPeter Wemm map_len = round_page_ps(offset + filsz, pagesize) - file_addr; 39752c24af7SPeter Wemm 39852c24af7SPeter Wemm if (map_len != 0) { 39952c24af7SPeter Wemm vm_object_reference(object); 400fa7dd9c5SMatthew Dillon 401fa7dd9c5SMatthew Dillon /* cow flags: don't dump readonly sections in core */ 402fa7dd9c5SMatthew Dillon cow = MAP_COPY_ON_WRITE | MAP_PREFAULT | 403fa7dd9c5SMatthew Dillon (prot & VM_PROT_WRITE ? 0 : MAP_DISABLE_COREDUMP); 404fa7dd9c5SMatthew Dillon 4053ebc1248SPeter Wemm rv = __elfN(map_insert)(&vmspace->vm_map, 40652c24af7SPeter Wemm object, 40752c24af7SPeter Wemm file_addr, /* file offset */ 40852c24af7SPeter Wemm map_addr, /* virtual start */ 40952c24af7SPeter Wemm map_addr + map_len,/* virtual end */ 41052c24af7SPeter Wemm prot, 41152c24af7SPeter Wemm VM_PROT_ALL, 412fa7dd9c5SMatthew Dillon cow); 413a3021f91SBoris Popov if (rv != KERN_SUCCESS) { 414a3021f91SBoris Popov vm_object_deallocate(object); 415a7cddfedSJake Burkholder return (EINVAL); 416a3021f91SBoris Popov } 41752c24af7SPeter Wemm 41852c24af7SPeter Wemm /* we can stop now if we've covered it all */ 41923955314SAlfred Perlstein if (memsz == filsz) { 420a7cddfedSJake Burkholder return (0); 42152c24af7SPeter Wemm } 42223955314SAlfred Perlstein } 42352c24af7SPeter Wemm 42452c24af7SPeter Wemm 42552c24af7SPeter Wemm /* 42652c24af7SPeter Wemm * We have to get the remaining bit of the file into the first part 42752c24af7SPeter Wemm * of the oversized map segment. This is normally because the .data 42852c24af7SPeter Wemm * segment in the file is extended to provide bss. It's a neat idea 42952c24af7SPeter Wemm * to try and save a page, but it's a pain in the behind to implement. 430e1743d02SSøren Schmidt */ 4313ebc1248SPeter Wemm copy_len = (offset + filsz) - trunc_page_ps(offset + filsz, pagesize); 4323ebc1248SPeter Wemm map_addr = trunc_page_ps((vm_offset_t)vmaddr + filsz, pagesize); 433ca0387efSJake Burkholder map_len = round_page_ps((vm_offset_t)vmaddr + memsz, pagesize) - 434ca0387efSJake Burkholder map_addr; 435e1743d02SSøren Schmidt 43652c24af7SPeter Wemm /* This had damn well better be true! */ 4378191d577SPeter Wemm if (map_len != 0) { 43881f223caSJake Burkholder rv = __elfN(map_insert)(&vmspace->vm_map, NULL, 0, map_addr, 43981f223caSJake Burkholder map_addr + map_len, VM_PROT_ALL, VM_PROT_ALL, 0); 44023955314SAlfred Perlstein if (rv != KERN_SUCCESS) { 441a7cddfedSJake Burkholder return (EINVAL); 4428191d577SPeter Wemm } 44323955314SAlfred Perlstein } 444e1743d02SSøren Schmidt 44552c24af7SPeter Wemm if (copy_len != 0) { 4463ebc1248SPeter Wemm vm_offset_t off; 44752c24af7SPeter Wemm vm_object_reference(object); 44852c24af7SPeter Wemm rv = vm_map_find(exec_map, 44952c24af7SPeter Wemm object, 45052c24af7SPeter Wemm trunc_page(offset + filsz), 45152c24af7SPeter Wemm &data_buf, 452e1743d02SSøren Schmidt PAGE_SIZE, 45352c24af7SPeter Wemm TRUE, 454e1743d02SSøren Schmidt VM_PROT_READ, 45552c24af7SPeter Wemm VM_PROT_ALL, 456e972780aSAlan Cox MAP_COPY_ON_WRITE | MAP_PREFAULT_PARTIAL); 45752c24af7SPeter Wemm if (rv != KERN_SUCCESS) { 45852c24af7SPeter Wemm vm_object_deallocate(object); 459a7cddfedSJake Burkholder return (EINVAL); 46052c24af7SPeter Wemm } 461e1743d02SSøren Schmidt 46252c24af7SPeter Wemm /* send the page fragment to user space */ 46381f223caSJake Burkholder off = trunc_page_ps(offset + filsz, pagesize) - 46481f223caSJake Burkholder trunc_page(offset + filsz); 4653ebc1248SPeter Wemm error = copyout((caddr_t)data_buf + off, (caddr_t)map_addr, 4663ebc1248SPeter Wemm copy_len); 46752c24af7SPeter Wemm vm_map_remove(exec_map, data_buf, data_buf + PAGE_SIZE); 46823955314SAlfred Perlstein if (error) { 46952c24af7SPeter Wemm return (error); 47052c24af7SPeter Wemm } 47123955314SAlfred Perlstein } 472e1743d02SSøren Schmidt 473e1743d02SSøren Schmidt /* 4743ebc1248SPeter Wemm * set it to the specified protection. 4753ebc1248SPeter Wemm * XXX had better undo the damage from pasting over the cracks here! 476e1743d02SSøren Schmidt */ 4773ebc1248SPeter Wemm vm_map_protect(&vmspace->vm_map, trunc_page(map_addr), 4783ebc1248SPeter Wemm round_page(map_addr + map_len), prot, FALSE); 4798191d577SPeter Wemm 480a7cddfedSJake Burkholder return (error); 481e1743d02SSøren Schmidt } 482e1743d02SSøren Schmidt 483c33fe779SJohn Polstra /* 484c33fe779SJohn Polstra * Load the file "file" into memory. It may be either a shared object 485c33fe779SJohn Polstra * or an executable. 486c33fe779SJohn Polstra * 487c33fe779SJohn Polstra * The "addr" reference parameter is in/out. On entry, it specifies 488c33fe779SJohn Polstra * the address where a shared object should be loaded. If the file is 489c33fe779SJohn Polstra * an executable, this value is ignored. On exit, "addr" specifies 490c33fe779SJohn Polstra * where the file was actually loaded. 491c33fe779SJohn Polstra * 492c33fe779SJohn Polstra * The "entry" reference parameter is out only. On exit, it specifies 493c33fe779SJohn Polstra * the entry point for the loaded file. 494c33fe779SJohn Polstra */ 495e1743d02SSøren Schmidt static int 4963ebc1248SPeter Wemm __elfN(load_file)(struct proc *p, const char *file, u_long *addr, 4973ebc1248SPeter Wemm u_long *entry, size_t pagesize) 498e1743d02SSøren Schmidt { 499911c2be0SMark Peek struct { 500911c2be0SMark Peek struct nameidata nd; 501911c2be0SMark Peek struct vattr attr; 502911c2be0SMark Peek struct image_params image_params; 503911c2be0SMark Peek } *tempdata; 504d254af07SMatthew Dillon const Elf_Ehdr *hdr = NULL; 505d254af07SMatthew Dillon const Elf_Phdr *phdr = NULL; 506911c2be0SMark Peek struct nameidata *nd; 507e1743d02SSøren Schmidt struct vmspace *vmspace = p->p_vmspace; 508911c2be0SMark Peek struct vattr *attr; 509911c2be0SMark Peek struct image_params *imgp; 51052c24af7SPeter Wemm vm_prot_t prot; 511c33fe779SJohn Polstra u_long rbase; 512c33fe779SJohn Polstra u_long base_addr = 0; 513c33fe779SJohn Polstra int error, i, numsegs; 514e1743d02SSøren Schmidt 515b40ce416SJulian Elischer if (curthread->td_proc != p) 516b40ce416SJulian Elischer panic("elf_load_file - thread"); /* XXXKSE DIAGNOSTIC */ 517b40ce416SJulian Elischer 518911c2be0SMark Peek tempdata = malloc(sizeof(*tempdata), M_TEMP, M_WAITOK); 519911c2be0SMark Peek nd = &tempdata->nd; 520911c2be0SMark Peek attr = &tempdata->attr; 521911c2be0SMark Peek imgp = &tempdata->image_params; 522911c2be0SMark Peek 523c8a79999SPeter Wemm /* 524c8a79999SPeter Wemm * Initialize part of the common data 525c8a79999SPeter Wemm */ 526c8a79999SPeter Wemm imgp->proc = p; 527450ffb44SRobert Watson imgp->userspace_argv = NULL; 528450ffb44SRobert Watson imgp->userspace_envv = NULL; 529911c2be0SMark Peek imgp->attr = attr; 530c8a79999SPeter Wemm imgp->firstpage = NULL; 531c8a79999SPeter Wemm imgp->image_header = (char *)kmem_alloc_wait(exec_map, PAGE_SIZE); 5320b2ed1aeSJeff Roberson imgp->object = NULL; 5336d7bdc8dSRobert Watson imgp->execlabel = NULL; 534c8a79999SPeter Wemm 535c8a79999SPeter Wemm if (imgp->image_header == NULL) { 536911c2be0SMark Peek nd->ni_vp = NULL; 537c8a79999SPeter Wemm error = ENOMEM; 538c8a79999SPeter Wemm goto fail; 539c8a79999SPeter Wemm } 540c8a79999SPeter Wemm 541b40ce416SJulian Elischer /* XXXKSE */ 542b40ce416SJulian Elischer NDINIT(nd, LOOKUP, LOCKLEAF|FOLLOW, UIO_SYSSPACE, file, curthread); 543e1743d02SSøren Schmidt 544911c2be0SMark Peek if ((error = namei(nd)) != 0) { 545911c2be0SMark Peek nd->ni_vp = NULL; 546e1743d02SSøren Schmidt goto fail; 547e1743d02SSøren Schmidt } 548911c2be0SMark Peek NDFREE(nd, NDF_ONLY_PNBUF); 549911c2be0SMark Peek imgp->vp = nd->ni_vp; 550c8a79999SPeter Wemm 551e1743d02SSøren Schmidt /* 552e1743d02SSøren Schmidt * Check permissions, modes, uid, etc on the file, and "open" it. 553e1743d02SSøren Schmidt */ 554c8a79999SPeter Wemm error = exec_check_permissions(imgp); 555c8a79999SPeter Wemm if (error) { 556b40ce416SJulian Elischer VOP_UNLOCK(nd->ni_vp, 0, curthread); /* XXXKSE */ 557c8a79999SPeter Wemm goto fail; 558c8a79999SPeter Wemm } 559e1743d02SSøren Schmidt 560c8a79999SPeter Wemm error = exec_map_first_page(imgp); 56125ead034SBrian Feldman /* 56225ead034SBrian Feldman * Also make certain that the interpreter stays the same, so set 563e6e370a7SJeff Roberson * its VV_TEXT flag, too. 56425ead034SBrian Feldman */ 56525ead034SBrian Feldman if (error == 0) 566e6e370a7SJeff Roberson nd->ni_vp->v_vflag |= VV_TEXT; 567e6e370a7SJeff Roberson 5680b2ed1aeSJeff Roberson VOP_GETVOBJECT(nd->ni_vp, &imgp->object); 5690b2ed1aeSJeff Roberson vm_object_reference(imgp->object); 5700b2ed1aeSJeff Roberson 571b40ce416SJulian Elischer VOP_UNLOCK(nd->ni_vp, 0, curthread); /* XXXKSE */ 572e1743d02SSøren Schmidt if (error) 573e1743d02SSøren Schmidt goto fail; 574e1743d02SSøren Schmidt 575d254af07SMatthew Dillon hdr = (const Elf_Ehdr *)imgp->image_header; 5763ebc1248SPeter Wemm if ((error = __elfN(check_header)(hdr)) != 0) 577e1743d02SSøren Schmidt goto fail; 578c33fe779SJohn Polstra if (hdr->e_type == ET_DYN) 579c33fe779SJohn Polstra rbase = *addr; 580c33fe779SJohn Polstra else if (hdr->e_type == ET_EXEC) 581c33fe779SJohn Polstra rbase = 0; 582c33fe779SJohn Polstra else { 583c33fe779SJohn Polstra error = ENOEXEC; 584c33fe779SJohn Polstra goto fail; 585c33fe779SJohn Polstra } 586e1743d02SSøren Schmidt 587c8a79999SPeter Wemm /* Only support headers that fit within first page for now */ 58852c24af7SPeter Wemm if ((hdr->e_phoff > PAGE_SIZE) || 58952c24af7SPeter Wemm (hdr->e_phoff + hdr->e_phentsize * hdr->e_phnum) > PAGE_SIZE) { 590c8a79999SPeter Wemm error = ENOEXEC; 591e1743d02SSøren Schmidt goto fail; 592c8a79999SPeter Wemm } 593c8a79999SPeter Wemm 594d254af07SMatthew Dillon phdr = (const Elf_Phdr *)(imgp->image_header + hdr->e_phoff); 595e1743d02SSøren Schmidt 596c33fe779SJohn Polstra for (i = 0, numsegs = 0; i < hdr->e_phnum; i++) { 59752c24af7SPeter Wemm if (phdr[i].p_type == PT_LOAD) { /* Loadable segment */ 59852c24af7SPeter Wemm prot = 0; 599e1743d02SSøren Schmidt if (phdr[i].p_flags & PF_X) 600e1743d02SSøren Schmidt prot |= VM_PROT_EXECUTE; 601e1743d02SSøren Schmidt if (phdr[i].p_flags & PF_W) 602e1743d02SSøren Schmidt prot |= VM_PROT_WRITE; 603e1743d02SSøren Schmidt if (phdr[i].p_flags & PF_R) 604e1743d02SSøren Schmidt prot |= VM_PROT_READ; 605e1743d02SSøren Schmidt 60681f223caSJake Burkholder if ((error = __elfN(load_section)(p, vmspace, 60781f223caSJake Burkholder nd->ni_vp, imgp->object, phdr[i].p_offset, 60881f223caSJake Burkholder (caddr_t)(uintptr_t)phdr[i].p_vaddr + rbase, 60981f223caSJake Burkholder phdr[i].p_memsz, phdr[i].p_filesz, prot, 61081f223caSJake Burkholder pagesize)) != 0) 611e1743d02SSøren Schmidt goto fail; 612e1743d02SSøren Schmidt /* 613c33fe779SJohn Polstra * Establish the base address if this is the 614c33fe779SJohn Polstra * first segment. 615e1743d02SSøren Schmidt */ 616c33fe779SJohn Polstra if (numsegs == 0) 617ca0387efSJake Burkholder base_addr = trunc_page(phdr[i].p_vaddr + 618ca0387efSJake Burkholder rbase); 619c33fe779SJohn Polstra numsegs++; 620e1743d02SSøren Schmidt } 621e1743d02SSøren Schmidt } 622c33fe779SJohn Polstra *addr = base_addr; 623c33fe779SJohn Polstra *entry = (unsigned long)hdr->e_entry + rbase; 624e1743d02SSøren Schmidt 625e1743d02SSøren Schmidt fail: 626c8a79999SPeter Wemm if (imgp->firstpage) 627c8a79999SPeter Wemm exec_unmap_first_page(imgp); 628c8a79999SPeter Wemm if (imgp->image_header) 629c8a79999SPeter Wemm kmem_free_wakeup(exec_map, (vm_offset_t)imgp->image_header, 630c8a79999SPeter Wemm PAGE_SIZE); 6310b2ed1aeSJeff Roberson if (imgp->object) 6320b2ed1aeSJeff Roberson vm_object_deallocate(imgp->object); 6330b2ed1aeSJeff Roberson 634911c2be0SMark Peek if (nd->ni_vp) 635911c2be0SMark Peek vrele(nd->ni_vp); 636911c2be0SMark Peek 637911c2be0SMark Peek free(tempdata, M_TEMP); 638e1743d02SSøren Schmidt 639a7cddfedSJake Burkholder return (error); 640e1743d02SSøren Schmidt } 641e1743d02SSøren Schmidt 642303b270bSEivind Eklund static int 6433ebc1248SPeter Wemm __CONCAT(exec_, __elfN(imgact))(struct image_params *imgp) 644e1743d02SSøren Schmidt { 645ecbb00a2SDoug Rabson const Elf_Ehdr *hdr = (const Elf_Ehdr *)imgp->image_header; 646f5ef029eSPoul-Henning Kamp const Elf_Phdr *phdr; 647ecbb00a2SDoug Rabson Elf_Auxargs *elf_auxargs = NULL; 6485856e12eSJohn Dyson struct vmspace *vmspace; 64952c24af7SPeter Wemm vm_prot_t prot; 65021c2d047SMatthew Dillon u_long text_size = 0, data_size = 0, total_size = 0; 651e1743d02SSøren Schmidt u_long text_addr = 0, data_addr = 0; 652cac45152SMatthew Dillon u_long seg_size, seg_addr; 653e1743d02SSøren Schmidt u_long addr, entry = 0, proghdr = 0; 65452c24af7SPeter Wemm int error, i; 655c8a79999SPeter Wemm const char *interp = NULL; 656d1dbc694SJohn Polstra Elf_Brandinfo *brand_info; 657911c2be0SMark Peek char *path; 658619eb6e5SJeff Roberson struct thread *td = curthread; 6595fe3ed62SJake Burkholder struct sysentvec *sv; 660e1743d02SSøren Schmidt 6610cddd8f0SMatthew Dillon GIANT_REQUIRED; 6620cddd8f0SMatthew Dillon 663e1743d02SSøren Schmidt /* 664e1743d02SSøren Schmidt * Do we have a valid ELF header ? 665e1743d02SSøren Schmidt */ 6663ebc1248SPeter Wemm if (__elfN(check_header)(hdr) != 0 || hdr->e_type != ET_EXEC) 667a7cddfedSJake Burkholder return (-1); 668e1743d02SSøren Schmidt 669e1743d02SSøren Schmidt /* 670e1743d02SSøren Schmidt * From here on down, we return an errno, not -1, as we've 671e1743d02SSøren Schmidt * detected an ELF file. 672e1743d02SSøren Schmidt */ 673e1743d02SSøren Schmidt 674e1743d02SSøren Schmidt if ((hdr->e_phoff > PAGE_SIZE) || 67552c24af7SPeter Wemm (hdr->e_phoff + hdr->e_phentsize * hdr->e_phnum) > PAGE_SIZE) { 676c8a79999SPeter Wemm /* Only support headers in first page for now */ 677a7cddfedSJake Burkholder return (ENOEXEC); 678e1743d02SSøren Schmidt } 67952c24af7SPeter Wemm phdr = (const Elf_Phdr *)(imgp->image_header + hdr->e_phoff); 680e1743d02SSøren Schmidt 681e1743d02SSøren Schmidt /* 682e1743d02SSøren Schmidt * From this point on, we may have resources that need to be freed. 683e1743d02SSøren Schmidt */ 68425ead034SBrian Feldman 685619eb6e5SJeff Roberson VOP_UNLOCK(imgp->vp, 0, td); 68625ead034SBrian Feldman 6875fe3ed62SJake Burkholder for (i = 0; i < hdr->e_phnum; i++) { 6885fe3ed62SJake Burkholder switch (phdr[i].p_type) { 6895fe3ed62SJake Burkholder case PT_INTERP: /* Path to interpreter */ 6905fe3ed62SJake Burkholder if (phdr[i].p_filesz > MAXPATHLEN || 6915fe3ed62SJake Burkholder phdr[i].p_offset + phdr[i].p_filesz > PAGE_SIZE) { 6925fe3ed62SJake Burkholder error = ENOEXEC; 693e1743d02SSøren Schmidt goto fail; 6945fe3ed62SJake Burkholder } 6955fe3ed62SJake Burkholder interp = imgp->image_header + phdr[i].p_offset; 6965fe3ed62SJake Burkholder break; 6975fe3ed62SJake Burkholder default: 6983ebc1248SPeter Wemm break; 6993ebc1248SPeter Wemm } 7003ebc1248SPeter Wemm } 7013ebc1248SPeter Wemm 7025fe3ed62SJake Burkholder brand_info = __elfN(get_brandinfo)(hdr, interp); 7035fe3ed62SJake Burkholder if (brand_info == NULL) { 7045fe3ed62SJake Burkholder uprintf("ELF binary type \"%u\" not known.\n", 7055fe3ed62SJake Burkholder hdr->e_ident[EI_OSABI]); 7065fe3ed62SJake Burkholder error = ENOEXEC; 7075fe3ed62SJake Burkholder goto fail; 7083ebc1248SPeter Wemm } 7095fe3ed62SJake Burkholder sv = brand_info->sysvec; 7103ebc1248SPeter Wemm 7115fe3ed62SJake Burkholder if ((error = exec_extract_strings(imgp)) != 0) 7125fe3ed62SJake Burkholder goto fail; 7135fe3ed62SJake Burkholder 71405ba50f5SJake Burkholder exec_new_vmspace(imgp, sv); 715e1743d02SSøren Schmidt 7165856e12eSJohn Dyson vmspace = imgp->proc->p_vmspace; 7175856e12eSJohn Dyson 718e1743d02SSøren Schmidt for (i = 0; i < hdr->e_phnum; i++) { 719e1743d02SSøren Schmidt switch (phdr[i].p_type) { 720e1743d02SSøren Schmidt case PT_LOAD: /* Loadable segment */ 72152c24af7SPeter Wemm prot = 0; 722e1743d02SSøren Schmidt if (phdr[i].p_flags & PF_X) 723e1743d02SSøren Schmidt prot |= VM_PROT_EXECUTE; 724e1743d02SSøren Schmidt if (phdr[i].p_flags & PF_W) 725e1743d02SSøren Schmidt prot |= VM_PROT_WRITE; 726e1743d02SSøren Schmidt if (phdr[i].p_flags & PF_R) 727e1743d02SSøren Schmidt prot |= VM_PROT_READ; 728e1743d02SSøren Schmidt 7293ebc1248SPeter Wemm #if defined(__ia64__) && __ELF_WORD_SIZE == 32 && defined(IA32_ME_HARDER) 7303ebc1248SPeter Wemm /* 7313ebc1248SPeter Wemm * Some x86 binaries assume read == executable, 7323ebc1248SPeter Wemm * notably the M3 runtime and therefore cvsup 7333ebc1248SPeter Wemm */ 7343ebc1248SPeter Wemm if (prot & VM_PROT_READ) 7353ebc1248SPeter Wemm prot |= VM_PROT_EXECUTE; 7363ebc1248SPeter Wemm #endif 7373ebc1248SPeter Wemm 73881f223caSJake Burkholder if ((error = __elfN(load_section)(imgp->proc, vmspace, 73981f223caSJake Burkholder imgp->vp, imgp->object, phdr[i].p_offset, 7403ebc1248SPeter Wemm (caddr_t)(uintptr_t)phdr[i].p_vaddr, 74181f223caSJake Burkholder phdr[i].p_memsz, phdr[i].p_filesz, prot, 7425fe3ed62SJake Burkholder sv->sv_pagesize)) != 0) 743e1743d02SSøren Schmidt goto fail; 744e1743d02SSøren Schmidt 745cac45152SMatthew Dillon seg_addr = trunc_page(phdr[i].p_vaddr); 746cac45152SMatthew Dillon seg_size = round_page(phdr[i].p_memsz + 747cac45152SMatthew Dillon phdr[i].p_vaddr - seg_addr); 748cac45152SMatthew Dillon 749e1743d02SSøren Schmidt /* 75021c2d047SMatthew Dillon * Is this .text or .data? We can't use 75121c2d047SMatthew Dillon * VM_PROT_WRITE or VM_PROT_EXEC, it breaks the 75221c2d047SMatthew Dillon * alpha terribly and possibly does other bad 75321c2d047SMatthew Dillon * things so we stick to the old way of figuring 75421c2d047SMatthew Dillon * it out: If the segment contains the program 75521c2d047SMatthew Dillon * entry point, it's a text segment, otherwise it 75621c2d047SMatthew Dillon * is a data segment. 75721c2d047SMatthew Dillon * 75821c2d047SMatthew Dillon * Note that obreak() assumes that data_addr + 75921c2d047SMatthew Dillon * data_size == end of data load area, and the ELF 76021c2d047SMatthew Dillon * file format expects segments to be sorted by 76121c2d047SMatthew Dillon * address. If multiple data segments exist, the 76221c2d047SMatthew Dillon * last one will be used. 763e1743d02SSøren Schmidt */ 7649782ecbaSPeter Wemm if (hdr->e_entry >= phdr[i].p_vaddr && 7659782ecbaSPeter Wemm hdr->e_entry < (phdr[i].p_vaddr + 7669782ecbaSPeter Wemm phdr[i].p_memsz)) { 7679782ecbaSPeter Wemm text_size = seg_size; 7689782ecbaSPeter Wemm text_addr = seg_addr; 7699782ecbaSPeter Wemm entry = (u_long)hdr->e_entry; 7709782ecbaSPeter Wemm } else { 77121c2d047SMatthew Dillon data_size = seg_size; 772cac45152SMatthew Dillon data_addr = seg_addr; 773cac45152SMatthew Dillon } 77421c2d047SMatthew Dillon total_size += seg_size; 77596725dd0SAlexander Kabaev break; 77696725dd0SAlexander Kabaev case PT_PHDR: /* Program header table info */ 77796725dd0SAlexander Kabaev proghdr = phdr[i].p_vaddr; 77896725dd0SAlexander Kabaev break; 77996725dd0SAlexander Kabaev default: 78096725dd0SAlexander Kabaev break; 78196725dd0SAlexander Kabaev } 78296725dd0SAlexander Kabaev } 78396725dd0SAlexander Kabaev 78496725dd0SAlexander Kabaev if (data_addr == 0 && data_size == 0) { 78596725dd0SAlexander Kabaev data_addr = text_addr; 78696725dd0SAlexander Kabaev data_size = text_size; 78796725dd0SAlexander Kabaev } 788cac45152SMatthew Dillon 789cac45152SMatthew Dillon /* 790cac45152SMatthew Dillon * Check limits. It should be safe to check the 79196725dd0SAlexander Kabaev * limits after loading the segments since we do 79296725dd0SAlexander Kabaev * not actually fault in all the segments pages. 793cac45152SMatthew Dillon */ 794cac45152SMatthew Dillon if (data_size > 795cac45152SMatthew Dillon imgp->proc->p_rlimit[RLIMIT_DATA].rlim_cur || 796cac45152SMatthew Dillon text_size > maxtsiz || 79721c2d047SMatthew Dillon total_size > 798cac45152SMatthew Dillon imgp->proc->p_rlimit[RLIMIT_VMEM].rlim_cur) { 799cac45152SMatthew Dillon error = ENOMEM; 800cac45152SMatthew Dillon goto fail; 801cac45152SMatthew Dillon } 802e1743d02SSøren Schmidt 803e1743d02SSøren Schmidt vmspace->vm_tsize = text_size >> PAGE_SHIFT; 8047cd99438SBruce Evans vmspace->vm_taddr = (caddr_t)(uintptr_t)text_addr; 805e1743d02SSøren Schmidt vmspace->vm_dsize = data_size >> PAGE_SHIFT; 8067cd99438SBruce Evans vmspace->vm_daddr = (caddr_t)(uintptr_t)data_addr; 807e1743d02SSøren Schmidt 80847633640SJohn Polstra addr = ELF_RTLD_ADDR(vmspace); 809e1743d02SSøren Schmidt 810ea5a2b2eSSøren Schmidt imgp->entry_addr = entry; 811ea5a2b2eSSøren Schmidt 8125fe3ed62SJake Burkholder imgp->proc->p_sysent = sv; 813d1dbc694SJohn Polstra if (interp != NULL) { 814911c2be0SMark Peek path = malloc(MAXPATHLEN, M_TEMP, M_WAITOK); 8158cf03452SJake Burkholder snprintf(path, MAXPATHLEN, "%s%s", brand_info->emul_path, 8168cf03452SJake Burkholder interp); 8173ebc1248SPeter Wemm if ((error = __elfN(load_file)(imgp->proc, path, &addr, 8185fe3ed62SJake Burkholder &imgp->entry_addr, sv->sv_pagesize)) != 0) { 81981f223caSJake Burkholder if ((error = __elfN(load_file)(imgp->proc, interp, 8205fe3ed62SJake Burkholder &addr, &imgp->entry_addr, sv->sv_pagesize)) != 0) { 821ca0387efSJake Burkholder uprintf("ELF interpreter %s not found\n", 822ca0387efSJake Burkholder path); 823911c2be0SMark Peek free(path, M_TEMP); 824e1743d02SSøren Schmidt goto fail; 825e1743d02SSøren Schmidt } 826e1743d02SSøren Schmidt } 827911c2be0SMark Peek free(path, M_TEMP); 8287a583b02SMarcel Moolenaar } 829ea5a2b2eSSøren Schmidt 830e1743d02SSøren Schmidt /* 831e1743d02SSøren Schmidt * Construct auxargs table (used by the fixup routine) 832e1743d02SSøren Schmidt */ 833ecbb00a2SDoug Rabson elf_auxargs = malloc(sizeof(Elf_Auxargs), M_TEMP, M_WAITOK); 834e1743d02SSøren Schmidt elf_auxargs->execfd = -1; 835e1743d02SSøren Schmidt elf_auxargs->phdr = proghdr; 836e1743d02SSøren Schmidt elf_auxargs->phent = hdr->e_phentsize; 837e1743d02SSøren Schmidt elf_auxargs->phnum = hdr->e_phnum; 838e1743d02SSøren Schmidt elf_auxargs->pagesz = PAGE_SIZE; 839e1743d02SSøren Schmidt elf_auxargs->base = addr; 840e1743d02SSøren Schmidt elf_auxargs->flags = 0; 841e1743d02SSøren Schmidt elf_auxargs->entry = entry; 842e1743d02SSøren Schmidt elf_auxargs->trace = elf_trace; 843e1743d02SSøren Schmidt 844e1743d02SSøren Schmidt imgp->auxargs = elf_auxargs; 845e1743d02SSøren Schmidt imgp->interpreted = 0; 846e1743d02SSøren Schmidt 847e1743d02SSøren Schmidt fail: 848619eb6e5SJeff Roberson vn_lock(imgp->vp, LK_EXCLUSIVE | LK_RETRY, td); 849a7cddfedSJake Burkholder return (error); 850e1743d02SSøren Schmidt } 851e1743d02SSøren Schmidt 8523ebc1248SPeter Wemm #if __ELF_WORD_SIZE == 32 8533ebc1248SPeter Wemm #define suword suword32 8543ebc1248SPeter Wemm #define stacktype u_int32_t 8553ebc1248SPeter Wemm #else 8563ebc1248SPeter Wemm #define suword suword64 8573ebc1248SPeter Wemm #define stacktype u_int64_t 8583ebc1248SPeter Wemm #endif 8593ebc1248SPeter Wemm 8603ebc1248SPeter Wemm int 8613ebc1248SPeter Wemm __elfN(freebsd_fixup)(register_t **stack_base, struct image_params *imgp) 862e1743d02SSøren Schmidt { 863ecbb00a2SDoug Rabson Elf_Auxargs *args = (Elf_Auxargs *)imgp->auxargs; 8643ebc1248SPeter Wemm stacktype *base; 8653ebc1248SPeter Wemm stacktype *pos; 866e1743d02SSøren Schmidt 8673ebc1248SPeter Wemm base = (stacktype *)*stack_base; 8683ebc1248SPeter Wemm pos = base + (imgp->argc + imgp->envc + 2); 869e1743d02SSøren Schmidt 870e1743d02SSøren Schmidt if (args->trace) { 871e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_DEBUG, 1); 872e1743d02SSøren Schmidt } 873e1743d02SSøren Schmidt if (args->execfd != -1) { 874e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_EXECFD, args->execfd); 875e1743d02SSøren Schmidt } 876e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PHDR, args->phdr); 877e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PHENT, args->phent); 878e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PHNUM, args->phnum); 879e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_PAGESZ, args->pagesz); 880e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_FLAGS, args->flags); 881e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_ENTRY, args->entry); 882e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_BASE, args->base); 883e1743d02SSøren Schmidt AUXARGS_ENTRY(pos, AT_NULL, 0); 884e1743d02SSøren Schmidt 885e1743d02SSøren Schmidt free(imgp->auxargs, M_TEMP); 886e1743d02SSøren Schmidt imgp->auxargs = NULL; 887e1743d02SSøren Schmidt 8883ebc1248SPeter Wemm base--; 8893ebc1248SPeter Wemm suword(base, (long)imgp->argc); 8903ebc1248SPeter Wemm *stack_base = (register_t *)base; 891a7cddfedSJake Burkholder return (0); 892e1743d02SSøren Schmidt } 893e1743d02SSøren Schmidt 894e1743d02SSøren Schmidt /* 8958c64af4fSJohn Polstra * Code for generating ELF core dumps. 8968c64af4fSJohn Polstra */ 8978c64af4fSJohn Polstra 8984d77a549SAlfred Perlstein typedef void (*segment_callback)(vm_map_entry_t, void *); 8990ff27d31SJohn Polstra 9000ff27d31SJohn Polstra /* Closure for cb_put_phdr(). */ 9010ff27d31SJohn Polstra struct phdr_closure { 9020ff27d31SJohn Polstra Elf_Phdr *phdr; /* Program header to fill in */ 9030ff27d31SJohn Polstra Elf_Off offset; /* Offset of segment in core file */ 9040ff27d31SJohn Polstra }; 9050ff27d31SJohn Polstra 9060ff27d31SJohn Polstra /* Closure for cb_size_segment(). */ 9070ff27d31SJohn Polstra struct sseg_closure { 9080ff27d31SJohn Polstra int count; /* Count of writable segments. */ 9090ff27d31SJohn Polstra size_t size; /* Total size of all writable segments. */ 9100ff27d31SJohn Polstra }; 9110ff27d31SJohn Polstra 9124d77a549SAlfred Perlstein static void cb_put_phdr(vm_map_entry_t, void *); 9134d77a549SAlfred Perlstein static void cb_size_segment(vm_map_entry_t, void *); 9144d77a549SAlfred Perlstein static void each_writable_segment(struct proc *, segment_callback, void *); 9153ebc1248SPeter Wemm static int __elfN(corehdr)(struct thread *, struct vnode *, struct ucred *, 9164d77a549SAlfred Perlstein int, void *, size_t); 9173ebc1248SPeter Wemm static void __elfN(puthdr)(struct proc *, void *, size_t *, 9184d77a549SAlfred Perlstein const prstatus_t *, const prfpregset_t *, const prpsinfo_t *, int); 9193ebc1248SPeter Wemm static void __elfN(putnote)(void *, size_t *, const char *, int, 9204d77a549SAlfred Perlstein const void *, size_t); 9218c64af4fSJohn Polstra 9228c64af4fSJohn Polstra extern int osreldate; 9238c64af4fSJohn Polstra 9248c64af4fSJohn Polstra int 9253ebc1248SPeter Wemm __elfN(coredump)(td, vp, limit) 926b40ce416SJulian Elischer struct thread *td; 9278c64af4fSJohn Polstra register struct vnode *vp; 928fca666a1SJulian Elischer off_t limit; 929fca666a1SJulian Elischer { 930b40ce416SJulian Elischer register struct proc *p = td->td_proc; 931a854ed98SJohn Baldwin register struct ucred *cred = td->td_ucred; 932fca666a1SJulian Elischer int error = 0; 9330ff27d31SJohn Polstra struct sseg_closure seginfo; 9340ff27d31SJohn Polstra void *hdr; 9358c64af4fSJohn Polstra size_t hdrsize; 9368c64af4fSJohn Polstra 9370ff27d31SJohn Polstra /* Size the program segments. */ 9380ff27d31SJohn Polstra seginfo.count = 0; 9390ff27d31SJohn Polstra seginfo.size = 0; 9400ff27d31SJohn Polstra each_writable_segment(p, cb_size_segment, &seginfo); 9410ff27d31SJohn Polstra 9420ff27d31SJohn Polstra /* 9430ff27d31SJohn Polstra * Calculate the size of the core file header area by making 9440ff27d31SJohn Polstra * a dry run of generating it. Nothing is written, but the 9450ff27d31SJohn Polstra * size is calculated. 9460ff27d31SJohn Polstra */ 9470ff27d31SJohn Polstra hdrsize = 0; 9483ebc1248SPeter Wemm __elfN(puthdr)((struct proc *)NULL, (void *)NULL, &hdrsize, 9490ff27d31SJohn Polstra (const prstatus_t *)NULL, (const prfpregset_t *)NULL, 9500ff27d31SJohn Polstra (const prpsinfo_t *)NULL, seginfo.count); 9510ff27d31SJohn Polstra 952fca666a1SJulian Elischer if (hdrsize + seginfo.size >= limit) 9538c64af4fSJohn Polstra return (EFAULT); 9540ff27d31SJohn Polstra 9550ff27d31SJohn Polstra /* 9560ff27d31SJohn Polstra * Allocate memory for building the header, fill it up, 9570ff27d31SJohn Polstra * and write it out. 9580ff27d31SJohn Polstra */ 9590ff27d31SJohn Polstra hdr = malloc(hdrsize, M_TEMP, M_WAITOK); 9600ff27d31SJohn Polstra if (hdr == NULL) { 961a7cddfedSJake Burkholder return (EINVAL); 9620ff27d31SJohn Polstra } 9633ebc1248SPeter Wemm error = __elfN(corehdr)(td, vp, cred, seginfo.count, hdr, hdrsize); 9640ff27d31SJohn Polstra 9650ff27d31SJohn Polstra /* Write the contents of all of the writable segments. */ 9660ff27d31SJohn Polstra if (error == 0) { 9670ff27d31SJohn Polstra Elf_Phdr *php; 9680ff27d31SJohn Polstra off_t offset; 9690ff27d31SJohn Polstra int i; 9700ff27d31SJohn Polstra 9710ff27d31SJohn Polstra php = (Elf_Phdr *)((char *)hdr + sizeof(Elf_Ehdr)) + 1; 9720ff27d31SJohn Polstra offset = hdrsize; 9730ff27d31SJohn Polstra for (i = 0; i < seginfo.count; i++) { 97406ae1e91SMatthew Dillon error = vn_rdwr_inchunks(UIO_WRITE, vp, 9753ebc1248SPeter Wemm (caddr_t)(uintptr_t)php->p_vaddr, 9760ff27d31SJohn Polstra php->p_filesz, offset, UIO_USERSPACE, 9779ca43589SRobert Watson IO_UNIT | IO_DIRECT, cred, NOCRED, (int *)NULL, 9789ca43589SRobert Watson curthread); /* XXXKSE */ 9790ff27d31SJohn Polstra if (error != 0) 9800ff27d31SJohn Polstra break; 9810ff27d31SJohn Polstra offset += php->p_filesz; 9820ff27d31SJohn Polstra php++; 9830ff27d31SJohn Polstra } 9840ff27d31SJohn Polstra } 9850ff27d31SJohn Polstra free(hdr, M_TEMP); 9860ff27d31SJohn Polstra 987a7cddfedSJake Burkholder return (error); 9888c64af4fSJohn Polstra } 9898c64af4fSJohn Polstra 9900ff27d31SJohn Polstra /* 9910ff27d31SJohn Polstra * A callback for each_writable_segment() to write out the segment's 9920ff27d31SJohn Polstra * program header entry. 9930ff27d31SJohn Polstra */ 9940ff27d31SJohn Polstra static void 9950ff27d31SJohn Polstra cb_put_phdr(entry, closure) 9960ff27d31SJohn Polstra vm_map_entry_t entry; 9970ff27d31SJohn Polstra void *closure; 9980ff27d31SJohn Polstra { 9990ff27d31SJohn Polstra struct phdr_closure *phc = (struct phdr_closure *)closure; 10000ff27d31SJohn Polstra Elf_Phdr *phdr = phc->phdr; 10010ff27d31SJohn Polstra 10020ff27d31SJohn Polstra phc->offset = round_page(phc->offset); 10030ff27d31SJohn Polstra 10040ff27d31SJohn Polstra phdr->p_type = PT_LOAD; 10050ff27d31SJohn Polstra phdr->p_offset = phc->offset; 10060ff27d31SJohn Polstra phdr->p_vaddr = entry->start; 10070ff27d31SJohn Polstra phdr->p_paddr = 0; 10080ff27d31SJohn Polstra phdr->p_filesz = phdr->p_memsz = entry->end - entry->start; 10090ff27d31SJohn Polstra phdr->p_align = PAGE_SIZE; 10100ff27d31SJohn Polstra phdr->p_flags = 0; 10110ff27d31SJohn Polstra if (entry->protection & VM_PROT_READ) 10120ff27d31SJohn Polstra phdr->p_flags |= PF_R; 10130ff27d31SJohn Polstra if (entry->protection & VM_PROT_WRITE) 10140ff27d31SJohn Polstra phdr->p_flags |= PF_W; 10150ff27d31SJohn Polstra if (entry->protection & VM_PROT_EXECUTE) 10160ff27d31SJohn Polstra phdr->p_flags |= PF_X; 10170ff27d31SJohn Polstra 10180ff27d31SJohn Polstra phc->offset += phdr->p_filesz; 10190ff27d31SJohn Polstra phc->phdr++; 10200ff27d31SJohn Polstra } 10210ff27d31SJohn Polstra 10220ff27d31SJohn Polstra /* 10230ff27d31SJohn Polstra * A callback for each_writable_segment() to gather information about 10240ff27d31SJohn Polstra * the number of segments and their total size. 10250ff27d31SJohn Polstra */ 10260ff27d31SJohn Polstra static void 10270ff27d31SJohn Polstra cb_size_segment(entry, closure) 10280ff27d31SJohn Polstra vm_map_entry_t entry; 10290ff27d31SJohn Polstra void *closure; 10300ff27d31SJohn Polstra { 10310ff27d31SJohn Polstra struct sseg_closure *ssc = (struct sseg_closure *)closure; 10320ff27d31SJohn Polstra 10330ff27d31SJohn Polstra ssc->count++; 10340ff27d31SJohn Polstra ssc->size += entry->end - entry->start; 10350ff27d31SJohn Polstra } 10360ff27d31SJohn Polstra 10370ff27d31SJohn Polstra /* 10380ff27d31SJohn Polstra * For each writable segment in the process's memory map, call the given 10390ff27d31SJohn Polstra * function with a pointer to the map entry and some arbitrary 10400ff27d31SJohn Polstra * caller-supplied data. 10410ff27d31SJohn Polstra */ 10420ff27d31SJohn Polstra static void 10430ff27d31SJohn Polstra each_writable_segment(p, func, closure) 10440ff27d31SJohn Polstra struct proc *p; 10450ff27d31SJohn Polstra segment_callback func; 10460ff27d31SJohn Polstra void *closure; 10470ff27d31SJohn Polstra { 10480ff27d31SJohn Polstra vm_map_t map = &p->p_vmspace->vm_map; 10490ff27d31SJohn Polstra vm_map_entry_t entry; 10500ff27d31SJohn Polstra 10510ff27d31SJohn Polstra for (entry = map->header.next; entry != &map->header; 10520ff27d31SJohn Polstra entry = entry->next) { 10530ff27d31SJohn Polstra vm_object_t obj; 10540ff27d31SJohn Polstra 1055fa7dd9c5SMatthew Dillon /* 1056fa7dd9c5SMatthew Dillon * Don't dump inaccessible mappings, deal with legacy 1057fa7dd9c5SMatthew Dillon * coredump mode. 1058fa7dd9c5SMatthew Dillon * 1059fa7dd9c5SMatthew Dillon * Note that read-only segments related to the elf binary 1060fa7dd9c5SMatthew Dillon * are marked MAP_ENTRY_NOCOREDUMP now so we no longer 1061fa7dd9c5SMatthew Dillon * need to arbitrarily ignore such segments. 1062fa7dd9c5SMatthew Dillon */ 1063fa7dd9c5SMatthew Dillon if (elf_legacy_coredump) { 1064fa7dd9c5SMatthew Dillon if ((entry->protection & VM_PROT_RW) != VM_PROT_RW) 10650ff27d31SJohn Polstra continue; 1066fa7dd9c5SMatthew Dillon } else { 1067fa7dd9c5SMatthew Dillon if ((entry->protection & VM_PROT_ALL) == 0) 1068fa7dd9c5SMatthew Dillon continue; 1069fa7dd9c5SMatthew Dillon } 10700ff27d31SJohn Polstra 10719730a5daSPaul Saab /* 1072fa7dd9c5SMatthew Dillon * Dont include memory segment in the coredump if 1073fa7dd9c5SMatthew Dillon * MAP_NOCORE is set in mmap(2) or MADV_NOCORE in 1074fa7dd9c5SMatthew Dillon * madvise(2). Do not dump submaps (i.e. parts of the 1075fa7dd9c5SMatthew Dillon * kernel map). 10769730a5daSPaul Saab */ 1077fa7dd9c5SMatthew Dillon if (entry->eflags & (MAP_ENTRY_NOCOREDUMP|MAP_ENTRY_IS_SUB_MAP)) 10789730a5daSPaul Saab continue; 10799730a5daSPaul Saab 10800ff27d31SJohn Polstra if ((obj = entry->object.vm_object) == NULL) 10810ff27d31SJohn Polstra continue; 10820ff27d31SJohn Polstra 10830ff27d31SJohn Polstra /* Find the deepest backing object. */ 10840ff27d31SJohn Polstra while (obj->backing_object != NULL) 10850ff27d31SJohn Polstra obj = obj->backing_object; 10860ff27d31SJohn Polstra 10870ff27d31SJohn Polstra /* Ignore memory-mapped devices and such things. */ 10880ff27d31SJohn Polstra if (obj->type != OBJT_DEFAULT && 10890ff27d31SJohn Polstra obj->type != OBJT_SWAP && 10900ff27d31SJohn Polstra obj->type != OBJT_VNODE) 10910ff27d31SJohn Polstra continue; 10920ff27d31SJohn Polstra 10930ff27d31SJohn Polstra (*func)(entry, closure); 10940ff27d31SJohn Polstra } 10950ff27d31SJohn Polstra } 10960ff27d31SJohn Polstra 10970ff27d31SJohn Polstra /* 10980ff27d31SJohn Polstra * Write the core file header to the file, including padding up to 10990ff27d31SJohn Polstra * the page boundary. 11000ff27d31SJohn Polstra */ 11018c64af4fSJohn Polstra static int 11023ebc1248SPeter Wemm __elfN(corehdr)(td, vp, cred, numsegs, hdr, hdrsize) 1103b40ce416SJulian Elischer struct thread *td; 11048c64af4fSJohn Polstra struct vnode *vp; 11058c64af4fSJohn Polstra struct ucred *cred; 11060ff27d31SJohn Polstra int numsegs; 11070ff27d31SJohn Polstra size_t hdrsize; 11080ff27d31SJohn Polstra void *hdr; 11098c64af4fSJohn Polstra { 1110911c2be0SMark Peek struct { 11118c64af4fSJohn Polstra prstatus_t status; 11128c64af4fSJohn Polstra prfpregset_t fpregset; 11138c64af4fSJohn Polstra prpsinfo_t psinfo; 1114911c2be0SMark Peek } *tempdata; 1115b40ce416SJulian Elischer struct proc *p = td->td_proc; 1116911c2be0SMark Peek size_t off; 1117911c2be0SMark Peek prstatus_t *status; 1118911c2be0SMark Peek prfpregset_t *fpregset; 1119911c2be0SMark Peek prpsinfo_t *psinfo; 1120911c2be0SMark Peek 1121546a92c4SBrian Somers tempdata = malloc(sizeof(*tempdata), M_TEMP, M_ZERO | M_WAITOK); 1122911c2be0SMark Peek status = &tempdata->status; 1123911c2be0SMark Peek fpregset = &tempdata->fpregset; 1124911c2be0SMark Peek psinfo = &tempdata->psinfo; 11258c64af4fSJohn Polstra 11268c64af4fSJohn Polstra /* Gather the information for the header. */ 1127911c2be0SMark Peek status->pr_version = PRSTATUS_VERSION; 1128911c2be0SMark Peek status->pr_statussz = sizeof(prstatus_t); 1129911c2be0SMark Peek status->pr_gregsetsz = sizeof(gregset_t); 1130911c2be0SMark Peek status->pr_fpregsetsz = sizeof(fpregset_t); 1131911c2be0SMark Peek status->pr_osreldate = osreldate; 1132911c2be0SMark Peek status->pr_cursig = p->p_sig; 1133911c2be0SMark Peek status->pr_pid = p->p_pid; 1134b40ce416SJulian Elischer fill_regs(td, &status->pr_reg); 11358c64af4fSJohn Polstra 1136b40ce416SJulian Elischer fill_fpregs(td, fpregset); 11378c64af4fSJohn Polstra 1138911c2be0SMark Peek psinfo->pr_version = PRPSINFO_VERSION; 1139911c2be0SMark Peek psinfo->pr_psinfosz = sizeof(prpsinfo_t); 1140e80fb434SRobert Drehmel strlcpy(psinfo->pr_fname, p->p_comm, sizeof(psinfo->pr_fname)); 1141911c2be0SMark Peek 11420ff27d31SJohn Polstra /* XXX - We don't fill in the command line arguments properly yet. */ 1143e80fb434SRobert Drehmel strlcpy(psinfo->pr_psargs, p->p_comm, sizeof(psinfo->pr_psargs)); 11448c64af4fSJohn Polstra 11458c64af4fSJohn Polstra /* Fill in the header. */ 11460ff27d31SJohn Polstra bzero(hdr, hdrsize); 11478c64af4fSJohn Polstra off = 0; 11483ebc1248SPeter Wemm __elfN(puthdr)(p, hdr, &off, status, fpregset, psinfo, numsegs); 1149911c2be0SMark Peek 1150911c2be0SMark Peek free(tempdata, M_TEMP); 11518c64af4fSJohn Polstra 11528c64af4fSJohn Polstra /* Write it to the core file. */ 1153a7cddfedSJake Burkholder return (vn_rdwr_inchunks(UIO_WRITE, vp, hdr, hdrsize, (off_t)0, 11549ca43589SRobert Watson UIO_SYSSPACE, IO_UNIT | IO_DIRECT, cred, NOCRED, NULL, 1155a7cddfedSJake Burkholder td)); /* XXXKSE */ 1156dada0278SJohn Polstra } 1157dada0278SJohn Polstra 11588c64af4fSJohn Polstra static void 11593ebc1248SPeter Wemm __elfN(puthdr)(struct proc *p, void *dst, size_t *off, const prstatus_t *status, 11600ff27d31SJohn Polstra const prfpregset_t *fpregset, const prpsinfo_t *psinfo, int numsegs) 11618c64af4fSJohn Polstra { 11628c64af4fSJohn Polstra size_t ehoff; 11638c64af4fSJohn Polstra size_t phoff; 11648c64af4fSJohn Polstra size_t noteoff; 11658c64af4fSJohn Polstra size_t notesz; 11668c64af4fSJohn Polstra 11678c64af4fSJohn Polstra ehoff = *off; 11688c64af4fSJohn Polstra *off += sizeof(Elf_Ehdr); 11698c64af4fSJohn Polstra 11708c64af4fSJohn Polstra phoff = *off; 11710ff27d31SJohn Polstra *off += (numsegs + 1) * sizeof(Elf_Phdr); 11728c64af4fSJohn Polstra 11738c64af4fSJohn Polstra noteoff = *off; 11743ebc1248SPeter Wemm __elfN(putnote)(dst, off, "FreeBSD", NT_PRSTATUS, status, 11758c64af4fSJohn Polstra sizeof *status); 11763ebc1248SPeter Wemm __elfN(putnote)(dst, off, "FreeBSD", NT_FPREGSET, fpregset, 11778c64af4fSJohn Polstra sizeof *fpregset); 11783ebc1248SPeter Wemm __elfN(putnote)(dst, off, "FreeBSD", NT_PRPSINFO, psinfo, 11798c64af4fSJohn Polstra sizeof *psinfo); 11808c64af4fSJohn Polstra notesz = *off - noteoff; 11818c64af4fSJohn Polstra 11820ff27d31SJohn Polstra /* Align up to a page boundary for the program segments. */ 11838c64af4fSJohn Polstra *off = round_page(*off); 11848c64af4fSJohn Polstra 11858c64af4fSJohn Polstra if (dst != NULL) { 11868c64af4fSJohn Polstra Elf_Ehdr *ehdr; 11878c64af4fSJohn Polstra Elf_Phdr *phdr; 11880ff27d31SJohn Polstra struct phdr_closure phc; 11898c64af4fSJohn Polstra 11908c64af4fSJohn Polstra /* 11918c64af4fSJohn Polstra * Fill in the ELF header. 11928c64af4fSJohn Polstra */ 11938c64af4fSJohn Polstra ehdr = (Elf_Ehdr *)((char *)dst + ehoff); 11948c64af4fSJohn Polstra ehdr->e_ident[EI_MAG0] = ELFMAG0; 11958c64af4fSJohn Polstra ehdr->e_ident[EI_MAG1] = ELFMAG1; 11968c64af4fSJohn Polstra ehdr->e_ident[EI_MAG2] = ELFMAG2; 11978c64af4fSJohn Polstra ehdr->e_ident[EI_MAG3] = ELFMAG3; 11988c64af4fSJohn Polstra ehdr->e_ident[EI_CLASS] = ELF_CLASS; 11998c64af4fSJohn Polstra ehdr->e_ident[EI_DATA] = ELF_DATA; 12008c64af4fSJohn Polstra ehdr->e_ident[EI_VERSION] = EV_CURRENT; 1201c815a20cSDavid E. O'Brien ehdr->e_ident[EI_OSABI] = ELFOSABI_FREEBSD; 1202c815a20cSDavid E. O'Brien ehdr->e_ident[EI_ABIVERSION] = 0; 12038c64af4fSJohn Polstra ehdr->e_ident[EI_PAD] = 0; 12048c64af4fSJohn Polstra ehdr->e_type = ET_CORE; 12058c64af4fSJohn Polstra ehdr->e_machine = ELF_ARCH; 12068c64af4fSJohn Polstra ehdr->e_version = EV_CURRENT; 12078c64af4fSJohn Polstra ehdr->e_entry = 0; 12088c64af4fSJohn Polstra ehdr->e_phoff = phoff; 12098c64af4fSJohn Polstra ehdr->e_flags = 0; 12108c64af4fSJohn Polstra ehdr->e_ehsize = sizeof(Elf_Ehdr); 12118c64af4fSJohn Polstra ehdr->e_phentsize = sizeof(Elf_Phdr); 12120ff27d31SJohn Polstra ehdr->e_phnum = numsegs + 1; 12138c64af4fSJohn Polstra ehdr->e_shentsize = sizeof(Elf_Shdr); 12148c64af4fSJohn Polstra ehdr->e_shnum = 0; 12158c64af4fSJohn Polstra ehdr->e_shstrndx = SHN_UNDEF; 12168c64af4fSJohn Polstra 12178c64af4fSJohn Polstra /* 12188c64af4fSJohn Polstra * Fill in the program header entries. 12198c64af4fSJohn Polstra */ 12208c64af4fSJohn Polstra phdr = (Elf_Phdr *)((char *)dst + phoff); 12218c64af4fSJohn Polstra 12228c64af4fSJohn Polstra /* The note segement. */ 12238c64af4fSJohn Polstra phdr->p_type = PT_NOTE; 12248c64af4fSJohn Polstra phdr->p_offset = noteoff; 12258c64af4fSJohn Polstra phdr->p_vaddr = 0; 12268c64af4fSJohn Polstra phdr->p_paddr = 0; 12278c64af4fSJohn Polstra phdr->p_filesz = notesz; 12288c64af4fSJohn Polstra phdr->p_memsz = 0; 12298c64af4fSJohn Polstra phdr->p_flags = 0; 12308c64af4fSJohn Polstra phdr->p_align = 0; 12318c64af4fSJohn Polstra phdr++; 12328c64af4fSJohn Polstra 12330ff27d31SJohn Polstra /* All the writable segments from the program. */ 12340ff27d31SJohn Polstra phc.phdr = phdr; 12350ff27d31SJohn Polstra phc.offset = *off; 12360ff27d31SJohn Polstra each_writable_segment(p, cb_put_phdr, &phc); 12378c64af4fSJohn Polstra } 12388c64af4fSJohn Polstra } 12398c64af4fSJohn Polstra 12408c64af4fSJohn Polstra static void 12413ebc1248SPeter Wemm __elfN(putnote)(void *dst, size_t *off, const char *name, int type, 12428c64af4fSJohn Polstra const void *desc, size_t descsz) 12438c64af4fSJohn Polstra { 12448c64af4fSJohn Polstra Elf_Note note; 12458c64af4fSJohn Polstra 12468c64af4fSJohn Polstra note.n_namesz = strlen(name) + 1; 12478c64af4fSJohn Polstra note.n_descsz = descsz; 12488c64af4fSJohn Polstra note.n_type = type; 12498c64af4fSJohn Polstra if (dst != NULL) 12508c64af4fSJohn Polstra bcopy(¬e, (char *)dst + *off, sizeof note); 12518c64af4fSJohn Polstra *off += sizeof note; 12528c64af4fSJohn Polstra if (dst != NULL) 12538c64af4fSJohn Polstra bcopy(name, (char *)dst + *off, note.n_namesz); 12548c64af4fSJohn Polstra *off += roundup2(note.n_namesz, sizeof(Elf_Size)); 12558c64af4fSJohn Polstra if (dst != NULL) 12568c64af4fSJohn Polstra bcopy(desc, (char *)dst + *off, note.n_descsz); 12578c64af4fSJohn Polstra *off += roundup2(note.n_descsz, sizeof(Elf_Size)); 12588c64af4fSJohn Polstra } 12598c64af4fSJohn Polstra 12608c64af4fSJohn Polstra /* 1261e1743d02SSøren Schmidt * Tell kern_execve.c about it, with a little help from the linker. 1262e1743d02SSøren Schmidt */ 12633ebc1248SPeter Wemm #if __ELF_WORD_SIZE == 32 12643ebc1248SPeter Wemm static struct execsw elf_execsw = {exec_elf32_imgact, "ELF32"}; 12653ebc1248SPeter Wemm EXEC_SET(elf32, elf_execsw); 12663ebc1248SPeter Wemm #else 12673ebc1248SPeter Wemm static struct execsw elf_execsw = {exec_elf64_imgact, "ELF64"}; 12683ebc1248SPeter Wemm EXEC_SET(elf64, elf_execsw); 12693ebc1248SPeter Wemm #endif 1270