1 /*- 2 * Copyright (c) 1998 Michael Smith <msmith@freebsd.org> 3 * Copyright (c) 1998 Peter Wemm <peter@freebsd.org> 4 * All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions 8 * are met: 9 * 1. Redistributions of source code must retain the above copyright 10 * notice, this list of conditions and the following disclaimer. 11 * 2. Redistributions in binary form must reproduce the above copyright 12 * notice, this list of conditions and the following disclaimer in the 13 * documentation and/or other materials provided with the distribution. 14 * 15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 16 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 18 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 19 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 20 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 21 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 22 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 23 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 24 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 25 * SUCH DAMAGE. 26 */ 27 28 #include <sys/param.h> 29 #include <sys/endian.h> 30 #include <sys/exec.h> 31 #include <sys/linker.h> 32 #include <sys/module.h> 33 #include <sys/stdint.h> 34 #include <string.h> 35 #include <machine/elf.h> 36 #include <stand.h> 37 #include <sys/link_elf.h> 38 39 #include "bootstrap.h" 40 41 #define COPYOUT(s,d,l) archsw.arch_copyout((vm_offset_t)(s), d, l) 42 43 #if defined(__i386__) && __ELF_WORD_SIZE == 64 44 #undef ELF_TARG_CLASS 45 #undef ELF_TARG_MACH 46 #define ELF_TARG_CLASS ELFCLASS64 47 #define ELF_TARG_MACH EM_X86_64 48 #endif 49 50 typedef struct elf_file { 51 Elf_Phdr *ph; 52 Elf_Ehdr *ehdr; 53 Elf_Sym *symtab; 54 Elf_Hashelt *hashtab; 55 Elf_Hashelt nbuckets; 56 Elf_Hashelt nchains; 57 Elf_Hashelt *buckets; 58 Elf_Hashelt *chains; 59 Elf_Rel *rel; 60 size_t relsz; 61 Elf_Rela *rela; 62 size_t relasz; 63 char *strtab; 64 size_t strsz; 65 int fd; 66 caddr_t firstpage; 67 size_t firstlen; 68 int kernel; 69 uint64_t off; 70 #ifdef LOADER_VERIEXEC_VECTX 71 struct vectx *vctx; 72 #endif 73 } *elf_file_t; 74 75 #ifdef LOADER_VERIEXEC_VECTX 76 #define VECTX_HANDLE(ef) (ef)->vctx 77 #else 78 #define VECTX_HANDLE(ef) (ef)->fd 79 #endif 80 81 static int __elfN(loadimage)(struct preloaded_file *mp, elf_file_t ef, 82 uint64_t loadaddr); 83 static int __elfN(lookup_symbol)(elf_file_t ef, const char* name, 84 Elf_Sym *sym, unsigned char type); 85 static int __elfN(reloc_ptr)(struct preloaded_file *mp, elf_file_t ef, 86 Elf_Addr p, void *val, size_t len); 87 static int __elfN(parse_modmetadata)(struct preloaded_file *mp, elf_file_t ef, 88 Elf_Addr p_start, Elf_Addr p_end); 89 static symaddr_fn __elfN(symaddr); 90 static char *fake_modname(const char *name); 91 92 const char *__elfN(kerneltype) = "elf kernel"; 93 const char *__elfN(moduletype) = "elf module"; 94 95 uint64_t __elfN(relocation_offset) = 0; 96 97 #ifdef __powerpc__ 98 extern void elf_wrong_field_size(void); 99 #define CONVERT_FIELD(b, f, e) \ 100 switch (sizeof((b)->f)) { \ 101 case 2: \ 102 (b)->f = e ## 16toh((b)->f); \ 103 break; \ 104 case 4: \ 105 (b)->f = e ## 32toh((b)->f); \ 106 break; \ 107 case 8: \ 108 (b)->f = e ## 64toh((b)->f); \ 109 break; \ 110 default: \ 111 /* Force a link time error. */ \ 112 elf_wrong_field_size(); \ 113 break; \ 114 } 115 116 #define CONVERT_SWITCH(h, d, f) \ 117 switch ((h)->e_ident[EI_DATA]) { \ 118 case ELFDATA2MSB: \ 119 f(d, be); \ 120 break; \ 121 case ELFDATA2LSB: \ 122 f(d, le); \ 123 break; \ 124 default: \ 125 return (EINVAL); \ 126 } 127 128 129 static int elf_header_convert(Elf_Ehdr *ehdr) 130 { 131 /* 132 * Fixup ELF header endianness. 133 * 134 * The Xhdr structure was loaded using block read call to optimize file 135 * accesses. It might happen, that the endianness of the system memory 136 * is different that endianness of the ELF header. Swap fields here to 137 * guarantee that Xhdr always contain valid data regardless of 138 * architecture. 139 */ 140 #define HEADER_FIELDS(b, e) \ 141 CONVERT_FIELD(b, e_type, e); \ 142 CONVERT_FIELD(b, e_machine, e); \ 143 CONVERT_FIELD(b, e_version, e); \ 144 CONVERT_FIELD(b, e_entry, e); \ 145 CONVERT_FIELD(b, e_phoff, e); \ 146 CONVERT_FIELD(b, e_shoff, e); \ 147 CONVERT_FIELD(b, e_flags, e); \ 148 CONVERT_FIELD(b, e_ehsize, e); \ 149 CONVERT_FIELD(b, e_phentsize, e); \ 150 CONVERT_FIELD(b, e_phnum, e); \ 151 CONVERT_FIELD(b, e_shentsize, e); \ 152 CONVERT_FIELD(b, e_shnum, e); \ 153 CONVERT_FIELD(b, e_shstrndx, e) 154 155 CONVERT_SWITCH(ehdr, ehdr, HEADER_FIELDS); 156 157 #undef HEADER_FIELDS 158 159 return (0); 160 } 161 162 static int elf_program_header_convert(const Elf_Ehdr *ehdr, Elf_Phdr *phdr) 163 { 164 #define PROGRAM_HEADER_FIELDS(b, e) \ 165 CONVERT_FIELD(b, p_type, e); \ 166 CONVERT_FIELD(b, p_flags, e); \ 167 CONVERT_FIELD(b, p_offset, e); \ 168 CONVERT_FIELD(b, p_vaddr, e); \ 169 CONVERT_FIELD(b, p_paddr, e); \ 170 CONVERT_FIELD(b, p_filesz, e); \ 171 CONVERT_FIELD(b, p_memsz, e); \ 172 CONVERT_FIELD(b, p_align, e) 173 174 CONVERT_SWITCH(ehdr, phdr, PROGRAM_HEADER_FIELDS); 175 176 #undef PROGRAM_HEADER_FIELDS 177 178 return (0); 179 } 180 181 static int elf_section_header_convert(const Elf_Ehdr *ehdr, Elf_Shdr *shdr) 182 { 183 #define SECTION_HEADER_FIELDS(b, e) \ 184 CONVERT_FIELD(b, sh_name, e); \ 185 CONVERT_FIELD(b, sh_type, e); \ 186 CONVERT_FIELD(b, sh_link, e); \ 187 CONVERT_FIELD(b, sh_info, e); \ 188 CONVERT_FIELD(b, sh_flags, e); \ 189 CONVERT_FIELD(b, sh_addr, e); \ 190 CONVERT_FIELD(b, sh_offset, e); \ 191 CONVERT_FIELD(b, sh_size, e); \ 192 CONVERT_FIELD(b, sh_addralign, e); \ 193 CONVERT_FIELD(b, sh_entsize, e) 194 195 CONVERT_SWITCH(ehdr, shdr, SECTION_HEADER_FIELDS); 196 197 #undef SECTION_HEADER_FIELDS 198 199 return (0); 200 } 201 #undef CONVERT_SWITCH 202 #undef CONVERT_FIELD 203 #else 204 static int elf_header_convert(Elf_Ehdr *ehdr) 205 { 206 return (0); 207 } 208 209 static int elf_program_header_convert(const Elf_Ehdr *ehdr, Elf_Phdr *phdr) 210 { 211 return (0); 212 } 213 214 static int elf_section_header_convert(const Elf_Ehdr *ehdr, Elf_Shdr *shdr) 215 { 216 return (0); 217 } 218 #endif 219 220 #ifdef __amd64__ 221 static bool 222 is_kernphys_relocatable(elf_file_t ef) 223 { 224 Elf_Sym sym; 225 226 return (__elfN(lookup_symbol)(ef, "kernphys", &sym, STT_OBJECT) == 0); 227 } 228 #endif 229 230 #ifdef __i386__ 231 static bool 232 is_tg_kernel_support(struct preloaded_file *fp, elf_file_t ef) 233 { 234 Elf_Sym sym; 235 Elf_Addr p_start, p_end, v, p; 236 char vd_name[16]; 237 int error; 238 239 if (__elfN(lookup_symbol)(ef, "__start_set_vt_drv_set", &sym, STT_NOTYPE) != 0) 240 return (false); 241 p_start = sym.st_value + ef->off; 242 if (__elfN(lookup_symbol)(ef, "__stop_set_vt_drv_set", &sym, STT_NOTYPE) != 0) 243 return (false); 244 p_end = sym.st_value + ef->off; 245 246 /* 247 * Walk through vt_drv_set, each vt driver structure starts with 248 * static 16 chars for driver name. If we have "vbefb", return true. 249 */ 250 for (p = p_start; p < p_end; p += sizeof(Elf_Addr)) { 251 COPYOUT(p, &v, sizeof(v)); 252 253 error = __elfN(reloc_ptr)(fp, ef, p, &v, sizeof(v)); 254 if (error == EOPNOTSUPP) 255 v += ef->off; 256 else if (error != 0) 257 return (false); 258 COPYOUT(v, &vd_name, sizeof(vd_name)); 259 if (strncmp(vd_name, "vbefb", sizeof(vd_name)) == 0) 260 return (true); 261 } 262 263 return (false); 264 } 265 #endif 266 267 static int 268 __elfN(load_elf_header)(char *filename, elf_file_t ef) 269 { 270 ssize_t bytes_read; 271 Elf_Ehdr *ehdr; 272 int err; 273 274 /* 275 * Open the image, read and validate the ELF header 276 */ 277 if (filename == NULL) /* can't handle nameless */ 278 return (EFTYPE); 279 if ((ef->fd = open(filename, O_RDONLY)) == -1) 280 return (errno); 281 ef->firstpage = malloc(PAGE_SIZE); 282 if (ef->firstpage == NULL) { 283 close(ef->fd); 284 return (ENOMEM); 285 } 286 preload(ef->fd); 287 #ifdef LOADER_VERIEXEC_VECTX 288 { 289 int verror; 290 291 ef->vctx = vectx_open(ef->fd, filename, 0L, NULL, &verror, __func__); 292 if (verror) { 293 printf("Unverified %s: %s\n", filename, ve_error_get()); 294 close(ef->fd); 295 free(ef->vctx); 296 return (EAUTH); 297 } 298 } 299 #endif 300 bytes_read = VECTX_READ(VECTX_HANDLE(ef), ef->firstpage, PAGE_SIZE); 301 ef->firstlen = (size_t)bytes_read; 302 if (bytes_read < 0 || ef->firstlen <= sizeof(Elf_Ehdr)) { 303 err = EFTYPE; /* could be EIO, but may be small file */ 304 goto error; 305 } 306 ehdr = ef->ehdr = (Elf_Ehdr *)ef->firstpage; 307 308 /* Is it ELF? */ 309 if (!IS_ELF(*ehdr)) { 310 err = EFTYPE; 311 goto error; 312 } 313 314 if (ehdr->e_ident[EI_CLASS] != ELF_TARG_CLASS || /* Layout ? */ 315 ehdr->e_ident[EI_DATA] != ELF_TARG_DATA || 316 ehdr->e_ident[EI_VERSION] != EV_CURRENT) /* Version ? */ { 317 err = EFTYPE; 318 goto error; 319 } 320 321 err = elf_header_convert(ehdr); 322 if (err) 323 goto error; 324 325 if (ehdr->e_version != EV_CURRENT || ehdr->e_machine != ELF_TARG_MACH) { 326 /* Machine ? */ 327 err = EFTYPE; 328 goto error; 329 } 330 331 #if defined(LOADER_VERIEXEC) && !defined(LOADER_VERIEXEC_VECTX) 332 if (verify_file(ef->fd, filename, bytes_read, VE_MUST, __func__) < 0) { 333 err = EAUTH; 334 goto error; 335 } 336 #endif 337 return (0); 338 339 error: 340 if (ef->firstpage != NULL) { 341 free(ef->firstpage); 342 ef->firstpage = NULL; 343 } 344 if (ef->fd != -1) { 345 #ifdef LOADER_VERIEXEC_VECTX 346 free(ef->vctx); 347 #endif 348 close(ef->fd); 349 ef->fd = -1; 350 } 351 return (err); 352 } 353 354 /* 355 * Attempt to load the file (file) as an ELF module. It will be stored at 356 * (dest), and a pointer to a module structure describing the loaded object 357 * will be saved in (result). 358 */ 359 int 360 __elfN(loadfile)(char *filename, uint64_t dest, struct preloaded_file **result) 361 { 362 return (__elfN(loadfile_raw)(filename, dest, result, 0)); 363 } 364 365 int 366 __elfN(loadfile_raw)(char *filename, uint64_t dest, 367 struct preloaded_file **result, int multiboot) 368 { 369 struct preloaded_file *fp, *kfp; 370 struct elf_file ef; 371 Elf_Ehdr *ehdr; 372 int err; 373 374 fp = NULL; 375 bzero(&ef, sizeof(struct elf_file)); 376 ef.fd = -1; 377 378 err = __elfN(load_elf_header)(filename, &ef); 379 if (err != 0) 380 return (err); 381 382 ehdr = ef.ehdr; 383 384 /* 385 * Check to see what sort of module we are. 386 */ 387 kfp = file_findfile(NULL, __elfN(kerneltype)); 388 #ifdef __powerpc__ 389 /* 390 * Kernels can be ET_DYN, so just assume the first loaded object is the 391 * kernel. This assumption will be checked later. 392 */ 393 if (kfp == NULL) 394 ef.kernel = 1; 395 #endif 396 if (ef.kernel || ehdr->e_type == ET_EXEC) { 397 /* Looks like a kernel */ 398 if (kfp != NULL) { 399 printf("elf" __XSTRING(__ELF_WORD_SIZE) 400 "_loadfile: kernel already loaded\n"); 401 err = EPERM; 402 goto oerr; 403 } 404 /* 405 * Calculate destination address based on kernel entrypoint. 406 * 407 * For ARM, the destination address is independent of any values 408 * in the elf header (an ARM kernel can be loaded at any 2MB 409 * boundary), so we leave dest set to the value calculated by 410 * archsw.arch_loadaddr() and passed in to this function. 411 */ 412 #ifndef __arm__ 413 if (ehdr->e_type == ET_EXEC) 414 dest = (ehdr->e_entry & ~PAGE_MASK); 415 #endif 416 if ((ehdr->e_entry & ~PAGE_MASK) == 0) { 417 printf("elf" __XSTRING(__ELF_WORD_SIZE) 418 "_loadfile: not a kernel (maybe static binary?)\n"); 419 err = EPERM; 420 goto oerr; 421 } 422 ef.kernel = 1; 423 424 } else if (ehdr->e_type == ET_DYN) { 425 /* Looks like a kld module */ 426 if (multiboot != 0) { 427 printf("elf" __XSTRING(__ELF_WORD_SIZE) 428 "_loadfile: can't load module as multiboot\n"); 429 err = EPERM; 430 goto oerr; 431 } 432 if (kfp == NULL) { 433 printf("elf" __XSTRING(__ELF_WORD_SIZE) 434 "_loadfile: can't load module before kernel\n"); 435 err = EPERM; 436 goto oerr; 437 } 438 if (strcmp(__elfN(kerneltype), kfp->f_type)) { 439 printf("elf" __XSTRING(__ELF_WORD_SIZE) 440 "_loadfile: can't load module with kernel type '%s'\n", 441 kfp->f_type); 442 err = EPERM; 443 goto oerr; 444 } 445 /* Looks OK, got ahead */ 446 ef.kernel = 0; 447 448 } else { 449 err = EFTYPE; 450 goto oerr; 451 } 452 453 if (archsw.arch_loadaddr != NULL) 454 dest = archsw.arch_loadaddr(LOAD_ELF, ehdr, dest); 455 else 456 dest = roundup(dest, PAGE_SIZE); 457 458 /* 459 * Ok, we think we should handle this. 460 */ 461 fp = file_alloc(); 462 if (fp == NULL) { 463 printf("elf" __XSTRING(__ELF_WORD_SIZE) 464 "_loadfile: cannot allocate module info\n"); 465 err = EPERM; 466 goto out; 467 } 468 if (ef.kernel == 1 && multiboot == 0) 469 setenv("kernelname", filename, 1); 470 fp->f_name = strdup(filename); 471 if (multiboot == 0) 472 fp->f_type = strdup(ef.kernel ? 473 __elfN(kerneltype) : __elfN(moduletype)); 474 else 475 fp->f_type = strdup("elf multiboot kernel"); 476 477 if (module_verbose >= MODULE_VERBOSE_FULL) { 478 if (ef.kernel) 479 printf("%s entry at 0x%jx\n", filename, 480 (uintmax_t)ehdr->e_entry); 481 } else if (module_verbose > MODULE_VERBOSE_SILENT) 482 printf("%s ", filename); 483 484 fp->f_size = __elfN(loadimage)(fp, &ef, dest); 485 if (fp->f_size == 0 || fp->f_addr == 0) 486 goto ioerr; 487 488 /* save exec header as metadata */ 489 file_addmetadata(fp, MODINFOMD_ELFHDR, sizeof(*ehdr), ehdr); 490 491 /* Load OK, return module pointer */ 492 *result = (struct preloaded_file *)fp; 493 err = 0; 494 #ifdef __amd64__ 495 fp->f_kernphys_relocatable = multiboot || is_kernphys_relocatable(&ef); 496 #endif 497 #ifdef __i386__ 498 fp->f_tg_kernel_support = is_tg_kernel_support(fp, &ef); 499 #endif 500 goto out; 501 502 ioerr: 503 err = EIO; 504 oerr: 505 file_discard(fp); 506 out: 507 if (ef.firstpage) 508 free(ef.firstpage); 509 if (ef.fd != -1) { 510 #ifdef LOADER_VERIEXEC_VECTX 511 if (!err && ef.vctx) { 512 int verror; 513 514 verror = vectx_close(ef.vctx, VE_MUST, __func__); 515 if (verror) { 516 err = EAUTH; 517 file_discard(fp); 518 } 519 } 520 #endif 521 close(ef.fd); 522 } 523 return (err); 524 } 525 526 /* 527 * With the file (fd) open on the image, and (ehdr) containing 528 * the Elf header, load the image at (off) 529 */ 530 static int 531 __elfN(loadimage)(struct preloaded_file *fp, elf_file_t ef, uint64_t off) 532 { 533 int i; 534 u_int j; 535 Elf_Ehdr *ehdr; 536 Elf_Phdr *phdr, *php; 537 Elf_Shdr *shdr; 538 char *shstr; 539 int ret; 540 vm_offset_t firstaddr; 541 vm_offset_t lastaddr; 542 size_t chunk; 543 ssize_t result; 544 Elf_Addr ssym, esym; 545 Elf_Dyn *dp; 546 Elf_Addr adp; 547 Elf_Addr ctors; 548 int ndp; 549 int symstrindex; 550 int symtabindex; 551 Elf_Size size; 552 u_int fpcopy; 553 Elf_Sym sym; 554 Elf_Addr p_start, p_end; 555 556 dp = NULL; 557 shdr = NULL; 558 ret = 0; 559 firstaddr = lastaddr = 0; 560 ehdr = ef->ehdr; 561 #ifdef __powerpc__ 562 if (ef->kernel) { 563 #else 564 if (ehdr->e_type == ET_EXEC) { 565 #endif 566 #if defined(__i386__) || defined(__amd64__) 567 #if __ELF_WORD_SIZE == 64 568 /* x86_64 relocates after locore */ 569 off = - (off & 0xffffffffff000000ull); 570 #else 571 /* i386 relocates after locore */ 572 off = - (off & 0xff000000u); 573 #endif 574 #elif defined(__powerpc__) 575 /* 576 * On the purely virtual memory machines like e500, the kernel 577 * is linked against its final VA range, which is most often 578 * not available at the loader stage, but only after kernel 579 * initializes and completes its VM settings. In such cases we 580 * cannot use p_vaddr field directly to load ELF segments, but 581 * put them at some 'load-time' locations. 582 */ 583 if (off & 0xf0000000u) { 584 off = -(off & 0xf0000000u); 585 /* 586 * XXX the physical load address should not be 587 * hardcoded. Note that the Book-E kernel assumes that 588 * it's loaded at a 16MB boundary for now... 589 */ 590 off += 0x01000000; 591 } 592 ehdr->e_entry += off; 593 if (module_verbose >= MODULE_VERBOSE_FULL) 594 printf("Converted entry 0x%jx\n", 595 (uintmax_t)ehdr->e_entry); 596 597 #elif defined(__arm__) && !defined(EFI) 598 /* 599 * The elf headers in arm kernels specify virtual addresses in 600 * all header fields, even the ones that should be physical 601 * addresses. We assume the entry point is in the first page, 602 * and masking the page offset will leave us with the virtual 603 * address the kernel was linked at. We subtract that from the 604 * load offset, making 'off' into the value which, when added 605 * to a virtual address in an elf header, translates it to a 606 * physical address. We do the va->pa conversion on the entry 607 * point address in the header now, so that later we can launch 608 * the kernel by just jumping to that address. 609 * 610 * When booting from UEFI the copyin and copyout functions 611 * handle adjusting the location relative to the first virtual 612 * address. Because of this there is no need to adjust the 613 * offset or entry point address as these will both be handled 614 * by the efi code. 615 */ 616 off -= ehdr->e_entry & ~PAGE_MASK; 617 ehdr->e_entry += off; 618 if (module_verbose >= MODULE_VERBOSE_FULL) 619 printf("ehdr->e_entry 0x%jx, va<->pa off %llx\n", 620 (uintmax_t)ehdr->e_entry, off); 621 #else 622 off = 0; /* other archs use direct mapped kernels */ 623 #endif 624 } 625 ef->off = off; 626 627 if (ef->kernel) 628 __elfN(relocation_offset) = off; 629 630 if ((ehdr->e_phoff + ehdr->e_phnum * sizeof(*phdr)) > ef->firstlen) { 631 printf("elf" __XSTRING(__ELF_WORD_SIZE) 632 "_loadimage: program header not within first page\n"); 633 goto out; 634 } 635 phdr = (Elf_Phdr *)(ef->firstpage + ehdr->e_phoff); 636 637 for (i = 0; i < ehdr->e_phnum; i++) { 638 if (elf_program_header_convert(ehdr, phdr)) 639 continue; 640 641 /* We want to load PT_LOAD segments only.. */ 642 if (phdr[i].p_type != PT_LOAD) 643 continue; 644 645 if (module_verbose >= MODULE_VERBOSE_FULL) { 646 printf("Segment: 0x%lx@0x%lx -> 0x%lx-0x%lx", 647 (long)phdr[i].p_filesz, (long)phdr[i].p_offset, 648 (long)(phdr[i].p_vaddr + off), 649 (long)(phdr[i].p_vaddr + off + phdr[i].p_memsz - 1)); 650 } else if (module_verbose > MODULE_VERBOSE_SILENT) { 651 if ((phdr[i].p_flags & PF_W) == 0) { 652 printf("text=0x%lx ", (long)phdr[i].p_filesz); 653 } else { 654 printf("data=0x%lx", (long)phdr[i].p_filesz); 655 if (phdr[i].p_filesz < phdr[i].p_memsz) 656 printf("+0x%lx", (long)(phdr[i].p_memsz - 657 phdr[i].p_filesz)); 658 printf(" "); 659 } 660 } 661 fpcopy = 0; 662 if (ef->firstlen > phdr[i].p_offset) { 663 fpcopy = ef->firstlen - phdr[i].p_offset; 664 archsw.arch_copyin(ef->firstpage + phdr[i].p_offset, 665 phdr[i].p_vaddr + off, fpcopy); 666 } 667 if (phdr[i].p_filesz > fpcopy) { 668 if (kern_pread(VECTX_HANDLE(ef), 669 phdr[i].p_vaddr + off + fpcopy, 670 phdr[i].p_filesz - fpcopy, 671 phdr[i].p_offset + fpcopy) != 0) { 672 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 673 "_loadimage: read failed\n"); 674 goto out; 675 } 676 } 677 /* clear space from oversized segments; eg: bss */ 678 if (phdr[i].p_filesz < phdr[i].p_memsz) { 679 if (module_verbose >= MODULE_VERBOSE_FULL) { 680 printf(" (bss: 0x%lx-0x%lx)", 681 (long)(phdr[i].p_vaddr + off + phdr[i].p_filesz), 682 (long)(phdr[i].p_vaddr + off + phdr[i].p_memsz -1)); 683 } 684 kern_bzero(phdr[i].p_vaddr + off + phdr[i].p_filesz, 685 phdr[i].p_memsz - phdr[i].p_filesz); 686 } 687 if (module_verbose >= MODULE_VERBOSE_FULL) 688 printf("\n"); 689 690 if (archsw.arch_loadseg != NULL) 691 archsw.arch_loadseg(ehdr, phdr + i, off); 692 693 if (firstaddr == 0 || firstaddr > (phdr[i].p_vaddr + off)) 694 firstaddr = phdr[i].p_vaddr + off; 695 if (lastaddr == 0 || lastaddr < 696 (phdr[i].p_vaddr + off + phdr[i].p_memsz)) 697 lastaddr = phdr[i].p_vaddr + off + phdr[i].p_memsz; 698 } 699 lastaddr = roundup(lastaddr, sizeof(long)); 700 701 /* 702 * Get the section headers. We need this for finding the .ctors 703 * section as well as for loading any symbols. Both may be hard 704 * to do if reading from a .gz file as it involves seeking. I 705 * think the rule is going to have to be that you must strip a 706 * file to remove symbols before gzipping it. 707 */ 708 chunk = (size_t)ehdr->e_shnum * (size_t)ehdr->e_shentsize; 709 if (chunk == 0 || ehdr->e_shoff == 0) 710 goto nosyms; 711 shdr = alloc_pread(VECTX_HANDLE(ef), ehdr->e_shoff, chunk); 712 if (shdr == NULL) { 713 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 714 "_loadimage: failed to read section headers"); 715 goto nosyms; 716 } 717 718 for (i = 0; i < ehdr->e_shnum; i++) 719 elf_section_header_convert(ehdr, &shdr[i]); 720 721 file_addmetadata(fp, MODINFOMD_SHDR, chunk, shdr); 722 723 /* 724 * Read the section string table and look for the .ctors section. 725 * We need to tell the kernel where it is so that it can call the 726 * ctors. 727 */ 728 chunk = shdr[ehdr->e_shstrndx].sh_size; 729 if (chunk) { 730 shstr = alloc_pread(VECTX_HANDLE(ef), 731 shdr[ehdr->e_shstrndx].sh_offset, chunk); 732 if (shstr) { 733 for (i = 0; i < ehdr->e_shnum; i++) { 734 if (strcmp(shstr + shdr[i].sh_name, 735 ".ctors") != 0) 736 continue; 737 ctors = shdr[i].sh_addr; 738 file_addmetadata(fp, MODINFOMD_CTORS_ADDR, 739 sizeof(ctors), &ctors); 740 size = shdr[i].sh_size; 741 file_addmetadata(fp, MODINFOMD_CTORS_SIZE, 742 sizeof(size), &size); 743 break; 744 } 745 free(shstr); 746 } 747 } 748 749 /* 750 * Now load any symbols. 751 */ 752 symtabindex = -1; 753 symstrindex = -1; 754 for (i = 0; i < ehdr->e_shnum; i++) { 755 if (shdr[i].sh_type != SHT_SYMTAB) 756 continue; 757 for (j = 0; j < ehdr->e_phnum; j++) { 758 if (phdr[j].p_type != PT_LOAD) 759 continue; 760 if (shdr[i].sh_offset >= phdr[j].p_offset && 761 (shdr[i].sh_offset + shdr[i].sh_size <= 762 phdr[j].p_offset + phdr[j].p_filesz)) { 763 shdr[i].sh_offset = 0; 764 shdr[i].sh_size = 0; 765 break; 766 } 767 } 768 if (shdr[i].sh_offset == 0 || shdr[i].sh_size == 0) 769 continue; /* alread loaded in a PT_LOAD above */ 770 /* Save it for loading below */ 771 symtabindex = i; 772 symstrindex = shdr[i].sh_link; 773 } 774 if (symtabindex < 0 || symstrindex < 0) 775 goto nosyms; 776 777 /* Ok, committed to a load. */ 778 if (module_verbose >= MODULE_VERBOSE_FULL) 779 printf("syms=["); 780 ssym = lastaddr; 781 for (i = symtabindex; i >= 0; i = symstrindex) { 782 char *secname; 783 784 switch(shdr[i].sh_type) { 785 case SHT_SYMTAB: /* Symbol table */ 786 secname = "symtab"; 787 break; 788 case SHT_STRTAB: /* String table */ 789 secname = "strtab"; 790 break; 791 default: 792 secname = "WHOA!!"; 793 break; 794 } 795 size = shdr[i].sh_size; 796 797 archsw.arch_copyin(&size, lastaddr, sizeof(size)); 798 lastaddr += sizeof(size); 799 800 if (module_verbose >= MODULE_VERBOSE_FULL) { 801 printf("\n%s: 0x%jx@0x%jx -> 0x%jx-0x%jx", secname, 802 (uintmax_t)shdr[i].sh_size, (uintmax_t)shdr[i].sh_offset, 803 (uintmax_t)lastaddr, 804 (uintmax_t)(lastaddr + shdr[i].sh_size)); 805 } else if (module_verbose > MODULE_VERBOSE_SILENT) { 806 if (i == symstrindex) 807 printf("+"); 808 printf("0x%lx+0x%lx", (long)sizeof(size), (long)size); 809 } 810 if (VECTX_LSEEK(VECTX_HANDLE(ef), (off_t)shdr[i].sh_offset, SEEK_SET) == -1) { 811 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 812 "_loadimage: could not seek for symbols - skipped!"); 813 lastaddr = ssym; 814 ssym = 0; 815 goto nosyms; 816 } 817 result = archsw.arch_readin(VECTX_HANDLE(ef), lastaddr, shdr[i].sh_size); 818 if (result < 0 || (size_t)result != shdr[i].sh_size) { 819 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 820 "_loadimage: could not read symbols - skipped! " 821 "(%ju != %ju)", (uintmax_t)result, 822 (uintmax_t)shdr[i].sh_size); 823 lastaddr = ssym; 824 ssym = 0; 825 goto nosyms; 826 } 827 /* Reset offsets relative to ssym */ 828 lastaddr += shdr[i].sh_size; 829 lastaddr = roundup(lastaddr, sizeof(size)); 830 if (i == symtabindex) 831 symtabindex = -1; 832 else if (i == symstrindex) 833 symstrindex = -1; 834 } 835 esym = lastaddr; 836 if (module_verbose >= MODULE_VERBOSE_FULL) 837 printf("]"); 838 839 file_addmetadata(fp, MODINFOMD_SSYM, sizeof(ssym), &ssym); 840 file_addmetadata(fp, MODINFOMD_ESYM, sizeof(esym), &esym); 841 842 nosyms: 843 if (module_verbose > MODULE_VERBOSE_SILENT) 844 printf("\n"); 845 846 ret = lastaddr - firstaddr; 847 fp->f_addr = firstaddr; 848 849 php = NULL; 850 for (i = 0; i < ehdr->e_phnum; i++) { 851 if (phdr[i].p_type == PT_DYNAMIC) { 852 php = phdr + i; 853 adp = php->p_vaddr; 854 file_addmetadata(fp, MODINFOMD_DYNAMIC, sizeof(adp), 855 &adp); 856 break; 857 } 858 } 859 860 if (php == NULL) /* this is bad, we cannot get to symbols or _DYNAMIC */ 861 goto out; 862 863 ndp = php->p_filesz / sizeof(Elf_Dyn); 864 if (ndp == 0) 865 goto out; 866 dp = malloc(php->p_filesz); 867 if (dp == NULL) 868 goto out; 869 archsw.arch_copyout(php->p_vaddr + off, dp, php->p_filesz); 870 871 ef->strsz = 0; 872 for (i = 0; i < ndp; i++) { 873 if (dp[i].d_tag == 0) 874 break; 875 switch (dp[i].d_tag) { 876 case DT_HASH: 877 ef->hashtab = 878 (Elf_Hashelt*)(uintptr_t)(dp[i].d_un.d_ptr + off); 879 break; 880 case DT_STRTAB: 881 ef->strtab = 882 (char *)(uintptr_t)(dp[i].d_un.d_ptr + off); 883 break; 884 case DT_STRSZ: 885 ef->strsz = dp[i].d_un.d_val; 886 break; 887 case DT_SYMTAB: 888 ef->symtab = 889 (Elf_Sym *)(uintptr_t)(dp[i].d_un.d_ptr + off); 890 break; 891 case DT_REL: 892 ef->rel = 893 (Elf_Rel *)(uintptr_t)(dp[i].d_un.d_ptr + off); 894 break; 895 case DT_RELSZ: 896 ef->relsz = dp[i].d_un.d_val; 897 break; 898 case DT_RELA: 899 ef->rela = 900 (Elf_Rela *)(uintptr_t)(dp[i].d_un.d_ptr + off); 901 break; 902 case DT_RELASZ: 903 ef->relasz = dp[i].d_un.d_val; 904 break; 905 default: 906 break; 907 } 908 } 909 if (ef->hashtab == NULL || ef->symtab == NULL || 910 ef->strtab == NULL || ef->strsz == 0) 911 goto out; 912 COPYOUT(ef->hashtab, &ef->nbuckets, sizeof(ef->nbuckets)); 913 COPYOUT(ef->hashtab + 1, &ef->nchains, sizeof(ef->nchains)); 914 ef->buckets = ef->hashtab + 2; 915 ef->chains = ef->buckets + ef->nbuckets; 916 917 if (__elfN(lookup_symbol)(ef, "__start_set_modmetadata_set", &sym, 918 STT_NOTYPE) != 0) 919 return 0; 920 p_start = sym.st_value + ef->off; 921 if (__elfN(lookup_symbol)(ef, "__stop_set_modmetadata_set", &sym, 922 STT_NOTYPE) != 0) 923 return 0; 924 p_end = sym.st_value + ef->off; 925 926 if (__elfN(parse_modmetadata)(fp, ef, p_start, p_end) == 0) 927 goto out; 928 929 if (ef->kernel) /* kernel must not depend on anything */ 930 goto out; 931 932 out: 933 if (dp) 934 free(dp); 935 if (shdr) 936 free(shdr); 937 return ret; 938 } 939 940 static char invalid_name[] = "bad"; 941 942 char * 943 fake_modname(const char *name) 944 { 945 const char *sp, *ep; 946 char *fp; 947 size_t len; 948 949 sp = strrchr(name, '/'); 950 if (sp) 951 sp++; 952 else 953 sp = name; 954 955 ep = strrchr(sp, '.'); 956 if (ep == NULL) { 957 ep = sp + strlen(sp); 958 } 959 if (ep == sp) { 960 sp = invalid_name; 961 ep = invalid_name + sizeof(invalid_name) - 1; 962 } 963 964 len = ep - sp; 965 fp = malloc(len + 1); 966 if (fp == NULL) 967 return NULL; 968 memcpy(fp, sp, len); 969 fp[len] = '\0'; 970 return fp; 971 } 972 973 #if (defined(__i386__) || defined(__powerpc__)) && __ELF_WORD_SIZE == 64 974 struct mod_metadata64 { 975 int md_version; /* structure version MDTV_* */ 976 int md_type; /* type of entry MDT_* */ 977 uint64_t md_data; /* specific data */ 978 uint64_t md_cval; /* common string label */ 979 }; 980 #endif 981 #if defined(__amd64__) && __ELF_WORD_SIZE == 32 982 struct mod_metadata32 { 983 int md_version; /* structure version MDTV_* */ 984 int md_type; /* type of entry MDT_* */ 985 uint32_t md_data; /* specific data */ 986 uint32_t md_cval; /* common string label */ 987 }; 988 #endif 989 990 int 991 __elfN(load_modmetadata)(struct preloaded_file *fp, uint64_t dest) 992 { 993 struct elf_file ef; 994 int err, i, j; 995 Elf_Shdr *sh_meta, *shdr = NULL; 996 Elf_Shdr *sh_data[2]; 997 char *shstrtab = NULL; 998 size_t size; 999 Elf_Addr p_start, p_end; 1000 1001 bzero(&ef, sizeof(struct elf_file)); 1002 ef.fd = -1; 1003 1004 err = __elfN(load_elf_header)(fp->f_name, &ef); 1005 if (err != 0) 1006 goto out; 1007 1008 if (ef.kernel == 1 || ef.ehdr->e_type == ET_EXEC) { 1009 ef.kernel = 1; 1010 } else if (ef.ehdr->e_type != ET_DYN) { 1011 err = EFTYPE; 1012 goto out; 1013 } 1014 1015 size = (size_t)ef.ehdr->e_shnum * (size_t)ef.ehdr->e_shentsize; 1016 shdr = alloc_pread(VECTX_HANDLE(&ef), ef.ehdr->e_shoff, size); 1017 if (shdr == NULL) { 1018 err = ENOMEM; 1019 goto out; 1020 } 1021 1022 /* Load shstrtab. */ 1023 shstrtab = alloc_pread(VECTX_HANDLE(&ef), shdr[ef.ehdr->e_shstrndx].sh_offset, 1024 shdr[ef.ehdr->e_shstrndx].sh_size); 1025 if (shstrtab == NULL) { 1026 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 1027 "load_modmetadata: unable to load shstrtab\n"); 1028 err = EFTYPE; 1029 goto out; 1030 } 1031 1032 /* Find set_modmetadata_set and data sections. */ 1033 sh_data[0] = sh_data[1] = sh_meta = NULL; 1034 for (i = 0, j = 0; i < ef.ehdr->e_shnum; i++) { 1035 if (strcmp(&shstrtab[shdr[i].sh_name], 1036 "set_modmetadata_set") == 0) { 1037 sh_meta = &shdr[i]; 1038 } 1039 if ((strcmp(&shstrtab[shdr[i].sh_name], ".data") == 0) || 1040 (strcmp(&shstrtab[shdr[i].sh_name], ".rodata") == 0)) { 1041 sh_data[j++] = &shdr[i]; 1042 } 1043 } 1044 if (sh_meta == NULL || sh_data[0] == NULL || sh_data[1] == NULL) { 1045 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 1046 "load_modmetadata: unable to find set_modmetadata_set or data sections\n"); 1047 err = EFTYPE; 1048 goto out; 1049 } 1050 1051 /* Load set_modmetadata_set into memory */ 1052 err = kern_pread(VECTX_HANDLE(&ef), dest, sh_meta->sh_size, sh_meta->sh_offset); 1053 if (err != 0) { 1054 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 1055 "load_modmetadata: unable to load set_modmetadata_set: %d\n", err); 1056 goto out; 1057 } 1058 p_start = dest; 1059 p_end = dest + sh_meta->sh_size; 1060 dest += sh_meta->sh_size; 1061 1062 /* Load data sections into memory. */ 1063 err = kern_pread(VECTX_HANDLE(&ef), dest, sh_data[0]->sh_size, 1064 sh_data[0]->sh_offset); 1065 if (err != 0) { 1066 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 1067 "load_modmetadata: unable to load data: %d\n", err); 1068 goto out; 1069 } 1070 1071 /* 1072 * We have to increment the dest, so that the offset is the same into 1073 * both the .rodata and .data sections. 1074 */ 1075 ef.off = -(sh_data[0]->sh_addr - dest); 1076 dest += (sh_data[1]->sh_addr - sh_data[0]->sh_addr); 1077 1078 err = kern_pread(VECTX_HANDLE(&ef), dest, sh_data[1]->sh_size, 1079 sh_data[1]->sh_offset); 1080 if (err != 0) { 1081 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 1082 "load_modmetadata: unable to load data: %d\n", err); 1083 goto out; 1084 } 1085 1086 err = __elfN(parse_modmetadata)(fp, &ef, p_start, p_end); 1087 if (err != 0) { 1088 printf("\nelf" __XSTRING(__ELF_WORD_SIZE) 1089 "load_modmetadata: unable to parse metadata: %d\n", err); 1090 goto out; 1091 } 1092 1093 out: 1094 if (shstrtab != NULL) 1095 free(shstrtab); 1096 if (shdr != NULL) 1097 free(shdr); 1098 if (ef.firstpage != NULL) 1099 free(ef.firstpage); 1100 if (ef.fd != -1) { 1101 #ifdef LOADER_VERIEXEC_VECTX 1102 if (!err && ef.vctx) { 1103 int verror; 1104 1105 verror = vectx_close(ef.vctx, VE_MUST, __func__); 1106 if (verror) { 1107 err = EAUTH; 1108 file_discard(fp); 1109 } 1110 } 1111 #endif 1112 close(ef.fd); 1113 } 1114 return (err); 1115 } 1116 1117 int 1118 __elfN(parse_modmetadata)(struct preloaded_file *fp, elf_file_t ef, 1119 Elf_Addr p_start, Elf_Addr p_end) 1120 { 1121 struct mod_metadata md; 1122 #if (defined(__i386__) || defined(__powerpc__)) && __ELF_WORD_SIZE == 64 1123 struct mod_metadata64 md64; 1124 #elif defined(__amd64__) && __ELF_WORD_SIZE == 32 1125 struct mod_metadata32 md32; 1126 #endif 1127 struct mod_depend *mdepend; 1128 struct mod_version mver; 1129 char *s; 1130 int error, modcnt, minfolen; 1131 Elf_Addr v, p; 1132 1133 modcnt = 0; 1134 p = p_start; 1135 while (p < p_end) { 1136 COPYOUT(p, &v, sizeof(v)); 1137 error = __elfN(reloc_ptr)(fp, ef, p, &v, sizeof(v)); 1138 if (error == EOPNOTSUPP) 1139 v += ef->off; 1140 else if (error != 0) 1141 return (error); 1142 #if (defined(__i386__) || defined(__powerpc__)) && __ELF_WORD_SIZE == 64 1143 COPYOUT(v, &md64, sizeof(md64)); 1144 error = __elfN(reloc_ptr)(fp, ef, v, &md64, sizeof(md64)); 1145 if (error == EOPNOTSUPP) { 1146 md64.md_cval += ef->off; 1147 md64.md_data += ef->off; 1148 } else if (error != 0) 1149 return (error); 1150 md.md_version = md64.md_version; 1151 md.md_type = md64.md_type; 1152 md.md_cval = (const char *)(uintptr_t)md64.md_cval; 1153 md.md_data = (void *)(uintptr_t)md64.md_data; 1154 #elif defined(__amd64__) && __ELF_WORD_SIZE == 32 1155 COPYOUT(v, &md32, sizeof(md32)); 1156 error = __elfN(reloc_ptr)(fp, ef, v, &md32, sizeof(md32)); 1157 if (error == EOPNOTSUPP) { 1158 md32.md_cval += ef->off; 1159 md32.md_data += ef->off; 1160 } else if (error != 0) 1161 return (error); 1162 md.md_version = md32.md_version; 1163 md.md_type = md32.md_type; 1164 md.md_cval = (const char *)(uintptr_t)md32.md_cval; 1165 md.md_data = (void *)(uintptr_t)md32.md_data; 1166 #else 1167 COPYOUT(v, &md, sizeof(md)); 1168 error = __elfN(reloc_ptr)(fp, ef, v, &md, sizeof(md)); 1169 if (error == EOPNOTSUPP) { 1170 md.md_cval += ef->off; 1171 md.md_data = (void *)((uintptr_t)md.md_data + 1172 (uintptr_t)ef->off); 1173 } else if (error != 0) 1174 return (error); 1175 #endif 1176 p += sizeof(Elf_Addr); 1177 switch(md.md_type) { 1178 case MDT_DEPEND: 1179 if (ef->kernel) /* kernel must not depend on anything */ 1180 break; 1181 s = strdupout((vm_offset_t)md.md_cval); 1182 minfolen = sizeof(*mdepend) + strlen(s) + 1; 1183 mdepend = malloc(minfolen); 1184 if (mdepend == NULL) 1185 return ENOMEM; 1186 COPYOUT((vm_offset_t)md.md_data, mdepend, 1187 sizeof(*mdepend)); 1188 strcpy((char*)(mdepend + 1), s); 1189 free(s); 1190 file_addmetadata(fp, MODINFOMD_DEPLIST, minfolen, 1191 mdepend); 1192 free(mdepend); 1193 break; 1194 case MDT_VERSION: 1195 s = strdupout((vm_offset_t)md.md_cval); 1196 COPYOUT((vm_offset_t)md.md_data, &mver, sizeof(mver)); 1197 file_addmodule(fp, s, mver.mv_version, NULL); 1198 free(s); 1199 modcnt++; 1200 break; 1201 } 1202 } 1203 if (modcnt == 0) { 1204 s = fake_modname(fp->f_name); 1205 file_addmodule(fp, s, 1, NULL); 1206 free(s); 1207 } 1208 return 0; 1209 } 1210 1211 static unsigned long 1212 elf_hash(const char *name) 1213 { 1214 const unsigned char *p = (const unsigned char *) name; 1215 unsigned long h = 0; 1216 unsigned long g; 1217 1218 while (*p != '\0') { 1219 h = (h << 4) + *p++; 1220 if ((g = h & 0xf0000000) != 0) 1221 h ^= g >> 24; 1222 h &= ~g; 1223 } 1224 return h; 1225 } 1226 1227 static const char __elfN(bad_symtable)[] = "elf" __XSTRING(__ELF_WORD_SIZE) 1228 "_lookup_symbol: corrupt symbol table\n"; 1229 int 1230 __elfN(lookup_symbol)(elf_file_t ef, const char* name, Elf_Sym *symp, 1231 unsigned char type) 1232 { 1233 Elf_Hashelt symnum; 1234 Elf_Sym sym; 1235 char *strp; 1236 unsigned long hash; 1237 1238 if (ef->nbuckets == 0) { 1239 printf(__elfN(bad_symtable)); 1240 return ENOENT; 1241 } 1242 1243 hash = elf_hash(name); 1244 COPYOUT(&ef->buckets[hash % ef->nbuckets], &symnum, sizeof(symnum)); 1245 1246 while (symnum != STN_UNDEF) { 1247 if (symnum >= ef->nchains) { 1248 printf(__elfN(bad_symtable)); 1249 return ENOENT; 1250 } 1251 1252 COPYOUT(ef->symtab + symnum, &sym, sizeof(sym)); 1253 if (sym.st_name == 0) { 1254 printf(__elfN(bad_symtable)); 1255 return ENOENT; 1256 } 1257 1258 strp = strdupout((vm_offset_t)(ef->strtab + sym.st_name)); 1259 if (strcmp(name, strp) == 0) { 1260 free(strp); 1261 if (sym.st_shndx != SHN_UNDEF && sym.st_value != 0 && 1262 ELF_ST_TYPE(sym.st_info) == type) { 1263 *symp = sym; 1264 return 0; 1265 } 1266 return ENOENT; 1267 } 1268 free(strp); 1269 COPYOUT(&ef->chains[symnum], &symnum, sizeof(symnum)); 1270 } 1271 return ENOENT; 1272 } 1273 1274 /* 1275 * Apply any intra-module relocations to the value. p is the load address 1276 * of the value and val/len is the value to be modified. This does NOT modify 1277 * the image in-place, because this is done by kern_linker later on. 1278 * 1279 * Returns EOPNOTSUPP if no relocation method is supplied. 1280 */ 1281 static int 1282 __elfN(reloc_ptr)(struct preloaded_file *mp, elf_file_t ef, 1283 Elf_Addr p, void *val, size_t len) 1284 { 1285 size_t n; 1286 Elf_Rela a; 1287 Elf_Rel r; 1288 int error; 1289 1290 /* 1291 * The kernel is already relocated, but we still want to apply 1292 * offset adjustments. 1293 */ 1294 if (ef->kernel) 1295 return (EOPNOTSUPP); 1296 1297 for (n = 0; n < ef->relsz / sizeof(r); n++) { 1298 COPYOUT(ef->rel + n, &r, sizeof(r)); 1299 1300 error = __elfN(reloc)(ef, __elfN(symaddr), &r, ELF_RELOC_REL, 1301 ef->off, p, val, len); 1302 if (error != 0) 1303 return (error); 1304 } 1305 for (n = 0; n < ef->relasz / sizeof(a); n++) { 1306 COPYOUT(ef->rela + n, &a, sizeof(a)); 1307 1308 error = __elfN(reloc)(ef, __elfN(symaddr), &a, ELF_RELOC_RELA, 1309 ef->off, p, val, len); 1310 if (error != 0) 1311 return (error); 1312 } 1313 1314 return (0); 1315 } 1316 1317 static Elf_Addr 1318 __elfN(symaddr)(struct elf_file *ef, Elf_Size symidx) 1319 { 1320 1321 /* Symbol lookup by index not required here. */ 1322 return (0); 1323 } 1324