1 /*- 2 * Copyright (c) 2015-2017 Ruslan Bukin <br@bsdpad.com> 3 * All rights reserved. 4 * 5 * This software was developed by SRI International and the University of 6 * Cambridge Computer Laboratory under DARPA/AFRL contract FA8750-10-C-0237 7 * ("CTSRD"), as part of the DARPA CRASH research programme. 8 * 9 * This software was developed by the University of Cambridge Computer 10 * Laboratory as part of the CTSRD Project, with support from the UK Higher 11 * Education Innovation Fund (HEIF). 12 * 13 * Redistribution and use in source and binary forms, with or without 14 * modification, are permitted provided that the following conditions 15 * are met: 16 * 1. Redistributions of source code must retain the above copyright 17 * notice, this list of conditions and the following disclaimer. 18 * 2. Redistributions in binary form must reproduce the above copyright 19 * notice, this list of conditions and the following disclaimer in the 20 * documentation and/or other materials provided with the distribution. 21 * 22 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 25 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 26 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 32 * SUCH DAMAGE. 33 */ 34 35 #include <sys/cdefs.h> 36 __FBSDID("$FreeBSD$"); 37 38 #include <sys/types.h> 39 40 #include <stdlib.h> 41 42 #include "debug.h" 43 #include "rtld.h" 44 #include "rtld_printf.h" 45 46 /* 47 * It is possible for the compiler to emit relocations for unaligned data. 48 * We handle this situation with these inlines. 49 */ 50 #define RELOC_ALIGNED_P(x) \ 51 (((uintptr_t)(x) & (sizeof(void *) - 1)) == 0) 52 53 void _exit(int); 54 55 uint64_t 56 set_gp(Obj_Entry *obj) 57 { 58 uint64_t old; 59 SymLook req; 60 uint64_t gp; 61 int res; 62 63 __asm __volatile("mv %0, gp" : "=r"(old)); 64 65 symlook_init(&req, "__global_pointer$"); 66 req.ventry = NULL; 67 req.flags = SYMLOOK_EARLY; 68 res = symlook_obj(&req, obj); 69 70 if (res == 0) { 71 gp = req.sym_out->st_value; 72 __asm __volatile("mv gp, %0" :: "r"(gp)); 73 } 74 75 return (old); 76 } 77 78 void 79 init_pltgot(Obj_Entry *obj) 80 { 81 82 if (obj->pltgot != NULL) { 83 obj->pltgot[0] = (Elf_Addr)&_rtld_bind_start; 84 obj->pltgot[1] = (Elf_Addr)obj; 85 } 86 } 87 88 int 89 do_copy_relocations(Obj_Entry *dstobj) 90 { 91 const Obj_Entry *srcobj, *defobj; 92 const Elf_Rela *relalim; 93 const Elf_Rela *rela; 94 const Elf_Sym *srcsym; 95 const Elf_Sym *dstsym; 96 const void *srcaddr; 97 const char *name; 98 void *dstaddr; 99 SymLook req; 100 size_t size; 101 int res; 102 103 /* 104 * COPY relocs are invalid outside of the main program 105 */ 106 assert(dstobj->mainprog); 107 108 relalim = (const Elf_Rela *)((char *)dstobj->rela + 109 dstobj->relasize); 110 for (rela = dstobj->rela; rela < relalim; rela++) { 111 if (ELF_R_TYPE(rela->r_info) != R_RISCV_COPY) 112 continue; 113 114 dstaddr = (void *)(dstobj->relocbase + rela->r_offset); 115 dstsym = dstobj->symtab + ELF_R_SYM(rela->r_info); 116 name = dstobj->strtab + dstsym->st_name; 117 size = dstsym->st_size; 118 119 symlook_init(&req, name); 120 req.ventry = fetch_ventry(dstobj, ELF_R_SYM(rela->r_info)); 121 req.flags = SYMLOOK_EARLY; 122 123 for (srcobj = globallist_next(dstobj); srcobj != NULL; 124 srcobj = globallist_next(srcobj)) { 125 res = symlook_obj(&req, srcobj); 126 if (res == 0) { 127 srcsym = req.sym_out; 128 defobj = req.defobj_out; 129 break; 130 } 131 } 132 if (srcobj == NULL) { 133 _rtld_error( 134 "Undefined symbol \"%s\" referenced from COPY relocation in %s", 135 name, dstobj->path); 136 return (-1); 137 } 138 139 srcaddr = (const void *)(defobj->relocbase + srcsym->st_value); 140 memcpy(dstaddr, srcaddr, size); 141 } 142 143 return (0); 144 } 145 146 /* 147 * Process the PLT relocations. 148 */ 149 int 150 reloc_plt(Obj_Entry *obj) 151 { 152 const Elf_Rela *relalim; 153 const Elf_Rela *rela; 154 155 relalim = (const Elf_Rela *)((char *)obj->pltrela + obj->pltrelasize); 156 for (rela = obj->pltrela; rela < relalim; rela++) { 157 Elf_Addr *where; 158 159 assert(ELF_R_TYPE(rela->r_info) == R_RISCV_JUMP_SLOT); 160 161 where = (Elf_Addr *)(obj->relocbase + rela->r_offset); 162 *where += (Elf_Addr)obj->relocbase; 163 } 164 165 return (0); 166 } 167 168 /* 169 * LD_BIND_NOW was set - force relocation for all jump slots 170 */ 171 int 172 reloc_jmpslots(Obj_Entry *obj, int flags, RtldLockState *lockstate) 173 { 174 const Obj_Entry *defobj; 175 const Elf_Rela *relalim; 176 const Elf_Rela *rela; 177 const Elf_Sym *def; 178 179 relalim = (const Elf_Rela *)((char *)obj->pltrela + obj->pltrelasize); 180 for (rela = obj->pltrela; rela < relalim; rela++) { 181 Elf_Addr *where; 182 183 where = (Elf_Addr *)(obj->relocbase + rela->r_offset); 184 switch(ELF_R_TYPE(rela->r_info)) { 185 case R_RISCV_JUMP_SLOT: 186 def = find_symdef(ELF_R_SYM(rela->r_info), obj, 187 &defobj, SYMLOOK_IN_PLT | flags, NULL, lockstate); 188 if (def == NULL) { 189 dbg("reloc_jmpslots: sym not found"); 190 return (-1); 191 } 192 193 *where = (Elf_Addr)(defobj->relocbase + def->st_value); 194 break; 195 default: 196 _rtld_error("Unknown relocation type %x in jmpslot", 197 (unsigned int)ELF_R_TYPE(rela->r_info)); 198 return (-1); 199 } 200 } 201 202 return (0); 203 } 204 205 int 206 reloc_iresolve(Obj_Entry *obj, struct Struct_RtldLockState *lockstate) 207 { 208 209 /* XXX not implemented */ 210 return (0); 211 } 212 213 int 214 reloc_gnu_ifunc(Obj_Entry *obj, int flags, 215 struct Struct_RtldLockState *lockstate) 216 { 217 218 /* XXX not implemented */ 219 return (0); 220 } 221 222 Elf_Addr 223 reloc_jmpslot(Elf_Addr *where, Elf_Addr target, const Obj_Entry *defobj, 224 const Obj_Entry *obj, const Elf_Rel *rel) 225 { 226 227 assert(ELF_R_TYPE(rel->r_info) == R_RISCV_JUMP_SLOT); 228 229 if (*where != target && !ld_bind_not) 230 *where = target; 231 return (target); 232 } 233 234 /* 235 * Process non-PLT relocations 236 */ 237 int 238 reloc_non_plt(Obj_Entry *obj, Obj_Entry *obj_rtld, int flags, 239 RtldLockState *lockstate) 240 { 241 const Obj_Entry *defobj; 242 const Elf_Rela *relalim; 243 const Elf_Rela *rela; 244 const Elf_Sym *def; 245 SymCache *cache; 246 Elf_Addr *where; 247 unsigned long symnum; 248 249 if ((flags & SYMLOOK_IFUNC) != 0) 250 /* XXX not implemented */ 251 return (0); 252 253 /* 254 * The dynamic loader may be called from a thread, we have 255 * limited amounts of stack available so we cannot use alloca(). 256 */ 257 if (obj == obj_rtld) 258 cache = NULL; 259 else 260 cache = calloc(obj->dynsymcount, sizeof(SymCache)); 261 /* No need to check for NULL here */ 262 263 relalim = (const Elf_Rela *)((caddr_t)obj->rela + obj->relasize); 264 for (rela = obj->rela; rela < relalim; rela++) { 265 where = (Elf_Addr *)(obj->relocbase + rela->r_offset); 266 symnum = ELF_R_SYM(rela->r_info); 267 268 switch (ELF_R_TYPE(rela->r_info)) { 269 case R_RISCV_JUMP_SLOT: 270 /* This will be handled by the plt/jmpslot routines */ 271 break; 272 case R_RISCV_NONE: 273 break; 274 case R_RISCV_64: 275 def = find_symdef(symnum, obj, &defobj, flags, cache, 276 lockstate); 277 if (def == NULL) 278 return (-1); 279 280 *where = (Elf_Addr)(defobj->relocbase + def->st_value + 281 rela->r_addend); 282 break; 283 case R_RISCV_TLS_DTPMOD64: 284 def = find_symdef(symnum, obj, &defobj, flags, cache, 285 lockstate); 286 if (def == NULL) 287 return -1; 288 289 *where += (Elf_Addr)defobj->tlsindex; 290 break; 291 case R_RISCV_COPY: 292 /* 293 * These are deferred until all other relocations have 294 * been done. All we do here is make sure that the 295 * COPY relocation is not in a shared library. They 296 * are allowed only in executable files. 297 */ 298 if (!obj->mainprog) { 299 _rtld_error("%s: Unexpected R_RISCV_COPY " 300 "relocation in shared library", obj->path); 301 return (-1); 302 } 303 break; 304 case R_RISCV_TLS_DTPREL64: 305 def = find_symdef(symnum, obj, &defobj, flags, cache, 306 lockstate); 307 if (def == NULL) 308 return (-1); 309 /* 310 * We lazily allocate offsets for static TLS as we 311 * see the first relocation that references the 312 * TLS block. This allows us to support (small 313 * amounts of) static TLS in dynamically loaded 314 * modules. If we run out of space, we generate an 315 * error. 316 */ 317 if (!defobj->tls_done) { 318 if (!allocate_tls_offset((Obj_Entry*) defobj)) { 319 _rtld_error( 320 "%s: No space available for static " 321 "Thread Local Storage", obj->path); 322 return (-1); 323 } 324 } 325 326 *where += (Elf_Addr)(def->st_value + rela->r_addend 327 - TLS_DTV_OFFSET); 328 break; 329 case R_RISCV_TLS_TPREL64: 330 def = find_symdef(symnum, obj, &defobj, flags, cache, 331 lockstate); 332 if (def == NULL) 333 return (-1); 334 335 /* 336 * We lazily allocate offsets for static TLS as we 337 * see the first relocation that references the 338 * TLS block. This allows us to support (small 339 * amounts of) static TLS in dynamically loaded 340 * modules. If we run out of space, we generate an 341 * error. 342 */ 343 if (!defobj->tls_done) { 344 if (!allocate_tls_offset((Obj_Entry*) defobj)) { 345 _rtld_error( 346 "%s: No space available for static " 347 "Thread Local Storage", obj->path); 348 return (-1); 349 } 350 } 351 352 *where = (def->st_value + rela->r_addend + 353 defobj->tlsoffset - TLS_TP_OFFSET); 354 break; 355 case R_RISCV_RELATIVE: 356 *where = (Elf_Addr)(obj->relocbase + rela->r_addend); 357 break; 358 default: 359 rtld_printf("%s: Unhandled relocation %lu\n", 360 obj->path, ELF_R_TYPE(rela->r_info)); 361 return (-1); 362 } 363 } 364 365 return (0); 366 } 367 368 void 369 ifunc_init(Elf_Auxinfo aux_info[__min_size(AT_COUNT)] __unused) 370 { 371 } 372 373 void 374 allocate_initial_tls(Obj_Entry *objs) 375 { 376 Elf_Addr **tp; 377 378 /* 379 * Fix the size of the static TLS block by using the maximum 380 * offset allocated so far and adding a bit for dynamic modules to 381 * use. 382 */ 383 tls_static_space = tls_last_offset + tls_last_size + 384 RTLD_STATIC_TLS_EXTRA; 385 386 tp = (Elf_Addr **) ((char *)allocate_tls(objs, NULL, TLS_TCB_SIZE, 16) 387 + TLS_TP_OFFSET + TLS_TCB_SIZE); 388 389 __asm __volatile("mv tp, %0" :: "r"(tp)); 390 } 391 392 void * 393 __tls_get_addr(tls_index* ti) 394 { 395 char *_tp; 396 void *p; 397 398 __asm __volatile("mv %0, tp" : "=r" (_tp)); 399 400 p = tls_get_addr_common((Elf_Addr**)((Elf_Addr)_tp - TLS_TP_OFFSET 401 - TLS_TCB_SIZE), ti->ti_module, ti->ti_offset); 402 403 return (p + TLS_DTV_OFFSET); 404 } 405