1 /*- 2 * Copyright (c) 1989, 1992, 1993 3 * The Regents of the University of California. All rights reserved. 4 * 5 * This code is derived from software developed by the Computer Systems 6 * Engineering group at Lawrence Berkeley Laboratory under DARPA contract 7 * BG 91-66 and contributed to Berkeley. 8 * 9 * Redistribution and use in source and binary forms, with or without 10 * modification, are permitted provided that the following conditions 11 * are met: 12 * 1. Redistributions of source code must retain the above copyright 13 * notice, this list of conditions and the following disclaimer. 14 * 2. Redistributions in binary form must reproduce the above copyright 15 * notice, this list of conditions and the following disclaimer in the 16 * documentation and/or other materials provided with the distribution. 17 * 3. Neither the name of the University nor the names of its contributors 18 * may be used to endorse or promote products derived from this software 19 * without specific prior written permission. 20 * 21 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 22 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 23 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 24 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 25 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 26 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 27 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 28 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 29 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 30 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 31 * SUCH DAMAGE. 32 */ 33 34 #include <sys/cdefs.h> 35 __FBSDID("$FreeBSD$"); 36 37 #if defined(LIBC_SCCS) && !defined(lint) 38 #if 0 39 static char sccsid[] = "@(#)kvm_hp300.c 8.1 (Berkeley) 6/4/93"; 40 #endif 41 #endif /* LIBC_SCCS and not lint */ 42 43 /* 44 * AMD64 machine dependent routines for kvm. Hopefully, the forthcoming 45 * vm code will one day obsolete this module. 46 */ 47 48 #include <sys/param.h> 49 #include <sys/endian.h> 50 #include <stdint.h> 51 #include <stdlib.h> 52 #include <string.h> 53 #include <unistd.h> 54 #include <vm/vm.h> 55 #include <kvm.h> 56 57 #include <limits.h> 58 59 #include "kvm_private.h" 60 #include "kvm_amd64.h" 61 62 struct vmstate { 63 size_t phnum; 64 GElf_Phdr *phdr; 65 amd64_pml4e_t *PML4; 66 }; 67 68 /* 69 * Translate a physical memory address to a file-offset in the crash-dump. 70 */ 71 static size_t 72 _kvm_pa2off(kvm_t *kd, uint64_t pa, off_t *ofs) 73 { 74 struct vmstate *vm = kd->vmst; 75 GElf_Phdr *p; 76 size_t n; 77 78 if (kd->rawdump) { 79 *ofs = pa; 80 return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK)); 81 } 82 83 p = vm->phdr; 84 n = vm->phnum; 85 while (n && (pa < p->p_paddr || pa >= p->p_paddr + p->p_memsz)) 86 p++, n--; 87 if (n == 0) 88 return (0); 89 *ofs = (pa - p->p_paddr) + p->p_offset; 90 return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK)); 91 } 92 93 static void 94 _amd64_freevtop(kvm_t *kd) 95 { 96 struct vmstate *vm = kd->vmst; 97 98 if (vm->PML4) 99 free(vm->PML4); 100 free(vm->phdr); 101 free(vm); 102 kd->vmst = NULL; 103 } 104 105 static int 106 _amd64_probe(kvm_t *kd) 107 { 108 109 return (_kvm_probe_elf_kernel(kd, ELFCLASS64, EM_X86_64) && 110 !_kvm_is_minidump(kd)); 111 } 112 113 static int 114 _amd64_initvtop(kvm_t *kd) 115 { 116 struct kvm_nlist nl[2]; 117 amd64_physaddr_t pa; 118 kvaddr_t kernbase; 119 amd64_pml4e_t *PML4; 120 121 kd->vmst = (struct vmstate *)_kvm_malloc(kd, sizeof(*kd->vmst)); 122 if (kd->vmst == NULL) { 123 _kvm_err(kd, kd->program, "cannot allocate vm"); 124 return (-1); 125 } 126 kd->vmst->PML4 = 0; 127 128 if (kd->rawdump == 0) { 129 if (_kvm_read_core_phdrs(kd, &kd->vmst->phnum, 130 &kd->vmst->phdr) == -1) 131 return (-1); 132 } 133 134 nl[0].n_name = "kernbase"; 135 nl[1].n_name = 0; 136 137 if (kvm_nlist2(kd, nl) != 0) { 138 _kvm_err(kd, kd->program, "bad namelist - no kernbase"); 139 return (-1); 140 } 141 kernbase = nl[0].n_value; 142 143 nl[0].n_name = "KPML4phys"; 144 nl[1].n_name = 0; 145 146 if (kvm_nlist2(kd, nl) != 0) { 147 _kvm_err(kd, kd->program, "bad namelist - no KPML4phys"); 148 return (-1); 149 } 150 if (kvm_read2(kd, (nl[0].n_value - kernbase), &pa, sizeof(pa)) != 151 sizeof(pa)) { 152 _kvm_err(kd, kd->program, "cannot read KPML4phys"); 153 return (-1); 154 } 155 pa = le64toh(pa); 156 PML4 = _kvm_malloc(kd, AMD64_PAGE_SIZE); 157 if (PML4 == NULL) { 158 _kvm_err(kd, kd->program, "cannot allocate PML4"); 159 return (-1); 160 } 161 if (kvm_read2(kd, pa, PML4, AMD64_PAGE_SIZE) != AMD64_PAGE_SIZE) { 162 _kvm_err(kd, kd->program, "cannot read KPML4phys"); 163 free(PML4); 164 return (-1); 165 } 166 kd->vmst->PML4 = PML4; 167 return (0); 168 } 169 170 static int 171 _amd64_vatop(kvm_t *kd, kvaddr_t va, off_t *pa) 172 { 173 struct vmstate *vm; 174 amd64_physaddr_t offset; 175 amd64_physaddr_t pdpe_pa; 176 amd64_physaddr_t pde_pa; 177 amd64_physaddr_t pte_pa; 178 amd64_pml4e_t pml4e; 179 amd64_pdpe_t pdpe; 180 amd64_pde_t pde; 181 amd64_pte_t pte; 182 kvaddr_t pml4eindex; 183 kvaddr_t pdpeindex; 184 kvaddr_t pdeindex; 185 kvaddr_t pteindex; 186 amd64_physaddr_t a; 187 off_t ofs; 188 size_t s; 189 190 vm = kd->vmst; 191 offset = va & AMD64_PAGE_MASK; 192 193 /* 194 * If we are initializing (kernel page table descriptor pointer 195 * not yet set) then return pa == va to avoid infinite recursion. 196 */ 197 if (vm->PML4 == NULL) { 198 s = _kvm_pa2off(kd, va, pa); 199 if (s == 0) { 200 _kvm_err(kd, kd->program, 201 "_amd64_vatop: bootstrap data not in dump"); 202 goto invalid; 203 } else 204 return (AMD64_PAGE_SIZE - offset); 205 } 206 207 pml4eindex = (va >> AMD64_PML4SHIFT) & (AMD64_NPML4EPG - 1); 208 pml4e = le64toh(vm->PML4[pml4eindex]); 209 if ((pml4e & AMD64_PG_V) == 0) { 210 _kvm_err(kd, kd->program, "_amd64_vatop: pml4e not valid"); 211 goto invalid; 212 } 213 214 pdpeindex = (va >> AMD64_PDPSHIFT) & (AMD64_NPDPEPG - 1); 215 pdpe_pa = (pml4e & AMD64_PG_FRAME) + (pdpeindex * sizeof(amd64_pdpe_t)); 216 217 s = _kvm_pa2off(kd, pdpe_pa, &ofs); 218 if (s < sizeof(pdpe)) { 219 _kvm_err(kd, kd->program, "_amd64_vatop: pdpe_pa not found"); 220 goto invalid; 221 } 222 if (pread(kd->pmfd, &pdpe, sizeof(pdpe), ofs) != sizeof(pdpe)) { 223 _kvm_syserr(kd, kd->program, "_amd64_vatop: read pdpe"); 224 goto invalid; 225 } 226 pdpe = le64toh(pdpe); 227 if ((pdpe & AMD64_PG_V) == 0) { 228 _kvm_err(kd, kd->program, "_amd64_vatop: pdpe not valid"); 229 goto invalid; 230 } 231 232 if (pdpe & AMD64_PG_PS) { 233 /* 234 * No next-level page table; pdpe describes one 1GB page. 235 */ 236 a = (pdpe & AMD64_PG_1GB_FRAME) + (va & AMD64_PDPMASK); 237 s = _kvm_pa2off(kd, a, pa); 238 if (s == 0) { 239 _kvm_err(kd, kd->program, 240 "_amd64_vatop: 1GB page address not in dump"); 241 goto invalid; 242 } else 243 return (AMD64_NBPDP - (va & AMD64_PDPMASK)); 244 } 245 246 pdeindex = (va >> AMD64_PDRSHIFT) & (AMD64_NPDEPG - 1); 247 pde_pa = (pdpe & AMD64_PG_FRAME) + (pdeindex * sizeof(amd64_pde_t)); 248 249 s = _kvm_pa2off(kd, pde_pa, &ofs); 250 if (s < sizeof(pde)) { 251 _kvm_syserr(kd, kd->program, "_amd64_vatop: pde_pa not found"); 252 goto invalid; 253 } 254 if (pread(kd->pmfd, &pde, sizeof(pde), ofs) != sizeof(pde)) { 255 _kvm_syserr(kd, kd->program, "_amd64_vatop: read pde"); 256 goto invalid; 257 } 258 pde = le64toh(pde); 259 if ((pde & AMD64_PG_V) == 0) { 260 _kvm_err(kd, kd->program, "_amd64_vatop: pde not valid"); 261 goto invalid; 262 } 263 264 if (pde & AMD64_PG_PS) { 265 /* 266 * No final-level page table; pde describes one 2MB page. 267 */ 268 a = (pde & AMD64_PG_PS_FRAME) + (va & AMD64_PDRMASK); 269 s = _kvm_pa2off(kd, a, pa); 270 if (s == 0) { 271 _kvm_err(kd, kd->program, 272 "_amd64_vatop: 2MB page address not in dump"); 273 goto invalid; 274 } else 275 return (AMD64_NBPDR - (va & AMD64_PDRMASK)); 276 } 277 278 pteindex = (va >> AMD64_PAGE_SHIFT) & (AMD64_NPTEPG - 1); 279 pte_pa = (pde & AMD64_PG_FRAME) + (pteindex * sizeof(amd64_pte_t)); 280 281 s = _kvm_pa2off(kd, pte_pa, &ofs); 282 if (s < sizeof(pte)) { 283 _kvm_err(kd, kd->program, "_amd64_vatop: pte_pa not found"); 284 goto invalid; 285 } 286 if (pread(kd->pmfd, &pte, sizeof(pte), ofs) != sizeof(pte)) { 287 _kvm_syserr(kd, kd->program, "_amd64_vatop: read"); 288 goto invalid; 289 } 290 if ((pte & AMD64_PG_V) == 0) { 291 _kvm_err(kd, kd->program, "_amd64_vatop: pte not valid"); 292 goto invalid; 293 } 294 295 a = (pte & AMD64_PG_FRAME) + offset; 296 s = _kvm_pa2off(kd, a, pa); 297 if (s == 0) { 298 _kvm_err(kd, kd->program, "_amd64_vatop: address not in dump"); 299 goto invalid; 300 } else 301 return (AMD64_PAGE_SIZE - offset); 302 303 invalid: 304 _kvm_err(kd, 0, "invalid address (0x%jx)", (uintmax_t)va); 305 return (0); 306 } 307 308 static int 309 _amd64_kvatop(kvm_t *kd, kvaddr_t va, off_t *pa) 310 { 311 312 if (ISALIVE(kd)) { 313 _kvm_err(kd, 0, "kvm_kvatop called in live kernel!"); 314 return (0); 315 } 316 return (_amd64_vatop(kd, va, pa)); 317 } 318 319 int 320 _amd64_native(kvm_t *kd __unused) 321 { 322 323 #ifdef __amd64__ 324 return (1); 325 #else 326 return (0); 327 #endif 328 } 329 330 static struct kvm_arch kvm_amd64 = { 331 .ka_probe = _amd64_probe, 332 .ka_initvtop = _amd64_initvtop, 333 .ka_freevtop = _amd64_freevtop, 334 .ka_kvatop = _amd64_kvatop, 335 .ka_native = _amd64_native, 336 }; 337 338 KVM_ARCH(kvm_amd64); 339