Searched refs:kernel_map (Results 1 – 11 of 11) sorted by relevance
32 #define PAGE_OFFSET kernel_map.page_offset107 extern struct kernel_mapping kernel_map;112 ((x) >= kernel_map.virt_addr && (x) < (kernel_map.virt_addr + kernel_map.size))118 #define linear_mapping_pa_to_va(x) ((void *)((unsigned long)(x) + kernel_map.va_pa_offset))123 #define kernel_mapping_pa_to_va(y) ((void *)((unsigned long)(y) + kernel_map.va_kernel_pa_offset))128 #define linear_mapping_va_to_pa(x) ((unsigned long)(x) - kernel_map.va_pa_offset)133 #define kernel_mapping_va_to_pa(y) ((unsigned long)(y) - kernel_map.va_kernel_pa_offset)166 return kernel_map.virt_offset; in kaslr_offset()
24 unsigned long kernel_start = kernel_map.virt_addr; in __phys_addr_symbol()25 unsigned long kernel_end = kernel_start + kernel_map.size; in __phys_addr_symbol()39 BUG_ON(!kernel_map.va_pa_offset); in linear_mapping_va_to_pa()41 return ((unsigned long)(x) - kernel_map.va_pa_offset); in linear_mapping_va_to_pa()47 BUG_ON(!kernel_map.va_pa_offset); in linear_mapping_pa_to_va()49 return ((void *)((unsigned long)(x) + kernel_map.va_pa_offset)); in linear_mapping_pa_to_va()
42 struct kernel_mapping kernel_map __ro_after_init;43 EXPORT_SYMBOL(kernel_map);155 print_ml("kernel", (unsigned long)kernel_map.virt_addr, in print_vm_layout()243 kernel_map.va_pa_offset = PAGE_OFFSET - phys_ram_base; in setup_bootmem()312 uintptr_t reloc_offset = kernel_map.virt_addr - KERNEL_LINK_ADDR; in relocate_kernel()317 uintptr_t va_kernel_link_pa_offset = KERNEL_LINK_ADDR - kernel_map.phys_addr; in relocate_kernel()472 BUG_ON((va - kernel_map.virt_addr) >> PUD_SHIFT); in alloc_pmd_early()537 BUG_ON((va - kernel_map.virt_addr) >> PGDIR_SHIFT); in alloc_pud_early()575 BUG_ON((va - kernel_map.virt_addr) >> PGDIR_SHIFT); in alloc_p4d_early()765 kernel_map.page_offset = PAGE_OFFSET_L4; in disable_pgtable_l5()[all …]
443 address_markers[KERNEL_MAPPING_NR].start_address = kernel_map.virt_addr; in ptdump_init()
28 kernel_map.va_kernel_pa_offset); in arch_crash_save_vmcoreinfo()
180 this_hart_id, kernel_map.va_pa_offset); in machine_kexec()
392 kernel_map.virt_offset, in dump_kernel_offset()
76 la a1, kernel_map
543 struct map *kernel_map = machine__kernel_map(machine); in map__rip_2objdump() local545 if (kernel_map) in map__rip_2objdump()546 map = kernel_map; in map__rip_2objdump()
617 struct map *kernel_map = machine__kernel_map(&rep->session->machines.host); in evlist__tty_browse_hists() 618 struct kmap *kernel_kmap = kernel_map ? map__kmap(kernel_map) : NULL; in evlist__tty_browse_hists() 623 if (kernel_map == NULL || in report__warn_kptr_restrict() 624 (dso__hit(map__dso(kernel_map)) && in report__warn_kptr_restrict() 631 if (kernel_map && map__has_symbols(kernel_map)) { in report__warn_kptr_restrict() 622 struct map *kernel_map = machine__kernel_map(&rep->session->machines.host); report__warn_kptr_restrict() local
346 struct map *kernel_map; in build_alloc_func_list() local 363 kernel_map = machine__kernel_map(machine); in build_alloc_func_list() 364 if (map__load(kernel_map) < 0) { in build_alloc_func_list() 369 map__for_each_symbol(kernel_map, sym, node) { in build_alloc_func_list()