Lines Matching refs:va

324 void moea_dumpsys_map(vm_paddr_t pa, size_t sz, void **va);
457 tlbie(vm_offset_t va) in tlbie() argument
462 __asm __volatile("tlbie %0" :: "r"(va)); in tlbie()
470 vm_offset_t va; in tlbia() local
472 for (va = 0; va < 0x00040000; va += 0x00001000) { in tlbia()
473 __asm __volatile("tlbie %0" :: "r"(va)); in tlbia()
481 va_to_sr(u_int *sr, vm_offset_t va) in va_to_sr() argument
483 return (sr[(uintptr_t)va >> ADDR_SR_SHFT]); in va_to_sr()
536 moea_pte_match(struct pte *pt, u_int sr, vm_offset_t va, int which) in moea_pte_match() argument
540 ((va >> ADDR_API_SHFT) & PTE_API) | which); in moea_pte_match()
544 moea_pte_create(struct pte *pt, u_int sr, vm_offset_t va, u_int pte_lo) in moea_pte_create() argument
556 (((va & ADDR_PIDX) >> ADDR_API_SHFT) & PTE_API); in moea_pte_create()
569 moea_pte_clear(struct pte *pt, vm_offset_t va, int ptebit) in moea_pte_clear() argument
578 tlbie(va); in moea_pte_clear()
601 moea_pte_unset(struct pte *pt, struct pte *pvo_pt, vm_offset_t va) in moea_pte_unset() argument
617 tlbie(va); in moea_pte_unset()
627 moea_pte_change(struct pte *pt, struct pte *pvo_pt, vm_offset_t va) in moea_pte_change() argument
633 moea_pte_unset(pt, pvo_pt, va); in moea_pte_change()
704 vm_offset_t pa, va, off; in moea_bootstrap() local
962 va = virtual_avail + KSTACK_GUARD_PAGES * PAGE_SIZE; in moea_bootstrap()
963 virtual_avail = va + kstack_pages * PAGE_SIZE; in moea_bootstrap()
964 CTR2(KTR_PMAP, "moea_bootstrap: kstack0 at %#x (%#x)", pa, va); in moea_bootstrap()
965 thread0.td_kstack = va; in moea_bootstrap()
968 moea_kenter(va, pa); in moea_bootstrap()
970 va += PAGE_SIZE; in moea_bootstrap()
978 va = virtual_avail; in moea_bootstrap()
980 while (va < virtual_avail) { in moea_bootstrap()
981 moea_kenter(va, pa); in moea_bootstrap()
983 va += PAGE_SIZE; in moea_bootstrap()
991 va = virtual_avail; in moea_bootstrap()
993 while (va < virtual_avail) { in moea_bootstrap()
994 moea_kenter(va, pa); in moea_bootstrap()
996 va += PAGE_SIZE; in moea_bootstrap()
1103 void *va = (void *)(pa + off); in moea_zero_page_area() local
1105 bzero(va, size); in moea_zero_page_area()
1138 moea_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, in moea_enter() argument
1146 error = moea_enter_locked(pmap, va, m, prot, flags, psind); in moea_enter()
1166 moea_enter_locked(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, in moea_enter_locked() argument
1207 error = moea_pvo_enter(pmap, zone, pvo_head, va, VM_PAGE_TO_PHYS(m), in moea_enter_locked()
1240 vm_offset_t va; in moea_enter_object() local
1251 va = start + ptoa(m->pindex - m_start->pindex); in moea_enter_object()
1252 moea_enter_locked(pm, va, m, prot & in moea_enter_object()
1262 moea_enter_quick(pmap_t pm, vm_offset_t va, vm_page_t m, in moea_enter_quick() argument
1268 moea_enter_locked(pm, va, m, prot & (VM_PROT_READ | VM_PROT_EXECUTE), in moea_enter_quick()
1275 moea_extract(pmap_t pm, vm_offset_t va) in moea_extract() argument
1281 pvo = moea_pvo_find_va(pm, va & ~ADDR_POFF, NULL); in moea_extract()
1285 pa = PVO_PADDR(pvo) | (va & ADDR_POFF); in moea_extract()
1296 moea_extract_and_hold(pmap_t pmap, vm_offset_t va, vm_prot_t prot) in moea_extract_and_hold() argument
1303 pvo = moea_pvo_find_va(pmap, va & ~ADDR_POFF, NULL); in moea_extract_and_hold()
1362 moea_is_prefaultable(pmap_t pmap, vm_offset_t va) in moea_is_prefaultable() argument
1368 pvo = moea_pvo_find_va(pmap, va & ~ADDR_POFF, NULL); in moea_is_prefaultable()
1507 moea_kenter(vm_offset_t va, vm_paddr_t pa) in moea_kenter() argument
1510 moea_kenter_attr(va, pa, VM_MEMATTR_DEFAULT); in moea_kenter()
1514 moea_kenter_attr(vm_offset_t va, vm_paddr_t pa, vm_memattr_t ma) in moea_kenter_attr() argument
1520 if (va < VM_MIN_KERNEL_ADDRESS) in moea_kenter_attr()
1522 va); in moea_kenter_attr()
1529 &moea_pvo_kunmanaged, va, pa, pte_lo, PVO_WIRED); in moea_kenter_attr()
1532 panic("moea_kenter: failed to enter va %#x pa %#x: %d", va, in moea_kenter_attr()
1543 moea_kextract(vm_offset_t va) in moea_kextract() argument
1551 if (va < VM_MIN_KERNEL_ADDRESS) { in moea_kextract()
1552 return (va); in moea_kextract()
1556 pvo = moea_pvo_find_va(kernel_pmap, va & ~ADDR_POFF, NULL); in moea_kextract()
1558 pa = PVO_PADDR(pvo) | (va & ADDR_POFF); in moea_kextract()
1567 moea_kremove(vm_offset_t va) in moea_kremove() argument
1570 moea_remove(kernel_pmap, va, va + PAGE_SIZE); in moea_kremove()
1650 vm_offset_t sva, va; in moea_map() local
1653 va = sva; in moea_map()
1654 for (; pa_start < pa_end; pa_start += PAGE_SIZE, va += PAGE_SIZE) in moea_map()
1655 moea_kenter(va, pa_start); in moea_map()
1656 *virt = va; in moea_map()
1852 vm_offset_t va; in moea_qenter() local
1854 va = sva; in moea_qenter()
1856 moea_kenter(va, VM_PAGE_TO_PHYS(*m)); in moea_qenter()
1857 va += PAGE_SIZE; in moea_qenter()
1869 vm_offset_t va; in moea_qremove() local
1871 va = sva; in moea_qremove()
1873 moea_kremove(va); in moea_qremove()
1874 va += PAGE_SIZE; in moea_qremove()
1947 moea_mincore(pmap_t pm, vm_offset_t va, vm_paddr_t *pap) in moea_mincore() argument
1957 pvo = moea_pvo_find_va(pm, va & ~ADDR_POFF, NULL); in moea_mincore()
2041 vm_offset_t va, vm_paddr_t pa, u_int pte_lo, int flags) in moea_pvo_enter() argument
2057 va &= ~ADDR_POFF; in moea_pvo_enter()
2058 sr = va_to_sr(pm->pm_sr, va); in moea_pvo_enter()
2059 ptegidx = va_to_pteg(sr, va); in moea_pvo_enter()
2067 if (pvo->pvo_pmap == pm && PVO_VADDR(pvo) == va) { in moea_pvo_enter()
2115 pvo->pvo_vaddr = va; in moea_pvo_enter()
2126 moea_pte_create(&pvo->pvo_pte.pte, sr, va, pa | pte_lo); in moea_pvo_enter()
2235 moea_pvo_find_va(pmap_t pm, vm_offset_t va, int *pteidx_p) in moea_pvo_find_va() argument
2241 va &= ~ADDR_POFF; in moea_pvo_find_va()
2242 sr = va_to_sr(pm->pm_sr, va); in moea_pvo_find_va()
2243 ptegidx = va_to_pteg(sr, va); in moea_pvo_find_va()
2247 if (pvo->pvo_pmap == pm && PVO_VADDR(pvo) == va) { in moea_pvo_find_va()
2699 vm_offset_t va, tmpva, ppa, offset; in moea_mapdev_attr() local
2716 va = kva_alloc(size); in moea_mapdev_attr()
2717 if (!va) in moea_mapdev_attr()
2720 for (tmpva = va; size > 0;) { in moea_mapdev_attr()
2728 return ((void *)(va + offset)); in moea_mapdev_attr()
2734 vm_offset_t base, offset, va; in moea_unmapdev() local
2740 va = (vm_offset_t)p; in moea_unmapdev()
2741 if ((va >= VM_MIN_KERNEL_ADDRESS) && (va <= virtual_end)) { in moea_unmapdev()
2742 base = trunc_page(va); in moea_unmapdev()
2743 offset = va & PAGE_MASK; in moea_unmapdev()
2751 moea_sync_icache(pmap_t pm, vm_offset_t va, vm_size_t sz) in moea_sync_icache() argument
2760 lim = round_page(va + 1); in moea_sync_icache()
2761 len = MIN(lim - va, sz); in moea_sync_icache()
2762 pvo = moea_pvo_find_va(pm, va & ~ADDR_POFF, NULL); in moea_sync_icache()
2764 pa = PVO_PADDR(pvo) | (va & ADDR_POFF); in moea_sync_icache()
2767 va += len; in moea_sync_icache()
2774 moea_dumpsys_map(vm_paddr_t pa, size_t sz, void **va) in moea_dumpsys_map() argument
2777 *va = (void *)pa; in moea_dumpsys_map()
2786 vm_offset_t va; in moea_scan_init() local
2813 va = dump_map[1].pa_start + dump_map[1].pa_size; in moea_scan_init()
2815 while (va < virtual_end) { in moea_scan_init()
2817 if (va >= kmi.buffer_sva && va < kmi.buffer_eva) { in moea_scan_init()
2818 va = kmi.buffer_eva; in moea_scan_init()
2821 pvo = moea_pvo_find_va(kernel_pmap, va & ~ADDR_POFF, NULL); in moea_scan_init()
2824 va += PAGE_SIZE; in moea_scan_init()
2826 if (va < virtual_end) { in moea_scan_init()
2827 dump_map[2].pa_start = va; in moea_scan_init()
2828 va += PAGE_SIZE; in moea_scan_init()
2830 while (va < virtual_end) { in moea_scan_init()
2832 if (va == kmi.buffer_sva) in moea_scan_init()
2834 pvo = moea_pvo_find_va(kernel_pmap, va & ~ADDR_POFF, in moea_scan_init()
2839 va += PAGE_SIZE; in moea_scan_init()
2841 dump_map[2].pa_size = va - dump_map[2].pa_start; in moea_scan_init()