Lines Matching refs:mmu
128 window_size = mmu_btop(map_len) * mmu.pte_size; in hat_kmap_init()
137 mmu.kmap_htables = in hat_kmap_init()
149 mmu.kmap_htables[i] = ht; in hat_kmap_init()
164 mmu.kmap_addr = map_addr; in hat_kmap_init()
165 mmu.kmap_eaddr = map_eaddr; in hat_kmap_init()
166 mmu.kmap_ptes = (x86pte_t *)ptes; in hat_kmap_init()
242 level_t lpagel = mmu.max_page_level; in hat_kern_alloc()
292 nwindows = MAX(nwindows, mmu.max_level); in hat_kern_alloc()
300 mmu.pwin_base = vmem_xalloc(heap_arena, nwindows * MMU_PAGESIZE, in hat_kern_alloc()
302 ASSERT(nwindows <= MMU_PAGESIZE / mmu.pte_size); in hat_kern_alloc()
303 mmu.pwin_pte_va = vmem_xalloc(heap_arena, MMU_PAGESIZE, in hat_kern_alloc()
310 (void) find_pte((uintptr_t)mmu.pwin_base, &paddr, 0, 0); in hat_kern_alloc()
313 mmu.pwin_pte_pa = paddr; in hat_kern_alloc()
315 (void) find_pte((uintptr_t)mmu.pwin_pte_va, NULL, 0, 0); in hat_kern_alloc()
316 kbm_read_only((uintptr_t)mmu.pwin_pte_va, mmu.pwin_pte_pa); in hat_kern_alloc()
318 kbm_map((uintptr_t)mmu.pwin_pte_va, mmu.pwin_pte_pa, 0, 1); in hat_kern_alloc()
333 while (start_level <= mmu.max_page_level) { in hat_kern_alloc()
339 for (l = start_level; l < mmu.max_level; ++l) { in hat_kern_alloc()
355 table_cnt += mmu.top_level_count - ((kernelbase >> in hat_kern_alloc()
356 LEVEL_SHIFT(mmu.max_level)) & (mmu.top_level_count - 1)); in hat_kern_alloc()
369 mapping_cnt = (table_cnt * mmu.ptes_per_table) >> 4; in hat_kern_alloc()
397 htable_attach(kas.a_hat, 0, mmu.max_level, NULL, in hat_kern_setup()