Lines Matching refs:mmu

209 	window_size = mmu_btop(map_len) * mmu.pte_size;  in hat_kmap_init()
218 mmu.kmap_htables = in hat_kmap_init()
230 mmu.kmap_htables[i] = ht; in hat_kmap_init()
245 mmu.kmap_addr = map_addr; in hat_kmap_init()
246 mmu.kmap_eaddr = map_eaddr; in hat_kmap_init()
247 mmu.kmap_ptes = (x86pte_t *)ptes; in hat_kmap_init()
323 level_t lpagel = mmu.max_page_level; in hat_kern_alloc()
373 nwindows = MAX(nwindows, mmu.max_level); in hat_kern_alloc()
381 mmu.pwin_base = vmem_xalloc(heap_arena, nwindows * MMU_PAGESIZE, in hat_kern_alloc()
383 ASSERT(nwindows <= MMU_PAGESIZE / mmu.pte_size); in hat_kern_alloc()
384 mmu.pwin_pte_va = vmem_xalloc(heap_arena, MMU_PAGESIZE, in hat_kern_alloc()
391 (void) find_pte((uintptr_t)mmu.pwin_base, &paddr, 0, 0); in hat_kern_alloc()
394 mmu.pwin_pte_pa = paddr; in hat_kern_alloc()
396 (void) find_pte((uintptr_t)mmu.pwin_pte_va, NULL, 0, 0); in hat_kern_alloc()
397 kbm_read_only((uintptr_t)mmu.pwin_pte_va, mmu.pwin_pte_pa); in hat_kern_alloc()
399 kbm_map((uintptr_t)mmu.pwin_pte_va, mmu.pwin_pte_pa, 0, 1); in hat_kern_alloc()
414 while (start_level <= mmu.max_page_level) { in hat_kern_alloc()
420 for (l = start_level; l < mmu.max_level; ++l) { in hat_kern_alloc()
436 table_cnt += mmu.top_level_count - ((kernelbase >> in hat_kern_alloc()
437 LEVEL_SHIFT(mmu.max_level)) & (mmu.top_level_count - 1)); in hat_kern_alloc()
447 if (mmu.pae_hat != 0 && (kernelbase & LEVEL_OFFSET(mmu.max_level)) != 0) in hat_kern_alloc()
448 table_cnt += mmu.ptes_per_table - in hat_kern_alloc()
449 ((kernelbase & LEVEL_OFFSET(mmu.max_level)) >> in hat_kern_alloc()
450 LEVEL_SHIFT(mmu.max_level - 1)); in hat_kern_alloc()
464 mapping_cnt = (table_cnt * mmu.ptes_per_table) >> 4; in hat_kern_alloc()
492 htable_attach(kas.a_hat, 0, mmu.max_level, NULL, in hat_kern_setup()