Lines Matching defs:vaddr
1144 uintptr_t vaddr = (uintptr_t)0;
1158 ASSERT(IS_PAGEALIGNED(vaddr));
1164 while (vaddr < eaddr) {
1165 (void) htable_walk(hat, &ht, &vaddr, eaddr);
1169 ASSERT(!IN_VA_HOLE(vaddr));
1176 vaddr = ht->ht_vaddr + LEVEL_SIZE(l + 1);
1186 hat_unload(hat, (caddr_t)vaddr, LEVEL_SIZE(l),
1194 vaddr = ht->ht_vaddr + LEVEL_SIZE(1);
1196 vaddr += LEVEL_SIZE(l);
1873 uintptr_t vaddr = (uintptr_t)addr;
1874 uintptr_t eaddr = vaddr + len;
1881 ASSERT(IS_PAGEALIGNED(vaddr));
1890 while (vaddr < eaddr) {
1891 (void) htable_walk(hat, &ht, &vaddr, eaddr);
1895 ASSERT(!IN_VA_HOLE(vaddr));
1899 "htable=%p, vaddr=%p\n", (void *)ht, (void *)vaddr);
1902 vaddr += LEVEL_SIZE(ht->ht_level);
2381 uintptr_t vaddr = (uintptr_t)addr;
2382 uintptr_t eaddr = vaddr + len;
2392 ASSERT(IS_PAGEALIGNED(vaddr));
2400 ht = htable_getpte(hat, vaddr, &entry, &old_pte, 0);
2412 while (vaddr < eaddr) {
2413 old_pte = htable_walk(hat, &ht, &vaddr, eaddr);
2417 ASSERT(!IN_VA_HOLE(vaddr));
2419 if (vaddr < (uintptr_t)addr)
2425 if (vaddr != contig_va ||
2431 r[r_cnt].rng_va = vaddr;
2446 entry = htable_va2entry(vaddr, ht);
2449 vaddr += LEVEL_SIZE(ht->ht_level);
2450 contig_va = vaddr;
2503 uintptr_t vaddr = (uintptr_t)addr;
2504 uintptr_t eaddr = vaddr + len;
2512 ASSERT(!IN_VA_HOLE(vaddr));
2513 ASSERT(IS_PAGEALIGNED(vaddr));
2518 for (; vaddr < eaddr; vaddr += LEVEL_SIZE(ht->ht_level)) {
2520 pte = htable_walk(hat, &ht, &vaddr, eaddr);
2523 entry = htable_va2entry(vaddr, ht);
2593 uintptr_t vaddr = ALIGN2PAGE(addr);
2597 ASSERT(hat == kas.a_hat || vaddr <= _userlimit);
2599 if (IN_VA_HOLE(vaddr))
2602 ht = htable_getpte(hat, vaddr, NULL, &pte, mmu.max_page_level);
2634 uintptr_t vaddr = (uintptr_t)addr;
2642 ASSERT(IS_PAGEALIGNED(vaddr));
2645 for (; vaddr < eaddr; vaddr += LEVEL_SIZE(ht->ht_level)) {
2647 oldpte = htable_walk(hat, &ht, &vaddr, eaddr);
2717 entry = htable_va2entry(vaddr, ht);
2770 uintptr_t vaddr = ALIGN2PAGE(addr);
2774 ASSERT(hat == kas.a_hat || vaddr <= _userlimit);
2775 if (IN_VA_HOLE(vaddr))
2777 ht = htable_getpage(hat, vaddr, NULL);
2794 uintptr_t vaddr = ALIGN2PAGE(addr);
2799 ASSERT(hat == kas.a_hat || vaddr <= _userlimit);
2803 if (IN_VA_HOLE(vaddr))
2812 if (mmu.kmap_addr <= vaddr && vaddr < mmu.kmap_eaddr) {
2816 pg_index = mmu_btop(vaddr - mmu.kmap_addr);
2825 ht = htable_getpage(hat, vaddr, &entry);
2830 ASSERT(vaddr >= ht->ht_vaddr);
2831 ASSERT(vaddr <= HTABLE_LAST_PAGE(ht));
2834 pfn += mmu_btop(vaddr & LEVEL_OFFSET(ht->ht_level));
2848 uintptr_t vaddr = ALIGN2PAGE(addr);
2853 ASSERT(hat == kas.a_hat || vaddr <= _userlimit);
2855 if (IN_VA_HOLE(vaddr))
2862 if (mmu.kmap_addr <= vaddr && vaddr < mmu.kmap_eaddr) {
2863 pg_off = mmu_btop(vaddr - mmu.kmap_addr);
2871 ht = htable_getpage(hat, vaddr, &entry);
2915 uintptr_t vaddr;
2963 vaddr = vaddr_start + (ism_addr - ism_addr_start);
2964 ht = htable_lookup(hat, vaddr, l);
2990 (vaddr & LEVEL_OFFSET(l + 1)) != 0)
3019 ht = htable_create(hat, vaddr, l, ism_ht);
3040 if (LEVEL_SIZE(l) <= eaddr - vaddr &&
3041 (vaddr & LEVEL_OFFSET(l)) == 0)
3049 if (eaddr - vaddr >= LEVEL_SIZE(ism_ht->ht_level)) {
3052 pgcnt = mmu_btop(eaddr - vaddr);
3075 while (hati_load_common(hat, vaddr, pp, prot, flags,
3082 vaddr += LEVEL_SIZE(l);
3104 uint64_t vaddr = (uintptr_t)addr;
3105 uintptr_t eaddr = vaddr + len;
3113 ASSERT(IS_PAGEALIGNED(vaddr));
3128 for (vaddr = (uintptr_t)addr; vaddr < eaddr;
3129 vaddr = (vaddr & LEVEL_MASK(l + 1)) + LEVEL_SIZE(l + 1)) {
3130 ASSERT(!IN_VA_HOLE(vaddr));
3134 ht = htable_lookup(hat, vaddr, l);
4057 uintptr_t vaddr = low;
4067 while (vaddr < high) {
4068 pte = htable_walk(kas.a_hat, &ht, &vaddr, high);
4073 entry = htable_va2entry(vaddr, ht);
4085 vaddr += LEVEL_SIZE(ht->ht_level);
4201 * Map in a locked page and return the vaddr.
4207 caddr_t vaddr;
4221 vaddr = hat_kpm_page2va(pp, 1);
4223 return (vaddr);
4231 hat_kpm_mapout(struct page *pp, struct kpme *kpme, caddr_t vaddr)
4239 if (IS_KPM_ADDR(vaddr) == 0) {
4265 caddr_t paddr, vaddr;
4271 vaddr = (uintptr_t)kpm_vbase + paddr;
4273 return ((caddr_t)vaddr);
4289 uintptr_t vaddr = (uintptr_t)kpm_vbase + mmu_ptob(pfn);
4292 return ((caddr_t)vaddr);
4306 * Return the page frame number for the kpm virtual address vaddr.
4309 hat_kpm_va2pfn(caddr_t vaddr)
4313 ASSERT(IS_KPM_ADDR(vaddr));
4315 pfn = (pfn_t)btop(vaddr - kpm_vbase);
4322 * Return the page for the kpm virtual address vaddr.
4325 hat_kpm_vaddr2page(caddr_t vaddr)
4329 ASSERT(IS_KPM_ADDR(vaddr));
4331 pfn = hat_kpm_va2pfn(vaddr);
4341 hat_kpm_fault(hat_t *hat, caddr_t vaddr)
4343 panic("pagefault in seg_kpm. hat: 0x%p vaddr: 0x%p",
4344 (void *)hat, (void *)vaddr);