/freebsd/sys/compat/linuxkpi/common/src/ |
H A D | linux_xarray.c | 118 __xa_alloc(struct xarray *xa, uint32_t *pindex, void *ptr, uint32_t mask, gfp_t gfp) in __xa_alloc() argument 130 *pindex = (xa->xa_flags & XA_FLAGS_ALLOC1) != 0 ? 1 : 0; in __xa_alloc() 134 retval = radix_tree_insert(&xa->xa_head, *pindex, ptr); in __xa_alloc() 138 if (likely(*pindex != mask)) { in __xa_alloc() 139 (*pindex)++; in __xa_alloc() 157 xa_alloc(struct xarray *xa, uint32_t *pindex, void *ptr, uint32_t mask, gfp_t gfp) in xa_alloc() argument 165 retval = __xa_alloc(xa, pindex, ptr, mask, gfp); in xa_alloc() 178 __xa_alloc_cyclic(struct xarray *xa, uint32_t *pindex, void *ptr, uint32_t mask, in __xa_alloc_cyclic() argument 218 *pindex = *pnext_index; in __xa_alloc_cyclic() 224 xa_alloc_cyclic(struct xarray *xa, uint32_t *pindex, void *ptr, uint32_t mask, in xa_alloc_cyclic() argument [all …]
|
H A D | linux_page.c | 95 IDX_TO_OFF(page->pindex))); in linux_page_address() 301 vm_pindex_t pindex; in lkpi_vmf_insert_pfn_prot_locked() local 304 pindex = OFF_TO_IDX(addr - vma->vm_start); in lkpi_vmf_insert_pfn_prot_locked() 306 vma->vm_pfn_first = pindex; in lkpi_vmf_insert_pfn_prot_locked() 307 MPASS(pindex <= OFF_TO_IDX(vma->vm_end)); in lkpi_vmf_insert_pfn_prot_locked() 310 page = vm_page_grab(vm_obj, pindex, VM_ALLOC_NOCREAT); in lkpi_vmf_insert_pfn_prot_locked() 341 if (vm_page_insert(page, vm_obj, pindex)) { in lkpi_vmf_insert_pfn_prot_locked()
|
H A D | linux_shmemfs.c | 45 linux_shmem_read_mapping_page_gfp(vm_object_t obj, int pindex, gfp_t gfp) in linux_shmem_read_mapping_page_gfp() argument 54 rv = vm_page_grab_valid(&page, obj, pindex, VM_ALLOC_NORMAL | in linux_shmem_read_mapping_page_gfp()
|
/freebsd/sys/vm/ |
H A D | phys_pager.c | 57 vm_pindex_t pindex, int *before, int *after); 79 vm_pindex_t pindex; in phys_pager_allocate() local 88 pindex = OFF_TO_IDX(foff + PAGE_MASK + size); in phys_pager_allocate() 103 object1 = vm_object_allocate(OBJT_PHYS, pindex); in phys_pager_allocate() 112 if (pindex > object->size) in phys_pager_allocate() 113 object->size = pindex; in phys_pager_allocate() 127 if (pindex > object->size) in phys_pager_allocate() 128 object->size = pindex; in phys_pager_allocate() 133 object = vm_object_allocate(OBJT_PHYS, pindex); in phys_pager_allocate() 277 default_phys_pager_haspage(vm_object_t object, vm_pindex_t pindex, int *before, in default_phys_pager_haspage() argument [all …]
|
H A D | vm_reserv.c | 128 #define VM_RESERV_INDEX(object, pindex) \ argument 129 (((object)->pg_color + (pindex)) & (VM_LEVEL_0_NPAGES - 1)) 162 vm_pindex_t pindex; /* (o, r) offset in object */ member 282 vm_pindex_t pindex); 383 vm_reserv_insert(vm_reserv_t rv, vm_object_t object, vm_pindex_t pindex) in vm_reserv_insert() argument 400 rv->pindex = pindex; in vm_reserv_insert() 512 vm_reserv_from_object(vm_object_t object, vm_pindex_t pindex, in vm_reserv_from_object() argument 522 KASSERT(mpred->pindex < pindex, in vm_reserv_from_object() 525 if (rv->object == object && vm_reserv_has_pindex(rv, pindex)) in vm_reserv_from_object() 531 KASSERT(msucc->pindex > pindex, in vm_reserv_from_object() [all …]
|
H A D | vm_page.c | 168 vm_pindex_t pindex, const char *wmesg, int allocflags, bool locked); 175 vm_pindex_t pindex, vm_page_t mpred); 945 if (_vm_page_busy_sleep(obj, m, m->pindex, "vmpba", allocflags, in vm_page_busy_acquire() 1074 return (_vm_page_busy_sleep(obj, m, m->pindex, wmesg, allocflags, in vm_page_busy_sleep() 1091 vm_page_busy_sleep_unlocked(vm_object_t obj, vm_page_t m, vm_pindex_t pindex, in vm_page_busy_sleep_unlocked() argument 1096 (void)_vm_page_busy_sleep(obj, m, pindex, wmesg, allocflags, false); in vm_page_busy_sleep_unlocked() 1112 _vm_page_busy_sleep(vm_object_t obj, vm_page_t m, vm_pindex_t pindex, in _vm_page_busy_sleep() argument 1142 m->object != obj || m->pindex != pindex) { in _vm_page_busy_sleep() 1486 vm_page_insert_lookup(vm_page_t m, vm_object_t object, vm_pindex_t pindex, in vm_page_insert_lookup() argument 1499 m->pindex = pindex; in vm_page_insert_lookup() [all …]
|
H A D | vm_object.c | 1058 pi = p->pindex; in vm_object_page_clean() 1297 vm_object_madvise_freespace(vm_object_t object, int advice, vm_pindex_t pindex, in vm_object_madvise_freespace() argument 1302 vm_pager_freespace(object, pindex, size); in vm_object_madvise_freespace() 1326 vm_object_madvise(vm_object_t object, vm_pindex_t pindex, vm_pindex_t end, in vm_object_madvise() argument 1342 for (m = vm_page_find_least(object, pindex); pindex < end; pindex++) { in vm_object_madvise() 1351 if (m == NULL || pindex < m->pindex) { in vm_object_madvise() 1358 tpindex = (m != NULL && m->pindex < end) ? in vm_object_madvise() 1359 m->pindex : end; in vm_object_madvise() 1361 pindex, tpindex - pindex); in vm_object_madvise() 1362 if ((pindex = tpindex) == end) in vm_object_madvise() [all …]
|
H A D | swap_pager.c | 436 swap_pager_haspage(vm_object_t object, vm_pindex_t pindex, int *before, int *after); 495 vm_pindex_t pindex, vm_pindex_t count); 534 swblk_start(struct swblk *sb, vm_pindex_t pindex) in swblk_start() argument 536 return (sb == NULL || sb->p >= pindex ? in swblk_start() 537 0 : pindex % SWAP_META_PAGES); in swblk_start() 543 swblk_lookup(vm_object_t object, vm_pindex_t pindex) in swblk_lookup() argument 546 rounddown(pindex, SWAP_META_PAGES))); in swblk_lookup() 562 swblk_iter_lookup_ge(struct pctrie_iter *blks, vm_pindex_t pindex) in swblk_iter_lookup_ge() argument 565 rounddown(pindex, SWAP_META_PAGES))); in swblk_iter_lookup_ge() 579 vm_pindex_t pindex) in swblk_iter_init() argument [all …]
|
H A D | vm_domainset.c | 65 int *iter, struct vm_object *obj, vm_pindex_t pindex) in vm_domainset_iter_init() argument 79 pindex += obj->pg_color; in vm_domainset_iter_init() 81 pindex >>= VM_LEVEL_1_ORDER; in vm_domainset_iter_init() 83 pindex >>= VM_LEVEL_0_ORDER; in vm_domainset_iter_init() 86 pindex /= vm_domainset_default_stride; in vm_domainset_iter_init() 92 pindex += (((uintptr_t)obj) / sizeof(*obj)); in vm_domainset_iter_init() 93 di->di_offset = pindex; in vm_domainset_iter_init() 202 vm_pindex_t pindex, int *domain, int *req) in vm_domainset_iter_page_init() argument 215 vm_domainset_iter_init(di, dr->dr_policy, &dr->dr_iter, obj, pindex); in vm_domainset_iter_page_init() 350 vm_pindex_t pindex, int *domain, int *flags) in vm_domainset_iter_page_init() argument
|
H A D | device_pager.c | 139 vm_pindex_t pindex; in cdev_pager_allocate() local 162 pindex = OFF_TO_IDX(foff) + OFF_TO_IDX(size); in cdev_pager_allocate() 163 if (pindex > OBJ_MAX_SIZE || pindex < OFF_TO_IDX(foff) || in cdev_pager_allocate() 164 pindex < OFF_TO_IDX(size)) in cdev_pager_allocate() 183 object1 = vm_object_allocate(tp, pindex); in cdev_pager_allocate() 200 if (pindex > object->size) in cdev_pager_allocate() 201 object->size = pindex; in cdev_pager_allocate() 247 if (pindex > object->size) in cdev_pager_allocate() 248 object->size = pindex; in cdev_pager_allocate() 273 vm_radix_iter_lookup(&pages, m->pindex); in cdev_pager_free_page() [all …]
|
H A D | vm_glue.c | 230 vm_pindex_t pindex; in vm_imgact_hold_page() local 232 pindex = OFF_TO_IDX(offset); in vm_imgact_hold_page() 233 (void)vm_page_grab_valid_unlocked(&m, object, pindex, in vm_imgact_hold_page() 491 vm_pindex_t pindex; in vm_thread_stack_dispose() local 495 pindex = vm_kstack_pindex(ks, pages); in vm_thread_stack_dispose() 500 m = vm_page_lookup(obj, pindex + i); in vm_thread_stack_dispose() 586 vm_pindex_t pindex = atop(ks - VM_MIN_KERNEL_ADDRESS); in vm_kstack_pindex() local 589 return (pindex); in vm_kstack_pindex() 596 return (pindex); in vm_kstack_pindex() 598 KASSERT(pindex % (kpages + KSTACK_GUARD_PAGES) >= KSTACK_GUARD_PAGES, in vm_kstack_pindex() [all …]
|
H A D | sg_pager.c | 77 vm_pindex_t npages, pindex; in sg_pager_alloc() local 104 pindex = OFF_TO_IDX(foff) + OFF_TO_IDX(size); in sg_pager_alloc() 105 if (pindex > npages || pindex < OFF_TO_IDX(foff) || in sg_pager_alloc() 106 pindex < OFF_TO_IDX(size)) in sg_pager_alloc() 162 offset = m[0]->pindex; in sg_pager_getpages() 220 sg_pager_haspage(vm_object_t object, vm_pindex_t pindex, int *before, in sg_pager_haspage() argument
|
H A D | vnode_pager.c | 364 vnode_pager_haspage(vm_object_t object, vm_pindex_t pindex, int *before, in vnode_pager_haspage() argument 387 if (IDX_TO_OFF(pindex) >= object->un_pager.vnp.vnp_size) in vnode_pager_haspage() 394 reqblock = pindex / pagesperblock; in vnode_pager_haspage() 397 reqblock = pindex * blocksperpage; in vnode_pager_haspage() 407 poff = pindex - (reqblock * pagesperblock); in vnode_pager_haspage() 426 if (pindex + *after >= object->size) in vnode_pager_haspage() 427 *after = object->size - 1 - pindex; in vnode_pager_haspage() 696 address = IDX_TO_OFF(m->pindex) + i * bsize; in vnode_pager_input_smlfs() 774 if (IDX_TO_OFF(m->pindex) >= object->un_pager.vnp.vnp_size) { in vnode_pager_input_old() 778 if (IDX_TO_OFF(m->pindex) + size > object->un_pager.vnp.vnp_size) in vnode_pager_input_old() [all …]
|
H A D | vm_fault.c | 138 vm_pindex_t pindex; member 370 if (m->object != fs->first_object || m->pindex != fs->first_pindex) in vm_fault_soft_fast() 489 KASSERT(m != NULL && m->pindex == pidx, in vm_fault_populate_cleanup() 630 KASSERT(m != NULL && m->pindex == pidx, in vm_fault_populate() 669 m[i].pindex == fs->first_pindex) in vm_fault_populate() 674 m[i].pindex == fs->first_pindex) { in vm_fault_populate() 1106 fs->pindex = fs->first_pindex; in vm_fault_cow() 1160 fs->pindex += OFF_TO_IDX(fs->object->backing_object_offset); in vm_fault_next() 1178 fs->pindex = fs->first_pindex; in vm_fault_zerofill() 1245 if (fs->pindex >= fs->object->size) { in vm_fault_allocate() [all …]
|
H A D | vm_page.h | 238 vm_pindex_t pindex; /* offset into object (O,P) */ member 603 vm_pindex_t pindex, const char *wmesg, int allocflags); 614 vm_page_t vm_page_alloc_contig(vm_object_t object, vm_pindex_t pindex, int req, 618 vm_pindex_t pindex, int domain, int req, u_long npages, vm_paddr_t low, 631 int vm_page_grab_zero_partial(vm_object_t object, vm_pindex_t pindex, int base, 635 int vm_page_grab_pages(vm_object_t object, vm_pindex_t pindex, int allocflags, 637 int vm_page_grab_pages_unlocked(vm_object_t object, vm_pindex_t pindex, 639 int vm_page_grab_valid(vm_page_t *mp, vm_object_t object, vm_pindex_t pindex, 642 vm_pindex_t pindex, int allocflags); 689 vm_pindex_t pindex, vm_page_t mold);
|
H A D | vm_pager.c | 129 dead_pager_haspage(vm_object_t object, vm_pindex_t pindex, int *prev, int *next) in dead_pager_haspage() argument 308 KASSERT(m[i]->pindex == m[0]->pindex + i, in vm_pager_assert_in() 323 vm_pindex_t pindex = m[0]->pindex; in vm_pager_get_pages() local 341 KASSERT(m[i] == vm_page_relookup(object, pindex++), in vm_pager_get_pages() 342 ("%s: mismatch page %p pindex %ju", __func__, in vm_pager_get_pages() 343 m[i], (uintmax_t )pindex - 1)); in vm_pager_get_pages()
|
H A D | vm_reserv.h | 48 vm_page_t vm_reserv_alloc_contig(vm_object_t object, vm_pindex_t pindex, 52 vm_page_t vm_reserv_alloc_page(vm_object_t object, vm_pindex_t pindex,
|
H A D | swap_pager.h | 78 vm_pindex_t swap_pager_seek_data(vm_object_t object, vm_pindex_t pindex); 79 vm_pindex_t swap_pager_seek_hole(vm_object_t object, vm_pindex_t pindex);
|
H A D | vm_pager.h | 71 typedef boolean_t pgo_can_alloc_page_t(vm_object_t object, vm_pindex_t pindex); 276 vm_pager_can_alloc_page(vm_object_t object, vm_pindex_t pindex) in vm_pager_can_alloc_page() argument 281 return (method != NULL ? method(object, pindex) : true); in vm_pager_can_alloc_page() 312 boolean_t (*phys_pg_haspage)(vm_object_t obj, vm_pindex_t pindex,
|
/freebsd/sys/dev/drm2/ |
H A D | drm_scatter.c | 58 vm_pindex_t pindex; in drm_sg_alloc() local 90 for (pindex = 0; pindex < entry->pages; pindex++) { in drm_sg_alloc() 91 entry->busaddr[pindex] = in drm_sg_alloc() 92 vtophys((uintptr_t)entry->vaddr + IDX_TO_OFF(pindex)); in drm_sg_alloc()
|
/freebsd/sys/contrib/openzfs/module/os/freebsd/zfs/ |
H A D | dmu_os.c | 102 ASSERT3U(ptoa((*ma)->pindex), ==, in dmu_write_pages() 138 ASSERT3U(ma[0]->pindex + count - 1, ==, ma[count - 1]->pindex); in dmu_read_pages() 141 err = dmu_buf_hold_array(os, object, IDX_TO_OFF(ma[0]->pindex), in dmu_read_pages() 161 m = vm_page_grab_unlocked(vmobj, ma[0]->pindex - 1 - i, in dmu_read_pages() 174 bufoff = IDX_TO_OFF(m->pindex) % db->db_size; in dmu_read_pages() 188 bufoff = IDX_TO_OFF(ma[0]->pindex) % db->db_size; in dmu_read_pages() 205 ASSERT3U(IDX_TO_OFF(m->pindex) + pgoff, ==, in dmu_read_pages() 281 m = vm_page_grab_unlocked(vmobj, ma[count - 1]->pindex + 1 + i, in dmu_read_pages() 294 bufoff = IDX_TO_OFF(m->pindex) % db->db_size; in dmu_read_pages()
|
/freebsd/contrib/ncurses/menu/ |
H A D | m_driver.c | 47 (menu)->pattern[--((menu)->pindex)] = '\0' 51 { (menu)->pattern[((menu)->pindex)++] = (char) (ch);\ 52 (menu)->pattern[(menu)->pindex] = '\0'; } 135 if ((menu->pindex + 1) > menu->namelen) in _nc_Match_Next_Character_In_Item_Name() 191 if (ch && ch != BS && menu->pindex > 0) in _nc_Match_Next_Character_In_Item_Name() 390 if (menu->pindex > 0) in menu_driver()
|
H A D | m_cursor.c | 98 if ((menu->opt & O_SHOWMATCH) && (menu->pindex > 0)) in pos_menu_cursor() 99 x += (menu->pindex + menu->marklen - 1); in pos_menu_cursor()
|
/freebsd/usr.sbin/services_mkdb/ |
H A D | services_mkdb.c | 238 size_t linecap, line, pindex; in parseservices() local 309 pindex = getprotoindex(sl, proto); in parseservices() 310 if (svc[pnum][pindex] == NULL) in parseservices() 311 s = svc[pnum][pindex] = sl_init(); in parseservices() 313 s = svc[pnum][pindex]; in parseservices()
|
/freebsd/sys/x86/iommu/ |
H A D | amd_idpgtbl.c | 164 KASSERT(m->pindex != 0, in amdiommu_pgtbl_map_pte() 167 iommu_pgfree(domain->pgtbl_obj, m->pindex, in amdiommu_pgtbl_map_pte() 318 KASSERT(m->pindex != 0, in amdiommu_unmap_clear_pte() 321 iommu_pgfree(domain->pgtbl_obj, m->pindex, flags, entry); in amdiommu_unmap_clear_pte()
|