Lines Matching refs:fl

189 static void vm_phys_split_pages(vm_page_t m, int oind, struct vm_freelist *fl,
280 struct vm_freelist *fl; in sysctl_vm_phys_free() local
303 fl = vm_phys_free_queues[dom][flind][pind]; in sysctl_vm_phys_free()
305 fl[oind].lcnt); in sysctl_vm_phys_free()
394 vm_freelist_add(struct vm_freelist *fl, vm_page_t m, int order, int pool, in vm_freelist_add() argument
401 TAILQ_INSERT_TAIL(&fl[order].pl, m, listq); in vm_freelist_add()
403 TAILQ_INSERT_HEAD(&fl[order].pl, m, listq); in vm_freelist_add()
404 fl[order].lcnt++; in vm_freelist_add()
408 vm_freelist_rem(struct vm_freelist *fl, vm_page_t m, int order) in vm_freelist_rem() argument
411 TAILQ_REMOVE(&fl[order].pl, m, listq); in vm_freelist_rem()
412 fl[order].lcnt--; in vm_freelist_rem()
528 struct vm_freelist *fl; in vm_phys_init() local
650 fl = vm_phys_free_queues[dom][flind][pind]; in vm_phys_init()
652 TAILQ_INIT(&fl[oind].pl); in vm_phys_init()
707 vm_phys_split_pages(vm_page_t m, int oind, struct vm_freelist *fl, int order, in vm_phys_split_pages() argument
718 vm_freelist_add(fl, m_buddy, oind, pool, tail); in vm_phys_split_pages()
723 vm_phys_enq_chunk(struct vm_freelist *fl, vm_page_t m, int order, int pool, in vm_phys_enq_chunk() argument
729 vm_freelist_add(fl, m, order, pool, tail); in vm_phys_enq_chunk()
760 vm_phys_enq_beg(vm_page_t m, u_int npages, struct vm_freelist *fl, int pool, in vm_phys_enq_beg() argument
777 vm_phys_enq_chunk(fl, m, order, pool, tail); in vm_phys_enq_beg()
797 vm_phys_enq_range(vm_page_t m, u_int npages, struct vm_freelist *fl, int pool, in vm_phys_enq_range() argument
812 vm_phys_enq_chunk(fl, m, order, pool, tail); in vm_phys_enq_range()
862 struct vm_freelist *alt, *fl; in vm_phys_alloc_npages() local
878 fl = vm_phys_free_queues[domain][flind][pool]; in vm_phys_alloc_npages()
880 while ((m = TAILQ_FIRST(&fl[oind].pl)) != NULL) { in vm_phys_alloc_npages()
881 vm_freelist_rem(fl, m, oind); in vm_phys_alloc_npages()
891 vm_phys_enq_range(m, avail - i, fl, in vm_phys_alloc_npages()
916 fl, pool, 1); in vm_phys_alloc_npages()
936 struct vm_freelist *alt, *fl; in vm_phys_alloc_freelist_pages() local
957 fl = &vm_phys_free_queues[domain][flind][pool][0]; in vm_phys_alloc_freelist_pages()
959 m = TAILQ_FIRST(&fl[oind].pl); in vm_phys_alloc_freelist_pages()
961 vm_freelist_rem(fl, m, oind); in vm_phys_alloc_freelist_pages()
963 vm_phys_split_pages(m, oind, fl, order, pool, 1); in vm_phys_alloc_freelist_pages()
982 vm_phys_split_pages(m, oind, fl, order, pool, 1); in vm_phys_alloc_freelist_pages()
1228 struct vm_freelist *fl; in vm_phys_free_pages() local
1251 fl = (*seg->free_queues)[m_buddy->pool]; in vm_phys_free_pages()
1252 vm_freelist_rem(fl, m_buddy, order); in vm_phys_free_pages()
1259 fl = (*seg->free_queues)[pool]; in vm_phys_free_pages()
1260 vm_freelist_add(fl, m, order, pool, 1); in vm_phys_free_pages()
1275 struct vm_freelist *fl; in vm_phys_lazy_init_domain() local
1297 fl = vm_phys_free_queues[domain][flind][pind]; in vm_phys_lazy_init_domain()
1299 if (atomic_load_int(&fl[oind].lcnt) == 0) in vm_phys_lazy_init_domain()
1301 while ((m = TAILQ_FIRST(&fl[oind].pl)) != NULL) { in vm_phys_lazy_init_domain()
1307 vm_freelist_rem(fl, m, oind); in vm_phys_lazy_init_domain()
1374 struct vm_freelist *fl; in vm_phys_enqueue_contig() local
1386 fl = (*seg->free_queues)[pool]; in vm_phys_enqueue_contig()
1393 m = vm_phys_enq_range(m, roundup2(lo, 1 << order) - lo, fl, in vm_phys_enqueue_contig()
1403 vm_phys_enq_chunk(fl, m, order, pool, 1); in vm_phys_enqueue_contig()
1407 vm_phys_enq_beg(m, m_end - m, fl, pool, 1); in vm_phys_enqueue_contig()
1496 struct vm_freelist *fl; in vm_phys_unfree_page() local
1542 fl = (*seg->free_queues)[pool]; in vm_phys_unfree_page()
1544 vm_freelist_rem(fl, m_set, order); in vm_phys_unfree_page()
1554 vm_freelist_add(fl, m_tmp, order, pool, 0); in vm_phys_unfree_page()
1566 vm_phys_find_freelist_contig(struct vm_freelist *fl, u_long npages, in vm_phys_find_freelist_contig() argument
1586 TAILQ_FOREACH(m, &fl[max_order].pl, listq) { in vm_phys_find_freelist_contig()
1644 struct vm_freelist *fl; in vm_phys_find_queues_contig() local
1658 fl = (*queues)[pind]; in vm_phys_find_queues_contig()
1659 TAILQ_FOREACH(m_ret, &fl[oind].pl, listq) { in vm_phys_find_queues_contig()
1678 fl = (*queues)[pind]; in vm_phys_find_queues_contig()
1679 m_ret = vm_phys_find_freelist_contig(fl, npages, in vm_phys_find_queues_contig()
1703 struct vm_freelist *fl; in vm_phys_alloc_contig() local
1752 fl = (*queues)[m->pool]; in vm_phys_alloc_contig()
1754 vm_freelist_rem(fl, m, oind); in vm_phys_alloc_contig()
1758 fl = (*queues)[VM_FREEPOOL_DEFAULT]; in vm_phys_alloc_contig()
1759 vm_phys_enq_range(&m_run[npages], m - &m_run[npages], fl, in vm_phys_alloc_contig()
2057 struct vm_freelist *fl; in DB_SHOW_COMMAND_FLAGS() local
2076 fl = vm_phys_free_queues[dom][flind][pind]; in DB_SHOW_COMMAND_FLAGS()
2077 db_printf(" | %6.6d", fl[oind].lcnt); in DB_SHOW_COMMAND_FLAGS()