Lines Matching refs:fl
188 static void vm_phys_split_pages(vm_page_t m, int oind, struct vm_freelist *fl,
279 struct vm_freelist *fl; in sysctl_vm_phys_free() local
302 fl = vm_phys_free_queues[dom][flind][pind]; in sysctl_vm_phys_free()
304 fl[oind].lcnt); in sysctl_vm_phys_free()
393 vm_freelist_add(struct vm_freelist *fl, vm_page_t m, int order, int pool, in vm_freelist_add() argument
410 TAILQ_INSERT_TAIL(&fl[order].pl, m, plinks.q); in vm_freelist_add()
412 TAILQ_INSERT_HEAD(&fl[order].pl, m, plinks.q); in vm_freelist_add()
413 fl[order].lcnt++; in vm_freelist_add()
417 vm_freelist_rem(struct vm_freelist *fl, vm_page_t m, int order) in vm_freelist_rem() argument
420 TAILQ_REMOVE(&fl[order].pl, m, plinks.q); in vm_freelist_rem()
421 fl[order].lcnt--; in vm_freelist_rem()
537 struct vm_freelist *fl; in vm_phys_init() local
659 fl = vm_phys_free_queues[dom][flind][pind]; in vm_phys_init()
661 TAILQ_INIT(&fl[oind].pl); in vm_phys_init()
716 vm_phys_split_pages(vm_page_t m, int oind, struct vm_freelist *fl, int order, in vm_phys_split_pages() argument
727 vm_freelist_add(fl, m_buddy, oind, pool, tail); in vm_phys_split_pages()
732 vm_phys_enq_chunk(struct vm_freelist *fl, vm_page_t m, int order, int pool, in vm_phys_enq_chunk() argument
738 vm_freelist_add(fl, m, order, pool, tail); in vm_phys_enq_chunk()
769 vm_phys_enq_beg(vm_page_t m, u_int npages, struct vm_freelist *fl, int pool, in vm_phys_enq_beg() argument
786 vm_phys_enq_chunk(fl, m, order, pool, tail); in vm_phys_enq_beg()
806 vm_phys_enq_range(vm_page_t m, u_int npages, struct vm_freelist *fl, int pool, in vm_phys_enq_range() argument
821 vm_phys_enq_chunk(fl, m, order, pool, tail); in vm_phys_enq_range()
871 struct vm_freelist *alt, *fl; in vm_phys_alloc_npages() local
887 fl = vm_phys_free_queues[domain][flind][pool]; in vm_phys_alloc_npages()
889 while ((m = TAILQ_FIRST(&fl[oind].pl)) != NULL) { in vm_phys_alloc_npages()
890 vm_freelist_rem(fl, m, oind); in vm_phys_alloc_npages()
900 vm_phys_enq_range(m, avail - i, fl, in vm_phys_alloc_npages()
925 fl, pool, 1); in vm_phys_alloc_npages()
945 struct vm_freelist *alt, *fl; in vm_phys_alloc_freelist_pages() local
966 fl = &vm_phys_free_queues[domain][flind][pool][0]; in vm_phys_alloc_freelist_pages()
968 m = TAILQ_FIRST(&fl[oind].pl); in vm_phys_alloc_freelist_pages()
970 vm_freelist_rem(fl, m, oind); in vm_phys_alloc_freelist_pages()
972 vm_phys_split_pages(m, oind, fl, order, pool, 1); in vm_phys_alloc_freelist_pages()
991 vm_phys_split_pages(m, oind, fl, order, pool, 1); in vm_phys_alloc_freelist_pages()
1237 struct vm_freelist *fl; in vm_phys_free_pages() local
1260 fl = (*seg->free_queues)[m_buddy->pool]; in vm_phys_free_pages()
1261 vm_freelist_rem(fl, m_buddy, order); in vm_phys_free_pages()
1268 fl = (*seg->free_queues)[pool]; in vm_phys_free_pages()
1269 vm_freelist_add(fl, m, order, pool, 1); in vm_phys_free_pages()
1284 struct vm_freelist *fl; in vm_phys_lazy_init_domain() local
1306 fl = vm_phys_free_queues[domain][flind][pind]; in vm_phys_lazy_init_domain()
1308 if (atomic_load_int(&fl[oind].lcnt) == 0) in vm_phys_lazy_init_domain()
1310 while ((m = TAILQ_FIRST(&fl[oind].pl)) != NULL) { in vm_phys_lazy_init_domain()
1316 vm_freelist_rem(fl, m, oind); in vm_phys_lazy_init_domain()
1383 struct vm_freelist *fl; in vm_phys_enqueue_contig() local
1395 fl = (*seg->free_queues)[pool]; in vm_phys_enqueue_contig()
1402 m = vm_phys_enq_range(m, roundup2(lo, 1 << order) - lo, fl, in vm_phys_enqueue_contig()
1412 vm_phys_enq_chunk(fl, m, order, pool, 1); in vm_phys_enqueue_contig()
1416 vm_phys_enq_beg(m, m_end - m, fl, pool, 1); in vm_phys_enqueue_contig()
1505 struct vm_freelist *fl; in vm_phys_unfree_page() local
1551 fl = (*seg->free_queues)[pool]; in vm_phys_unfree_page()
1553 vm_freelist_rem(fl, m_set, order); in vm_phys_unfree_page()
1563 vm_freelist_add(fl, m_tmp, order, pool, 0); in vm_phys_unfree_page()
1575 vm_phys_find_freelist_contig(struct vm_freelist *fl, u_long npages, in vm_phys_find_freelist_contig() argument
1595 TAILQ_FOREACH(m, &fl[max_order].pl, plinks.q) { in vm_phys_find_freelist_contig()
1653 struct vm_freelist *fl; in vm_phys_find_queues_contig() local
1667 fl = (*queues)[pind]; in vm_phys_find_queues_contig()
1668 TAILQ_FOREACH(m_ret, &fl[oind].pl, plinks.q) { in vm_phys_find_queues_contig()
1687 fl = (*queues)[pind]; in vm_phys_find_queues_contig()
1688 m_ret = vm_phys_find_freelist_contig(fl, npages, in vm_phys_find_queues_contig()
1712 struct vm_freelist *fl; in vm_phys_alloc_contig() local
1761 fl = (*queues)[m->pool]; in vm_phys_alloc_contig()
1763 vm_freelist_rem(fl, m, oind); in vm_phys_alloc_contig()
1767 fl = (*queues)[VM_FREEPOOL_DEFAULT]; in vm_phys_alloc_contig()
1768 vm_phys_enq_range(&m_run[npages], m - &m_run[npages], fl, in vm_phys_alloc_contig()
2066 struct vm_freelist *fl; in DB_SHOW_COMMAND_FLAGS() local
2085 fl = vm_phys_free_queues[dom][flind][pind]; in DB_SHOW_COMMAND_FLAGS()
2086 db_printf(" | %6.6d", fl[oind].lcnt); in DB_SHOW_COMMAND_FLAGS()