Lines Matching defs:vmemmap_pages
29 * @vmemmap_pages: the list head of the vmemmap pages that can be freed
40 struct list_head *vmemmap_pages;
232 list_add(&page->lru, walk->vmemmap_pages);
264 page = list_first_entry(walk->vmemmap_pages, struct page, lru);
312 * @vmemmap_pages: list to deposit vmemmap pages to be freed. It is callers
320 struct list_head *vmemmap_pages,
327 .vmemmap_pages = vmemmap_pages,
345 list_add(&walk.reuse_page->lru, vmemmap_pages);
368 * vmemmap_pages contains pages from the previous
376 .vmemmap_pages = vmemmap_pages,
412 * to the page which is from the @vmemmap_pages
426 LIST_HEAD(vmemmap_pages);
430 .vmemmap_pages = &vmemmap_pages,
437 if (alloc_vmemmap_page_list(start, end, &vmemmap_pages))
565 struct list_head *vmemmap_pages,
602 * mapping the range to vmemmap_pages list so that they can be freed by
606 vmemmap_pages, flags);
627 LIST_HEAD(vmemmap_pages);
629 __hugetlb_vmemmap_optimize_folio(h, folio, &vmemmap_pages, VMEMMAP_SYNCHRONIZE_RCU);
630 free_vmemmap_page_list(&vmemmap_pages);
658 LIST_HEAD(vmemmap_pages);
711 ret = __hugetlb_vmemmap_optimize_folio(h, folio, &vmemmap_pages, flags);
723 if (ret == -ENOMEM && !list_empty(&vmemmap_pages)) {
725 free_vmemmap_page_list(&vmemmap_pages);
726 INIT_LIST_HEAD(&vmemmap_pages);
727 __hugetlb_vmemmap_optimize_folio(h, folio, &vmemmap_pages, flags);
733 free_vmemmap_page_list(&vmemmap_pages);