Lines Matching refs:new_order

3484 		int new_order)  in __split_folio_to_order()  argument
3487 const bool handle_hwpoison = folio_test_has_hwpoisoned(folio) && new_order; in __split_folio_to_order()
3488 long new_nr_pages = 1 << new_order; in __split_folio_to_order()
3565 if (new_order) { in __split_folio_to_order()
3566 prep_compound_page(new_head, new_order); in __split_folio_to_order()
3581 if (new_order) in __split_folio_to_order()
3582 folio_set_order(folio, new_order); in __split_folio_to_order()
3628 static int __split_unmapped_folio(struct folio *folio, int new_order, in __split_unmapped_folio() argument
3634 int start_order = split_type == SPLIT_TYPE_UNIFORM ? new_order : old_order - 1; in __split_unmapped_folio()
3643 split_order >= new_order; in __split_unmapped_folio()
3708 int folio_check_splittable(struct folio *folio, unsigned int new_order, in folio_check_splittable() argument
3724 if (new_order == 1) in folio_check_splittable()
3726 } else if (split_type == SPLIT_TYPE_NON_UNIFORM || new_order) { in folio_check_splittable()
3758 if ((split_type == SPLIT_TYPE_NON_UNIFORM || new_order) && folio_test_swapcache(folio)) { in folio_check_splittable()
3779 static int __folio_freeze_and_split_unmapped(struct folio *folio, unsigned int new_order, in __folio_freeze_and_split_unmapped() argument
3820 new_order < HPAGE_PMD_ORDER) { in __folio_freeze_and_split_unmapped()
3845 ret = __split_unmapped_folio(folio, new_order, split_at, xas, in __folio_freeze_and_split_unmapped()
3944 static int __folio_split(struct folio *folio, unsigned int new_order, in __folio_split() argument
3968 if (new_order >= old_order) { in __folio_split()
3973 ret = folio_check_splittable(folio, new_order, split_type); in __folio_split()
4001 if (new_order < min_order) { in __folio_split()
4015 xas_set_order(&xas, folio->index, new_order); in __folio_split()
4064 ret = __folio_freeze_and_split_unmapped(folio, new_order, split_at, &xas, mapping, in __folio_split()
4137 int folio_split_unmapped(struct folio *folio, unsigned int new_order) in folio_split_unmapped() argument
4150 ret = __folio_freeze_and_split_unmapped(folio, new_order, &folio->page, NULL, in folio_split_unmapped()
4205 unsigned int new_order) in __split_huge_page_to_list_to_order() argument
4209 return __folio_split(folio, new_order, &folio->page, page, list, in __split_huge_page_to_list_to_order()
4237 int folio_split(struct folio *folio, unsigned int new_order, in folio_split() argument
4240 return __folio_split(folio, new_order, split_at, &folio->page, list, in folio_split()
4584 unsigned long vaddr_end, unsigned int new_order, in split_huge_pages_pid() argument
4612 pid, vaddr_start, vaddr_end, new_order, in_folio_offset); in split_huge_pages_pid()
4624 unsigned int target_order = new_order; in split_huge_pages_pid()
4644 target_order = max(new_order, in split_huge_pages_pid()
4701 pgoff_t off_end, unsigned int new_order, in split_huge_pages_in_file() argument
4718 file_path, off_start, off_end, new_order, in_folio_offset); in split_huge_pages_in_file()
4722 target_order = max(new_order, min_order); in split_huge_pages_in_file()
4782 unsigned int new_order = 0; in split_huge_pages_write() local
4813 &new_order, &in_folio_offset); in split_huge_pages_write()
4819 new_order, in_folio_offset); in split_huge_pages_write()
4827 &vaddr_end, &new_order, &in_folio_offset); in split_huge_pages_write()
4837 ret = split_huge_pages_pid(pid, vaddr_start, vaddr_end, new_order, in split_huge_pages_write()