| /linux/mm/ |
| H A D | fail_page_alloc.c | 12 u32 min_order; member 17 .min_order = 1, 30 if (order < fail_page_alloc.min_order) in should_fail_alloc_page() 62 debugfs_create_u32("min-order", mode, dir, &fail_page_alloc.min_order); in fail_page_alloc_debugfs()
|
| H A D | readahead.c | 473 unsigned int min_order = mapping_min_folio_order(mapping); in page_cache_ra_order() 491 new_order = max(new_order, min_order); in page_cache_ra_order() 499 * If the new_order is greater than min_order and index is in page_cache_ra_order() 501 * aligned to new_order should also be aligned to min_order. in page_cache_ra_order() 513 while (order > min_order && index + (1UL << order) - 1 > limit) in page_cache_ra_order() 774 unsigned int min_order = mapping_min_folio_order(mapping); in readahead_expand() 791 folio = ractl_alloc_folio(ractl, gfp_mask, min_order); in readahead_expand() 820 folio = ractl_alloc_folio(ractl, gfp_mask, min_order); in readahead_expand() 470 unsigned int min_order = mapping_min_folio_order(mapping); page_cache_ra_order() local 771 unsigned int min_order = mapping_min_folio_order(mapping); readahead_expand() local
|
| H A D | truncate.c | 181 unsigned long min_order) in try_folio_split_or_unmap() argument 189 ret = try_folio_split_to_order(folio, split_at, min_order); in try_folio_split_or_unmap() 223 unsigned int min_order; in truncate_inode_partial_folio() local 253 min_order = mapping_min_folio_order(folio->mapping); in truncate_inode_partial_folio() 255 if (!try_folio_split_or_unmap(folio, split_at, min_order)) { in truncate_inode_partial_folio() 282 try_folio_split_or_unmap(folio2, split_at2, min_order); in truncate_inode_partial_folio()
|
| H A D | filemap.c | 1978 unsigned int min_order = mapping_min_folio_order(mapping); in __filemap_get_folio_mpol() local 1979 unsigned int order = max(min_order, FGF_GET_ORDER(fgp_flags)); in __filemap_get_folio_mpol() 2004 if (order > min_order) in __filemap_get_folio_mpol() 2021 } while (order-- > min_order); in __filemap_get_folio_mpol() 2600 unsigned int min_order = mapping_min_folio_order(mapping); in filemap_create_folio() local 2606 folio = filemap_alloc_folio(mapping_gfp_mask(mapping), min_order, NULL); in filemap_create_folio() 2626 index = (iocb->ki_pos >> (PAGE_SHIFT + min_order)) << min_order; in filemap_create_folio()
|
| H A D | huge_memory.c | 3990 unsigned int min_order; in __folio_split() local 3994 min_order = mapping_min_folio_order(folio->mapping); in __folio_split() 3995 if (new_order < min_order) { in __folio_split() 4703 unsigned int min_order; in split_huge_pages_in_file() local 4715 min_order = mapping_min_folio_order(mapping); in split_huge_pages_in_file() 4716 target_order = max(new_order, min_order); in split_huge_pages_in_file()
|
| H A D | slub.c | 7306 unsigned int min_order, unsigned int max_order, in calc_slab_order() argument 7311 for (order = min_order; order <= max_order; order++) { in calc_slab_order() 7330 unsigned int min_order; in calculate_order() local 7352 min_order = max_t(unsigned int, slub_min_order, in calculate_order() 7354 if (order_objects(min_order, size) > MAX_OBJS_PER_PAGE) in calculate_order() 7373 order = calc_slab_order(size, min_order, slub_max_order, in calculate_order()
|
| H A D | page_alloc.c | 2392 int min_order = order; in __rmqueue_claim() local 2402 min_order = pageblock_order; in __rmqueue_claim() 2409 for (current_order = MAX_PAGE_ORDER; current_order >= min_order; in __rmqueue_claim()
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_buddy.c | 228 unsigned int min_order) in __force_merge() argument 233 if (!min_order) in __force_merge() 236 if (min_order > mm->max_order) in __force_merge() 240 for (i = min_order - 1; i >= 0; i--) { in __force_merge() 278 if (order >= min_order) in __force_merge() 1107 unsigned int min_order, order; in drm_buddy_alloc_blocks() local 1152 min_order = ilog2(min_block_size) - ilog2(mm->chunk_size); in drm_buddy_alloc_blocks() 1166 BUG_ON(order < min_order); in drm_buddy_alloc_blocks() 1176 if (order-- == min_order) { in drm_buddy_alloc_blocks() 1179 !__force_merge(mm, start, end, min_order)) { in drm_buddy_alloc_blocks() [all …]
|
| /linux/fs/ext4/ |
| H A D | mballoc.c | 1066 int i, order, min_order; in ext4_mb_scan_groups_best_avail() local 1079 min_order = order - sbi->s_mb_best_avail_max_trim_order; in ext4_mb_scan_groups_best_avail() 1080 if (min_order < 0) in ext4_mb_scan_groups_best_avail() 1081 min_order = 0; in ext4_mb_scan_groups_best_avail() 1089 if (1 << min_order < num_stripe_clusters) in ext4_mb_scan_groups_best_avail() 1094 min_order = fls(num_stripe_clusters) - 1; in ext4_mb_scan_groups_best_avail() 1097 if (1 << min_order < ac->ac_o_ex.fe_len) in ext4_mb_scan_groups_best_avail() 1098 min_order = fls(ac->ac_o_ex.fe_len); in ext4_mb_scan_groups_best_avail() 1103 for (i = order; i >= min_order; i--) { in ext4_mb_scan_groups_best_avail()
|
| /linux/drivers/md/bcache/ |
| H A D | btree.c | 606 static int mca_reap(struct btree *b, unsigned int min_order, bool flush) in mca_reap() argument 618 if (b->keys.page_order < min_order) in mca_reap()
|