Home
last modified time | relevance | path

Searched refs:min_order (Results 1 – 11 of 11) sorted by relevance

/linux/mm/
H A Dfail_page_alloc.c12 u32 min_order; member
17 .min_order = 1,
30 if (order < fail_page_alloc.min_order) in should_fail_alloc_page()
62 debugfs_create_u32("min-order", mode, dir, &fail_page_alloc.min_order); in fail_page_alloc_debugfs()
H A Dreadahead.c473 unsigned int min_order = mapping_min_folio_order(mapping); in page_cache_ra_order()
491 new_order = max(new_order, min_order); in page_cache_ra_order()
499 * If the new_order is greater than min_order and index is in page_cache_ra_order()
501 * aligned to new_order should also be aligned to min_order. in page_cache_ra_order()
513 while (order > min_order && index + (1UL << order) - 1 > limit) in page_cache_ra_order()
774 unsigned int min_order = mapping_min_folio_order(mapping); in readahead_expand()
791 folio = ractl_alloc_folio(ractl, gfp_mask, min_order); in readahead_expand()
820 folio = ractl_alloc_folio(ractl, gfp_mask, min_order); in readahead_expand()
470 unsigned int min_order = mapping_min_folio_order(mapping); page_cache_ra_order() local
771 unsigned int min_order = mapping_min_folio_order(mapping); readahead_expand() local
H A Dtruncate.c181 unsigned long min_order) in try_folio_split_or_unmap() argument
189 ret = try_folio_split_to_order(folio, split_at, min_order); in try_folio_split_or_unmap()
223 unsigned int min_order; in truncate_inode_partial_folio() local
253 min_order = mapping_min_folio_order(folio->mapping); in truncate_inode_partial_folio()
255 if (!try_folio_split_or_unmap(folio, split_at, min_order)) { in truncate_inode_partial_folio()
282 try_folio_split_or_unmap(folio2, split_at2, min_order); in truncate_inode_partial_folio()
H A Dfilemap.c1983 unsigned int min_order = mapping_min_folio_order(mapping); in __filemap_get_folio_mpol() local
1984 unsigned int order = max(min_order, FGF_GET_ORDER(fgp_flags)); in __filemap_get_folio_mpol()
2009 if (order > min_order) in __filemap_get_folio_mpol()
2026 } while (order-- > min_order); in __filemap_get_folio_mpol()
2605 unsigned int min_order = mapping_min_folio_order(mapping); in filemap_create_folio() local
2611 folio = filemap_alloc_folio(mapping_gfp_mask(mapping), min_order, NULL); in filemap_create_folio()
2631 index = (iocb->ki_pos >> (PAGE_SHIFT + min_order)) << min_order; in filemap_create_folio()
H A Dhuge_memory.c3996 unsigned int min_order; in __folio_split() local
4000 min_order = mapping_min_folio_order(folio->mapping); in __folio_split()
4001 if (new_order < min_order) { in __folio_split()
4709 unsigned int min_order; in split_huge_pages_in_file() local
4721 min_order = mapping_min_folio_order(mapping); in split_huge_pages_in_file()
4722 target_order = max(new_order, min_order); in split_huge_pages_in_file()
H A Dslub.c7347 unsigned int min_order, unsigned int max_order, in calc_slab_order() argument
7352 for (order = min_order; order <= max_order; order++) { in calc_slab_order()
7371 unsigned int min_order; in calculate_order() local
7393 min_order = max_t(unsigned int, slub_min_order, in calculate_order()
7395 if (order_objects(min_order, size) > MAX_OBJS_PER_PAGE) in calculate_order()
7414 order = calc_slab_order(size, min_order, slub_max_order, in calculate_order()
H A Dpage_alloc.c2392 int min_order = order; in __rmqueue_claim() local
2402 min_order = pageblock_order; in __rmqueue_claim()
2409 for (current_order = MAX_PAGE_ORDER; current_order >= min_order; in __rmqueue_claim()
/linux/drivers/gpu/drm/
H A Ddrm_buddy.c228 unsigned int min_order) in __force_merge() argument
233 if (!min_order) in __force_merge()
236 if (min_order > mm->max_order) in __force_merge()
240 for (i = min_order - 1; i >= 0; i--) { in __force_merge()
278 if (order >= min_order) in __force_merge()
1107 unsigned int min_order, order; in drm_buddy_alloc_blocks() local
1152 min_order = ilog2(min_block_size) - ilog2(mm->chunk_size); in drm_buddy_alloc_blocks()
1166 BUG_ON(order < min_order); in drm_buddy_alloc_blocks()
1176 if (order-- == min_order) { in drm_buddy_alloc_blocks()
1179 !__force_merge(mm, start, end, min_order)) { in drm_buddy_alloc_blocks()
[all …]
/linux/fs/ext4/
H A Dmballoc.c1066 int i, order, min_order; in ext4_mb_scan_groups_best_avail() local
1079 min_order = order - sbi->s_mb_best_avail_max_trim_order; in ext4_mb_scan_groups_best_avail()
1080 if (min_order < 0) in ext4_mb_scan_groups_best_avail()
1081 min_order = 0; in ext4_mb_scan_groups_best_avail()
1089 if (1 << min_order < num_stripe_clusters) in ext4_mb_scan_groups_best_avail()
1094 min_order = fls(num_stripe_clusters) - 1; in ext4_mb_scan_groups_best_avail()
1097 if (1 << min_order < ac->ac_o_ex.fe_len) in ext4_mb_scan_groups_best_avail()
1098 min_order = fls(ac->ac_o_ex.fe_len); in ext4_mb_scan_groups_best_avail()
1103 for (i = order; i >= min_order; i--) { in ext4_mb_scan_groups_best_avail()
H A Dinode.c5117 u16 min_order, max_order; in ext4_set_inode_mapping_order() local
5123 min_order = EXT4_SB(sb)->s_min_folio_order; in ext4_set_inode_mapping_order()
5124 if (!min_order && !S_ISREG(inode->i_mode)) in ext4_set_inode_mapping_order()
5128 max_order = min_order; in ext4_set_inode_mapping_order()
5130 mapping_set_folio_order_range(inode->i_mapping, min_order, max_order); in ext4_set_inode_mapping_order()
/linux/drivers/md/bcache/
H A Dbtree.c606 static int mca_reap(struct btree *b, unsigned int min_order, bool flush) in mca_reap() argument
618 if (b->keys.page_order < min_order) in mca_reap()