| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_mr.c | 92 for (o = order; o <= buddy->max_order; ++o) in mthca_buddy_alloc() 94 m = 1 << (buddy->max_order - o); in mthca_buddy_alloc() 140 static int mthca_buddy_init(struct mthca_buddy *buddy, int max_order) in mthca_buddy_init() argument 144 buddy->max_order = max_order; in mthca_buddy_init() 147 buddy->bits = kcalloc(buddy->max_order + 1, sizeof(*buddy->bits), in mthca_buddy_init() 149 buddy->num_free = kcalloc((buddy->max_order + 1), sizeof *buddy->num_free, in mthca_buddy_init() 154 for (i = 0; i <= buddy->max_order; ++i) { in mthca_buddy_init() 155 buddy->bits[i] = bitmap_zalloc(1 << (buddy->max_order - i), in mthca_buddy_init() 161 __set_bit(0, buddy->bits[buddy->max_order]); in mthca_buddy_init() 162 buddy->num_free[buddy->max_order] = 1; in mthca_buddy_init() [all …]
|
| H A D | mthca_dev.h | 206 int max_order; member
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 760 bool *is_prepared, int *max_order) in __kvm_gmem_get_pfn() argument 787 if (max_order) in __kvm_gmem_get_pfn() 788 *max_order = 0; in __kvm_gmem_get_pfn() 796 int *max_order) in kvm_gmem_get_pfn() argument 807 folio = __kvm_gmem_get_pfn(file, slot, index, pfn, &is_prepared, max_order); in kvm_gmem_get_pfn() 832 int ret = 0, max_order; in kvm_gmem_populate() local 851 for (i = 0; i < npages; i += (1 << max_order)) { in kvm_gmem_populate() 863 folio = __kvm_gmem_get_pfn(file, slot, index, &pfn, &is_prepared, &max_order); in kvm_gmem_populate() 877 WARN_ON(!IS_ALIGNED(gfn, 1 << max_order) || in kvm_gmem_populate() 878 (npages - i) < (1 << max_order)); in kvm_gmem_populate() [all …]
|
| /linux/tools/testing/radix-tree/ |
| H A D | iteration_check.c | 19 static int max_order; variable 29 for (order = max_order; order >= 0; order--) { in my_item_insert() 172 max_order = order; in iteration_test()
|
| /linux/mm/ |
| H A D | workingset.c | 785 unsigned int max_order; in workingset_init() local 797 max_order = fls_long(totalram_pages() - 1); in workingset_init() 798 if (max_order > timestamp_bits) in workingset_init() 799 bucket_order = max_order - timestamp_bits; in workingset_init() 801 timestamp_bits, max_order, bucket_order); in workingset_init()
|
| H A D | slub.c | 7580 unsigned int min_order, unsigned int max_order, in calc_slab_order() 7585 for (order = min_order; order <= max_order; order++) { in calc_slab_order() 7578 calc_slab_order(unsigned int size,unsigned int min_order,unsigned int max_order,unsigned int fract_leftover) calc_slab_order() argument
|
| /linux/drivers/dma-buf/heaps/ |
| H A D | system_heap.c | 319 unsigned int max_order) in alloc_largest_available() argument 327 if (max_order < orders[i]) in alloc_largest_available() 346 unsigned int max_order = orders[0]; in system_heap_allocate() local 375 page = alloc_largest_available(size_remaining, max_order); in system_heap_allocate() 381 max_order = compound_order(page); in system_heap_allocate()
|
| /linux/lib/ |
| H A D | test_xarray.c | 178 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 8 : 1; in check_xa_mark_1() local 207 for (order = 2; order < max_order; order++) { in check_xa_mark_1() 329 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 15 : 1; in check_xa_shrink() local 353 for (order = 0; order < max_order; order++) { in check_xa_shrink() 654 unsigned int max_order = (sizeof(long) == 4) ? 30 : 60; in check_multi_store() local 698 for (i = 0; i < max_order; i++) { in check_multi_store() 699 for (j = 0; j < max_order; j++) { in check_multi_store() 703 for (k = 0; k < max_order; k++) { in check_multi_store() 894 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 20 : 1; in check_multi_store_advanced() local 902 for (i = 0; i < max_order; i++) { in check_multi_store_advanced() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| H A D | mlx5dr.h | 171 u32 max_order; member 186 unsigned int max_order);
|
| /linux/kernel/events/ |
| H A D | ring_buffer.c | 689 int max_order = 0; in rb_alloc_aux() local 713 max_order = get_order(watermark); in rb_alloc_aux() 721 max_order = ilog2(nr_pages); in rb_alloc_aux() 741 order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); in rb_alloc_aux() 761 if (page_private(page) != max_order) in rb_alloc_aux()
|
| /linux/drivers/pci/ |
| H A D | setup-bus.c | 1209 int max_order) in calculate_mem_align() argument 1215 for (order = 0; order <= max_order; order++) { in calculate_mem_align() 1318 int order, max_order; in pbus_size_mem() local 1334 max_order = 0; in pbus_size_mem() 1384 if (order > max_order) in pbus_size_mem() 1385 max_order = order; in pbus_size_mem() 1397 min_align = calculate_mem_align(aligns, max_order); in pbus_size_mem() 1408 relaxed_align = 1ULL << (max_order + __ffs(SZ_1M)); in pbus_size_mem() 1424 relaxed_align = 1ULL << (max_order + __ffs(SZ_1M)); in pbus_size_mem()
|
| /linux/tools/testing/kunit/test_data/ |
| H A D | test_is_test_passed-no_tests_run_no_header.log | 31 workingset: timestamp_bits=62 max_order=13 bucket_order=0
|
| H A D | test_output_isolated_correctly.log | 56 workingset: timestamp_bits=62 max_order=16 bucket_order=0
|
| /linux/drivers/accel/qaic/ |
| H A D | qaic_data.c | 451 int max_order; in create_sgt() local 464 max_order = min(MAX_PAGE_ORDER, get_order(size)); in create_sgt() 469 max_order = 0; in create_sgt() 486 order = min(get_order(nr_pages * PAGE_SIZE), max_order); in create_sgt() 500 max_order = order; in create_sgt()
|
| /linux/include/linux/ |
| H A D | kvm_host.h | 2545 int *max_order); 2550 int *max_order) in kvm_gmem_get_pfn() argument 2558 int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_order);
|
| /linux/fs/ext4/ |
| H A D | inode.c | 5150 u16 min_order, max_order; in ext4_set_inode_mapping_order() local 5152 max_order = EXT4_SB(sb)->s_max_folio_order; in ext4_set_inode_mapping_order() 5153 if (!max_order) in ext4_set_inode_mapping_order() 5161 max_order = min_order; in ext4_set_inode_mapping_order() 5163 mapping_set_folio_order_range(inode->i_mapping, min_order, max_order); in ext4_set_inode_mapping_order()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | sev.c | 4952 int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order) in sev_gmem_prepare() argument 4972 __func__, gfn, pfn, max_order, level); in sev_gmem_prepare() 4976 if (is_large_rmp_possible(kvm, pfn, max_order)) { in sev_gmem_prepare() 4994 __func__, gfn, pfn, pfn_aligned, max_order, level); in sev_gmem_prepare()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | mlx4.h | 256 u32 max_order; member
|
| /linux/arch/x86/include/asm/ |
| H A D | kvm_host.h | 1940 int (*gmem_prepare)(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | mmu.c | 4571 int max_order, r; in kvm_mmu_faultin_pfn_gmem() local 4579 &fault->refcounted_page, &max_order); in kvm_mmu_faultin_pfn_gmem() 4586 fault->max_level = kvm_max_level_for_order(max_order); in kvm_mmu_faultin_pfn_gmem()
|
| /linux/block/ |
| H A D | blk-mq.c | 3590 unsigned int i, j, entries_per_page, max_order = 4; in blk_mq_alloc_rqs() local 3606 int this_order = max_order; in blk_mq_alloc_rqs()
|