| /linux/mm/ |
| H A D | page_alloc.c | 321 int alloc_flags); 1895 unsigned int alloc_flags) in prep_new_page() argument 1908 if (alloc_flags & ALLOC_NO_WATERMARKS) in prep_new_page() 2314 int block_type, unsigned int alloc_flags) in try_to_claim_block() argument 2335 if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD)) in try_to_claim_block() 2388 unsigned int alloc_flags) in __rmqueue_claim() argument 2401 if (order < pageblock_order && alloc_flags & ALLOC_NOFRAGMENT) in __rmqueue_claim() 2426 alloc_flags); in __rmqueue_claim() 2479 unsigned int alloc_flags, enum rmqueue_mode *mode) in __rmqueue() argument 2489 if (alloc_flags & ALLOC_CMA && in __rmqueue() [all …]
|
| H A D | compaction.c | 2299 cc->highest_zoneidx, cc->alloc_flags, in __compact_finished() 2433 int alloc_flags) in compaction_zonelist_suitable() argument 2471 int highest_zoneidx, unsigned int alloc_flags, in compaction_suit_allocation_order() argument 2482 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); in compaction_suit_allocation_order() 2484 alloc_flags, free_pages)) in compaction_suit_allocation_order() 2497 !(alloc_flags & ALLOC_CMA)) { in compaction_suit_allocation_order() 2539 cc->alloc_flags, in compact_zone() 2751 unsigned int alloc_flags, int highest_zoneidx, in compact_zone_order() argument 2762 .alloc_flags = alloc_flags, in compact_zone_order() 2815 unsigned int alloc_flags, const struct alloc_context *ac, in try_to_compact_pages() argument [all …]
|
| /linux/drivers/base/regmap/ |
| H A D | regcache-maple.c | 76 entry = kmalloc_array(last - index + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_write() 94 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_write() 138 map->alloc_flags); in regcache_maple_drop() 151 map->alloc_flags); in regcache_maple_drop() 165 ret = mas_store_gfp(&mas, lower, map->alloc_flags); in regcache_maple_drop() 173 ret = mas_store_gfp(&mas, upper, map->alloc_flags); in regcache_maple_drop() 207 buf = kmalloc_array(max - min, val_bytes, map->alloc_flags); in regcache_maple_sync_block() 297 mt = kmalloc_obj(*mt, map->alloc_flags); in regcache_maple_init() 340 entry = kmalloc_array(last - first + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_insert_block() 351 ret = mas_store_gfp(&mas, entry, map->alloc_flags); in regcache_maple_insert_block()
|
| H A D | regcache-rbtree.c | 188 map->cache = kmalloc_obj(*rbtree_ctx, map->alloc_flags); in regcache_rbtree_init() 280 blk = krealloc_array(rbnode->block, blklen, map->cache_word_size, map->alloc_flags); in regcache_rbtree_insert_to_block() 289 map->alloc_flags); in regcache_rbtree_insert_to_block() 323 rbnode = kzalloc_obj(*rbnode, map->alloc_flags); in regcache_rbtree_node_alloc() 349 map->alloc_flags); in regcache_rbtree_node_alloc() 355 map->alloc_flags); in regcache_rbtree_node_alloc()
|
| H A D | regcache-flat.c | 39 cache = kzalloc_flex(*cache, data, cache_size, map->alloc_flags); in regcache_flat_init() 43 cache->valid = bitmap_zalloc(cache_size, map->alloc_flags); in regcache_flat_init()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_amdkfd_gpuvm.c | 312 u32 alloc_flags = bo->kfd_bo->alloc_flags; in amdgpu_amdkfd_release_notify() local 315 amdgpu_amdkfd_unreserve_mem_limit(adev, size, alloc_flags, in amdgpu_amdkfd_release_notify() 342 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) in create_dmamap_sg_bo() 519 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE) in get_pte_flags() 521 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE) in get_pte_flags() 564 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_userptr() 660 mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP); in kfd_mem_dmamap_sg_bo() 666 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_sg_bo() 725 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmaunmap_userptr() 787 dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmaunmap_sg_bo() [all …]
|
| /linux/drivers/gpu/drm/ttm/tests/ |
| H A D | ttm_kunit_helpers.c | 120 unsigned int alloc_flags, in ttm_device_kunit_init_with_funcs() argument 129 alloc_flags); in ttm_device_kunit_init_with_funcs() 145 unsigned int alloc_flags) in ttm_device_kunit_init() argument 147 return ttm_device_kunit_init_with_funcs(priv, ttm, alloc_flags, in ttm_device_kunit_init()
|
| /linux/include/linux/ |
| H A D | stackdepot.h | 159 gfp_t alloc_flags, 178 unsigned int nr_entries, gfp_t alloc_flags);
|
| H A D | compaction.h | 93 unsigned int order, unsigned int alloc_flags, 104 int alloc_flags);
|
| H A D | gfp.h | 444 acr_flags_t alloc_flags, gfp_t gfp_mask); 449 acr_flags_t alloc_flags, gfp_t gfp_mask);
|
| H A D | mmzone.h | 1538 int highest_zoneidx, unsigned int alloc_flags, 1542 unsigned int alloc_flags);
|
| /linux/lib/ |
| H A D | stackdepot.c | 643 gfp_t alloc_flags, in stack_depot_save_flags() argument 651 bool allow_spin = gfpflags_allow_spinning(alloc_flags); in stack_depot_save_flags() 685 page = alloc_pages(gfp_nested_mask(alloc_flags), in stack_depot_save_flags() 745 gfp_t alloc_flags) in stack_depot_save() argument 747 return stack_depot_save_flags(entries, nr_entries, alloc_flags, in stack_depot_save()
|
| /linux/drivers/md/ |
| H A D | dm-zoned-reclaim.c | 286 int alloc_flags = DMZ_ALLOC_SEQ; in dmz_reclaim_rnd_data() local 292 alloc_flags | DMZ_ALLOC_RECLAIM); in dmz_reclaim_rnd_data() 293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { in dmz_reclaim_rnd_data() 294 alloc_flags = DMZ_ALLOC_RND; in dmz_reclaim_rnd_data()
|
| H A D | dm-zoned-metadata.c | 2050 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_mapping() local 2065 dzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_mapping() 2162 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_buffer() local 2171 bzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_buffer()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_chardev.c | 1972 bo_bucket->alloc_flags = (uint32_t)kgd_mem->alloc_flags; in criu_checkpoint_bos() 1975 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { in criu_checkpoint_bos() 1983 if (bo_bucket->alloc_flags in criu_checkpoint_bos() 1986 bo_bucket->alloc_flags & in criu_checkpoint_bos() 1995 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) in criu_checkpoint_bos() 1998 else if (bo_bucket->alloc_flags & in criu_checkpoint_bos() 2016 bo_bucket->alloc_flags, in criu_checkpoint_bos() 2329 if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) { in criu_restore_memory_of_gpu() 2337 } else if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP) { in criu_restore_memory_of_gpu() 2348 } else if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) { in criu_restore_memory_of_gpu() [all …]
|
| /linux/fs/xfs/ |
| H A D | xfs_extent_busy.c | 613 uint32_t alloc_flags) in xfs_extent_busy_flush() argument 625 if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH) in xfs_extent_busy_flush() 631 if (alloc_flags & XFS_ALLOC_FLAG_FREEING) in xfs_extent_busy_flush()
|
| H A D | xfs_extent_busy.h | 59 unsigned busy_gen, uint32_t alloc_flags);
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_alloc.h | 198 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_pool.c | 1077 int nid, unsigned int alloc_flags) in ttm_pool_init() argument 1085 pool->alloc_flags = alloc_flags; in ttm_pool_init()
|
| /linux/fs/btrfs/ |
| H A D | block-group.c | 3047 u64 alloc_flags; in btrfs_inc_block_group_ro() local 3095 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro() 3096 if (alloc_flags != cache->flags) { in btrfs_inc_block_group_ro() 3097 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, in btrfs_inc_block_group_ro() 3125 alloc_flags = btrfs_get_alloc_profile(fs_info, space_info->flags); in btrfs_inc_block_group_ro() 3126 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_inc_block_group_ro() 3142 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro() 3144 check_system_chunk(trans, alloc_flags); in btrfs_inc_block_group_ro() 4019 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); in btrfs_force_chunk_alloc() local 4028 return btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_force_chunk_alloc()
|
| H A D | zoned.c | 2755 u64 alloc_flags; in btrfs_zoned_reserve_data_reloc_bg() local 2770 alloc_flags = btrfs_get_alloc_profile(fs_info, space_info->flags); in btrfs_zoned_reserve_data_reloc_bg() 2771 index = btrfs_bg_flags_to_raid_index(alloc_flags); in btrfs_zoned_reserve_data_reloc_bg() 2835 ret = btrfs_chunk_alloc(trans, space_info, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_zoned_reserve_data_reloc_bg()
|
| /linux/fs/xfs/scrub/ |
| H A D | repair.h | 49 int xrep_fix_freelist(struct xfs_scrub *sc, int alloc_flags);
|
| /linux/drivers/iommu/ |
| H A D | dma-iommu.c | 902 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages() local 906 alloc_flags |= __GFP_NORETRY; in __iommu_dma_alloc_pages() 907 page = alloc_pages_node(nid, alloc_flags, order); in __iommu_dma_alloc_pages()
|
| /linux/include/uapi/linux/ |
| H A D | kfd_ioctl.h | 733 __u32 alloc_flags; 727 __u32 alloc_flags; global() member
|
| /linux/include/uapi/drm/ |
| H A D | amdgpu_drm.h | 861 __u64 alloc_flags; member
|