| /linux/include/linux/ |
| H A D | gfp.h | 265 static inline void warn_if_node_offline(int this_node, gfp_t gfp_mask) in warn_if_node_offline() argument 267 gfp_t warn_gfp = gfp_mask & (__GFP_THISNODE|__GFP_NOWARN); in warn_if_node_offline() 275 pr_warn("%pGg allocation from offline node %d\n", &gfp_mask, this_node); in warn_if_node_offline() 284 __alloc_pages_node_noprof(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node_noprof() argument 287 warn_if_node_offline(nid, gfp_mask); in __alloc_pages_node_noprof() 289 return __alloc_pages_noprof(gfp_mask, order, nid, NULL); in __alloc_pages_node_noprof() 310 static inline struct page *alloc_pages_node_noprof(int nid, gfp_t gfp_mask, in alloc_pages_node_noprof() argument 316 return __alloc_pages_node_noprof(nid, gfp_mask, order); in alloc_pages_node_noprof() 329 static inline struct page *alloc_pages_noprof(gfp_t gfp_mask, unsigned int order) in alloc_pages_noprof() argument 331 return alloc_pages_node_noprof(numa_node_id(), gfp_mask, order); in alloc_pages_noprof() [all …]
|
| H A D | mempool.h | 15 typedef void * (mempool_alloc_t)(gfp_t gfp_mask, void *pool_data); 43 void *pool_data, gfp_t gfp_mask, int node_id); 54 void *pool_data, gfp_t gfp_mask, int nid); 65 void *mempool_alloc_noprof(struct mempool *pool, gfp_t gfp_mask) __malloc; 83 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data); 95 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data); 109 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data);
|
| H A D | cpuset.h | 88 extern bool cpuset_current_node_allowed(int node, gfp_t gfp_mask); 90 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() argument 92 return cpuset_current_node_allowed(zone_to_nid(z), gfp_mask); in __cpuset_zone_allowed() 95 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() argument 98 return __cpuset_zone_allowed(z, gfp_mask); in cpuset_zone_allowed() 232 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() argument 237 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() argument
|
| /linux/mm/ |
| H A D | mempool.c | 235 void *pool_data, gfp_t gfp_mask, int node_id) in mempool_init_node() argument 248 gfp_mask, node_id); in mempool_init_node() 259 element = pool->alloc(gfp_mask, pool->pool_data); in mempool_init_node() 315 void *pool_data, gfp_t gfp_mask, int node_id) in mempool_create_node_noprof() argument 319 pool = kmalloc_node_noprof(sizeof(*pool), gfp_mask | __GFP_ZERO, node_id); in mempool_create_node_noprof() 324 gfp_mask, node_id)) { in mempool_create_node_noprof() 414 gfp_t gfp_mask) in mempool_alloc_from_pool() argument 442 if (gfp_mask & __GFP_DIRECT_RECLAIM) { in mempool_alloc_from_pool() 471 static inline gfp_t mempool_adjust_gfp(gfp_t *gfp_mask) in mempool_adjust_gfp() argument 473 *gfp_mask |= __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN; in mempool_adjust_gfp() [all …]
|
| H A D | page_alloc.c | 3673 unsigned int alloc_flags, gfp_t gfp_mask) in zone_watermark_fast() argument 3740 alloc_flags_nofragment(struct zone *zone, gfp_t gfp_mask) in alloc_flags_nofragment() argument 3748 alloc_flags = (__force int) (gfp_mask & __GFP_KSWAPD_RECLAIM); in alloc_flags_nofragment() 3777 static inline unsigned int gfp_to_alloc_flags_cma(gfp_t gfp_mask, in gfp_to_alloc_flags_cma() argument 3781 if (gfp_migratetype(gfp_mask) == MIGRATE_MOVABLE) in gfp_to_alloc_flags_cma() 3792 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, in get_page_from_freelist() argument 3817 !__cpuset_zone_allowed(zone, gfp_mask)) in get_page_from_freelist() 3892 gfp_mask)) in get_page_from_freelist() 3901 gfp_mask)) { in get_page_from_freelist() 3924 ret = node_reclaim(zone->zone_pgdat, gfp_mask, order); in get_page_from_freelist() [all …]
|
| H A D | page_owner.c | 27 gfp_t gfp_mask; member 175 gfp_t gfp_mask) in add_stack_record_to_list() argument 180 if (!gfpflags_allow_spinning(gfp_mask)) in add_stack_record_to_list() 184 stack = kmalloc_obj(*stack, gfp_nested_mask(gfp_mask)); in add_stack_record_to_list() 206 static void inc_stack_record_count(depot_stack_handle_t handle, gfp_t gfp_mask, in inc_stack_record_count() argument 226 add_stack_record_to_list(stack_record, gfp_mask); in inc_stack_record_count() 247 gfp_t gfp_mask, in __update_page_owner_handle() argument 260 page_owner->gfp_mask = gfp_mask; in __update_page_owner_handle() 336 gfp_t gfp_mask) in __set_page_owner() argument 341 handle = save_stack(gfp_mask); in __set_page_owner() [all …]
|
| H A D | fail_page_alloc.c | 26 bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) in should_fail_alloc_page() argument 32 if (gfp_mask & __GFP_NOFAIL) in should_fail_alloc_page() 34 if (fail_page_alloc.ignore_gfp_highmem && (gfp_mask & __GFP_HIGHMEM)) in should_fail_alloc_page() 37 (gfp_mask & __GFP_DIRECT_RECLAIM)) in should_fail_alloc_page() 41 if (gfp_mask & __GFP_NOWARN) in should_fail_alloc_page()
|
| H A D | swap_state.c | 529 struct folio *swap_cache_alloc_folio(swp_entry_t entry, gfp_t gfp_mask, in swap_cache_alloc_folio() argument 548 folio = folio_alloc_mpol(gfp_mask, 0, mpol, ilx, numa_node_id()); in swap_cache_alloc_folio() 552 result = __swap_cache_prepare_and_add(entry, folio, gfp_mask, false); in swap_cache_alloc_folio() 593 struct folio *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in read_swap_cache_async() argument 608 folio = swap_cache_alloc_folio(entry, gfp_mask, mpol, ilx, in read_swap_cache_async() 698 struct folio *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask, in swap_cluster_readahead() argument 727 swp_entry(swp_type(entry), offset), gfp_mask, mpol, ilx, in swap_cluster_readahead() 745 folio = swap_cache_alloc_folio(entry, gfp_mask, mpol, ilx, in swap_cluster_readahead() 806 static struct folio *swap_vma_readahead(swp_entry_t targ_entry, gfp_t gfp_mask, in swap_vma_readahead() argument 850 folio = swap_cache_alloc_folio(entry, gfp_mask, mpol, ilx, in swap_vma_readahead() [all …]
|
| H A D | vmalloc.c | 687 gfp_t gfp_mask) in vmap_pages_range_noflush() argument 690 page_shift, gfp_mask); in vmap_pages_range_noflush() 699 gfp_t gfp_mask) in __vmap_pages_range() argument 703 err = vmap_pages_range_noflush(addr, end, prot, pages, page_shift, gfp_mask); in __vmap_pages_range() 1888 preload_this_cpu_lock(spinlock_t *lock, gfp_t gfp_mask, int node) in preload_this_cpu_lock() argument 1902 va = kmem_cache_alloc_node(vmap_area_cachep, gfp_mask, node); in preload_this_cpu_lock() 2026 int node, gfp_t gfp_mask, in alloc_vmap_area() argument 2045 gfp_mask = gfp_mask & GFP_RECLAIM_MASK; in alloc_vmap_area() 2046 allow_block = gfpflags_allow_blocking(gfp_mask); in alloc_vmap_area() 2059 va = kmem_cache_alloc_node(vmap_area_cachep, gfp_mask, node); in alloc_vmap_area() [all …]
|
| /linux/block/ |
| H A D | blk-map.c | 22 gfp_t gfp_mask) in bio_alloc_map_data() argument 29 bmd = kmalloc_flex(*bmd, iov, data->nr_segs, gfp_mask); in bio_alloc_map_data() 46 unsigned int nr_vecs, gfp_t gfp_mask) in blk_rq_map_bio_alloc() argument 51 bio = bio_alloc_bioset(bdev, nr_vecs, rq->cmd_flags, gfp_mask, in blk_rq_map_bio_alloc() 151 struct iov_iter *iter, gfp_t gfp_mask) in bio_copy_user_iov() argument 161 bmd = bio_alloc_map_data(iter, gfp_mask); in bio_copy_user_iov() 176 bio = blk_rq_map_bio_alloc(rq, nr_pages, gfp_mask); in bio_copy_user_iov() 203 page = alloc_page(GFP_NOIO | gfp_mask); in bio_copy_user_iov() 261 gfp_t gfp_mask) in bio_map_user_iov() argument 270 bio = blk_rq_map_bio_alloc(rq, nr_vecs, gfp_mask); in bio_map_user_iov() [all …]
|
| /linux/fs/nfs/blocklayout/ |
| H A D | dev.c | 290 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask); 295 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_simple() argument 301 dev = bl_resolve_deviceid(server, v, gfp_mask); in bl_parse_simple() 389 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_scsi() argument 445 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_slice() argument 450 ret = bl_parse_deviceid(server, d, volumes, v->slice.volume, gfp_mask); in bl_parse_slice() 461 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_concat() argument 468 v->concat.volumes_count, gfp_mask); in bl_parse_concat() 474 volumes, v->concat.volumes[i], gfp_mask); in bl_parse_concat() 490 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_stripe() argument [all …]
|
| /linux/fs/btrfs/ |
| H A D | ulist.h | 50 struct ulist *ulist_alloc(gfp_t gfp_mask); 53 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask); 55 u64 *old_aux, gfp_t gfp_mask); 60 void **old_aux, gfp_t gfp_mask) in ulist_add_merge_ptr() argument 64 int ret = ulist_add_merge(ulist, val, (uintptr_t)aux, &old64, gfp_mask); in ulist_add_merge_ptr() 68 return ulist_add_merge(ulist, val, (u64)aux, (u64 *)old_aux, gfp_mask); in ulist_add_merge_ptr()
|
| H A D | ulist.c | 99 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() argument 101 struct ulist *ulist = kmalloc_obj(*ulist, gfp_mask); in ulist_alloc() 111 void ulist_prealloc(struct ulist *ulist, gfp_t gfp_mask) in ulist_prealloc() argument 114 ulist->prealloc = kzalloc_obj(*ulist->prealloc, gfp_mask); in ulist_prealloc() 200 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() argument 202 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); in ulist_add() 206 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge() argument 222 node = kmalloc_obj(*node, gfp_mask); in ulist_add_merge()
|
| /linux/include/linux/sched/ |
| H A D | mm.h | 272 extern void fs_reclaim_acquire(gfp_t gfp_mask); 273 extern void fs_reclaim_release(gfp_t gfp_mask); 277 static inline void fs_reclaim_acquire(gfp_t gfp_mask) { } in fs_reclaim_acquire() argument 278 static inline void fs_reclaim_release(gfp_t gfp_mask) { } in fs_reclaim_release() argument 315 static inline void might_alloc(gfp_t gfp_mask) in might_alloc() argument 317 fs_reclaim_acquire(gfp_mask); in might_alloc() 318 fs_reclaim_release(gfp_mask); in might_alloc() 323 might_sleep_if(gfpflags_allow_blocking(gfp_mask)); in might_alloc()
|
| /linux/net/sunrpc/auth_gss/ |
| H A D | gss_krb5_keys.c | 152 const struct xdr_netobj *in_constant, gfp_t gfp_mask) in krb5_DK() argument 174 inblockdata = kmalloc(blocksize, gfp_mask); in krb5_DK() 178 outblockdata = kmalloc(blocksize, gfp_mask); in krb5_DK() 271 gfp_t gfp_mask) in krb5_derive_key_v2() argument 277 inblock.data = kmalloc(inblock.len, gfp_mask); in krb5_derive_key_v2() 281 ret = krb5_DK(gk5e, inkey, inblock.data, label, gfp_mask); in krb5_derive_key_v2() 372 gfp_t gfp_mask) in krb5_kdf_feedback_cmac() argument 401 step.data = kzalloc(step.len, gfp_mask); in krb5_kdf_feedback_cmac() 406 DR.data = kmalloc(DR.len, gfp_mask); in krb5_kdf_feedback_cmac() 504 gfp_t gfp_mask) in krb5_kdf_hmac_sha2() argument [all …]
|
| H A D | gss_krb5_mech.c | 297 gss_krb5_import_ctx_v2(struct krb5_ctx *ctx, gfp_t gfp_mask) in gss_krb5_import_ctx_v2() argument 306 keyout.data = kmalloc(GSS_KRB5_MAX_KEYLEN, gfp_mask); in gss_krb5_import_ctx_v2() 313 KEY_USAGE_SEED_ENCRYPTION, gfp_mask)) in gss_krb5_import_ctx_v2() 329 KEY_USAGE_SEED_ENCRYPTION, gfp_mask)) in gss_krb5_import_ctx_v2() 346 KEY_USAGE_SEED_CHECKSUM, gfp_mask)) in gss_krb5_import_ctx_v2() 354 KEY_USAGE_SEED_CHECKSUM, gfp_mask)) in gss_krb5_import_ctx_v2() 363 KEY_USAGE_SEED_INTEGRITY, gfp_mask)) in gss_krb5_import_ctx_v2() 371 KEY_USAGE_SEED_INTEGRITY, gfp_mask)) in gss_krb5_import_ctx_v2() 396 gfp_t gfp_mask) in gss_import_v2_context() argument 447 gss_kerberos_mech.gm_oid.len, gfp_mask); in gss_import_v2_context() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | icm.c | 99 gfp_t gfp_mask, int node) in mlx4_alloc_icm_pages() argument 103 page = alloc_pages_node(node, gfp_mask, order); in mlx4_alloc_icm_pages() 105 page = alloc_pages(gfp_mask, order); in mlx4_alloc_icm_pages() 115 int order, gfp_t gfp_mask) in mlx4_alloc_icm_coherent() argument 118 &buf->dma_addr, gfp_mask); in mlx4_alloc_icm_coherent() 133 gfp_t gfp_mask, int coherent) in mlx4_alloc_icm() argument 142 BUG_ON(coherent && (gfp_mask & __GFP_HIGHMEM)); in mlx4_alloc_icm() 145 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN), in mlx4_alloc_icm() 149 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN)); in mlx4_alloc_icm() 162 gfp_mask & ~(__GFP_HIGHMEM | in mlx4_alloc_icm() [all …]
|
| /linux/mm/kasan/ |
| H A D | shadow.c | 339 static int ___alloc_pages_bulk(struct page **pages, int nr_pages, gfp_t gfp_mask) in ___alloc_pages_bulk() argument 345 nr_populated = alloc_pages_bulk(gfp_mask, nr_pages, pages); in ___alloc_pages_bulk() 357 static int __kasan_populate_vmalloc_do(unsigned long start, unsigned long end, gfp_t gfp_mask) in __kasan_populate_vmalloc_do() argument 364 data.pages = (struct page **)__get_free_page(gfp_mask | __GFP_ZERO); in __kasan_populate_vmalloc_do() 370 ret = ___alloc_pages_bulk(data.pages, nr_pages, gfp_mask); in __kasan_populate_vmalloc_do() 380 flags = memalloc_apply_gfp_scope(gfp_mask); in __kasan_populate_vmalloc_do() 398 int __kasan_populate_vmalloc(unsigned long addr, unsigned long size, gfp_t gfp_mask) in __kasan_populate_vmalloc() argument 424 ret = __kasan_populate_vmalloc_do(shadow_start, shadow_end, gfp_mask); in __kasan_populate_vmalloc() 656 int kasan_alloc_module_shadow(void *addr, size_t size, gfp_t gfp_mask) in kasan_alloc_module_shadow() argument 689 kmemleak_vmalloc(vm, size, gfp_mask); in kasan_alloc_module_shadow()
|
| /linux/drivers/connector/ |
| H A D | connector.c | 62 gfp_t gfp_mask, netlink_filter_fn filter, in cn_netlink_send_mult() argument 97 skb = nlmsg_new(size, gfp_mask); in cn_netlink_send_mult() 115 gfp_mask, filter, in cn_netlink_send_mult() 118 !gfpflags_allow_blocking(gfp_mask)); in cn_netlink_send_mult() 124 gfp_t gfp_mask) in cn_netlink_send() argument 126 return cn_netlink_send_mult(msg, msg->len, portid, __group, gfp_mask, in cn_netlink_send()
|
| /linux/kernel/power/ |
| H A D | snapshot.c | 191 static void *get_image_page(gfp_t gfp_mask, int safe_needed) in get_image_page() argument 195 res = (void *)get_zeroed_page(gfp_mask); in get_image_page() 201 res = (void *)get_zeroed_page(gfp_mask); in get_image_page() 210 static void *__get_safe_page(gfp_t gfp_mask) in __get_safe_page() argument 219 return get_image_page(gfp_mask, PG_SAFE); in __get_safe_page() 222 unsigned long get_safe_page(gfp_t gfp_mask) in get_safe_page() argument 224 return (unsigned long)__get_safe_page(gfp_mask); in get_safe_page() 227 static struct page *alloc_image_page(gfp_t gfp_mask) in alloc_image_page() argument 231 page = alloc_page(gfp_mask); in alloc_image_page() 297 gfp_t gfp_mask; /* mask for allocating pages */ member [all …]
|
| /linux/tools/testing/selftests/net/bench/page_pool/ |
| H A D | bench_page_pool_simple.c | 109 gfp_t gfp_mask = GFP_ATOMIC; in pp_fill_ptr_ring() local 113 array = kcalloc(elems, sizeof(struct page *), gfp_mask); in pp_fill_ptr_ring() 116 array[i] = page_pool_alloc_pages(pp, gfp_mask); in pp_fill_ptr_ring() 130 gfp_t gfp_mask = GFP_ATOMIC; /* GFP_ATOMIC is not really needed */ in time_bench_page_pool() local 162 page = page_pool_alloc_pages(pp, gfp_mask); in time_bench_page_pool()
|
| /linux/lib/ |
| H A D | generic-radix-tree.c | 24 gfp_t gfp_mask) in __genradix_ptr_alloc() argument 44 new_node = genradix_alloc_node(gfp_mask); in __genradix_ptr_alloc() 69 new_node = genradix_alloc_node(gfp_mask); in __genradix_ptr_alloc() 211 gfp_t gfp_mask) in __genradix_prealloc() argument 216 if (!__genradix_ptr_alloc(radix, offset, NULL, gfp_mask)) in __genradix_prealloc()
|
| H A D | radix-tree.c | 233 radix_tree_node_alloc(gfp_t gfp_mask, struct radix_tree_node *parent, in radix_tree_node_alloc() argument 245 if (!gfpflags_allow_blocking(gfp_mask) && !in_interrupt()) { in radix_tree_node_alloc() 254 gfp_mask | __GFP_NOWARN); in radix_tree_node_alloc() 276 ret = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); in radix_tree_node_alloc() 322 static __must_check int __radix_tree_preload(gfp_t gfp_mask, unsigned nr) in __radix_tree_preload() argument 332 gfp_mask &= ~__GFP_ACCOUNT; in __radix_tree_preload() 338 node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); in __radix_tree_preload() 365 int radix_tree_preload(gfp_t gfp_mask) in radix_tree_preload() argument 368 WARN_ON_ONCE(!gfpflags_allow_blocking(gfp_mask)); in radix_tree_preload() 369 return __radix_tree_preload(gfp_mask, RADIX_TREE_PRELOAD_SIZE); in radix_tree_preload() [all …]
|
| /linux/include/trace/events/ |
| H A D | compaction.h | 169 gfp_t gfp_mask, 172 TP_ARGS(order, gfp_mask, prio), 176 __field(unsigned long, gfp_mask) 182 __entry->gfp_mask = (__force unsigned long)gfp_mask; 188 show_gfp_flags(__entry->gfp_mask),
|
| /linux/drivers/infiniband/core/ |
| H A D | sa_query.c | 844 static int ib_nl_make_request(struct ib_sa_query *query, gfp_t gfp_mask) in ib_nl_make_request() argument 864 skb = nlmsg_new(len, gfp_mask); in ib_nl_make_request() 882 gfp_flag = ((gfp_mask & GFP_ATOMIC) == GFP_ATOMIC) ? GFP_ATOMIC : in ib_nl_make_request() 1330 static int alloc_mad(struct ib_sa_query *query, gfp_t gfp_mask) in alloc_mad() argument 1356 gfp_mask, in alloc_mad() 1398 gfp_t gfp_mask) in send_mad() argument 1405 ret = __xa_alloc(&queries, &id, query, xa_limit_32b, gfp_mask); in send_mad() 1423 if (!ib_nl_make_request(query, gfp_mask)) in send_mad() 1667 unsigned long timeout_ms, gfp_t gfp_mask, in ib_sa_path_rec_get() argument 1692 query = kzalloc_obj(*query, gfp_mask); in ib_sa_path_rec_get() [all …]
|