Lines Matching refs:gfp_mask
197 gfp_t gfp_mask, int node_id)
207 gfp_mask, node_id);
217 element = pool->alloc(gfp_mask, pool->pool_data);
259 * @gfp_mask: memory allocation flags
272 gfp_t gfp_mask, int node_id)
276 pool = kmalloc_node_noprof(sizeof(*pool), gfp_mask | __GFP_ZERO, node_id);
281 gfp_mask, node_id)) {
374 * @gfp_mask: the usual allocation bitmask.
384 void *mempool_alloc_noprof(mempool_t *pool, gfp_t gfp_mask)
391 VM_WARN_ON_ONCE(gfp_mask & __GFP_ZERO);
392 might_alloc(gfp_mask);
394 gfp_mask |= __GFP_NOMEMALLOC; /* don't allocate emergency reserves */
395 gfp_mask |= __GFP_NORETRY; /* don't loop in __alloc_pages */
396 gfp_mask |= __GFP_NOWARN; /* failures are OK */
398 gfp_temp = gfp_mask & ~(__GFP_DIRECT_RECLAIM|__GFP_IO);
424 if (gfp_temp != gfp_mask) {
426 gfp_temp = gfp_mask;
431 if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) {
555 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data)
559 return kmem_cache_alloc_noprof(mem, gfp_mask);
574 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data)
577 return kmalloc_noprof(size, gfp_mask);
587 void *mempool_kvmalloc(gfp_t gfp_mask, void *pool_data)
590 return kvmalloc(size, gfp_mask);
604 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data)
607 return alloc_pages_noprof(gfp_mask, order);