Searched refs:slab_cache (Results 1 – 13 of 13) sorted by relevance
| /linux/mm/kasan/ |
| H A D | common.c | 485 poison_kmalloc_redzone(slab->slab_cache, object, size, flags); in __kasan_krealloc() 536 if (check_slab_allocation(slab->slab_cache, ptr, ip)) in __kasan_mempool_poison_object() 539 poison_slab_object(slab->slab_cache, ptr, false); in __kasan_mempool_poison_object() 564 unpoison_slab_object(slab->slab_cache, ptr, flags, false); in __kasan_mempool_unpoison_object() 567 if (is_kmalloc_cache(slab->slab_cache)) in __kasan_mempool_unpoison_object() 568 poison_kmalloc_redzone(slab->slab_cache, ptr, size, flags); in __kasan_mempool_unpoison_object()
|
| H A D | quarantine.c | 131 return virt_to_slab(qlink)->slab_cache; in qlink_to_cache()
|
| H A D | report.c | 502 info->cache = slab->slab_cache; in complete_report_info()
|
| H A D | generic.c | 549 cache = slab->slab_cache; in kasan_record_aux_stack()
|
| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_active.c | 24 static struct kmem_cache *slab_cache; variable 172 kmem_cache_free(slab_cache, it); in __active_retire() 319 node = kmem_cache_alloc(slab_cache, GFP_ATOMIC); in active_instance() 747 kmem_cache_free(slab_cache, ref->cache); in i915_active_fini() 867 node = kmem_cache_alloc(slab_cache, GFP_KERNEL); in i915_active_acquire_preallocate_barrier() 915 kmem_cache_free(slab_cache, node); in i915_active_acquire_preallocate_barrier() 1179 kmem_cache_destroy(slab_cache); in i915_active_module_exit() 1184 slab_cache = KMEM_CACHE(active_node, SLAB_HWCACHE_ALIGN); in i915_active_module_init() 1185 if (!slab_cache) in i915_active_module_init()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | split_page_table_lock.rst | 62 确保架构不使用slab分配器来分配页表:slab使用page->slab_cache来分配其页
|
| /linux/mm/ |
| H A D | slab.h | 77 struct kmem_cache *slab_cache; 97 SLAB_MATCH(compound_head, slab_cache); /* Ensure bit 0 is clear */ 71 struct kmem_cache *slab_cache; global() member
|
| H A D | slub.c | 1765 if (unlikely(s != slab->slab_cache)) { in free_consistency_checks() 1766 if (!slab->slab_cache) { in free_consistency_checks() 2053 unsigned int offs = obj_to_index(obj_exts_slab->slab_cache, in mark_objexts_empty() 2190 virt_to_slab(vec)->slab_cache == s); in alloc_slab_obj_exts() 2245 if (obj_exts_in_slab(slab->slab_cache, slab)) { in free_slab_obj_exts() 2524 s = slab->slab_cache; in memcg_slab_post_charge() 3462 slab->slab_cache = s; in allocate_slab() 3536 __free_slab(slab->slab_cache, slab, true); in rcu_free_slab() 6079 s = slab->slab_cache; in free_deferred_objects() 6184 s = slab->slab_cache; in slab_free_after_rcu_debug() [all …]
|
| H A D | slab_common.c | 1254 return slab ? slab->slab_cache : NULL; in ksize() 1606 s = slab->slab_cache; in kfree_rcu_work()
|
| H A D | memcontrol.c | 2634 off = obj_to_index(slab->slab_cache, slab, p); in mem_cgroup_from_obj_slab()
|
| /linux/tools/cgroup/ |
| H A D | memcg_slabinfo.py | 198 cache = slab.slab_cache
|
| /linux/mm/kfence/ |
| H A D | core.c | 492 slab->slab_cache = cache; in kfence_guarded_alloc()
|
| /linux/drivers/md/ |
| H A D | raid5.c | 2386 sh = alloc_stripe(conf->slab_cache, gfp, conf->pool_size, conf); in grow_one_stripe() 2392 free_stripe(conf->slab_cache, sh); in grow_one_stripe() 2425 conf->slab_cache = sc; in grow_stripes() 2598 free_stripe(conf->slab_cache, osh); in resize_stripes() 2606 kmem_cache_destroy(conf->slab_cache); in resize_stripes() 2636 conf->slab_cache = sc; in resize_stripes() 2694 free_stripe(conf->slab_cache, sh); in drop_one_stripe() 2706 kmem_cache_destroy(conf->slab_cache); in shrink_stripes() 2707 conf->slab_cache = NULL; in shrink_stripes()
|