Lines Matching refs:pool_global

76 static struct obj_pool pool_global = {  variable
238 struct pool_stats *stats = &pool_global.stats; in pcpu_refill_stats()
245 if (pool_global.cnt < stats->min_fill) in pcpu_refill_stats()
246 stats->min_fill = pool_global.cnt; in pcpu_refill_stats()
283 if (!pool_move_batch(pcp, &pool_global)) in pcpu_alloc()
313 if (!pool_move_batch(&pool_global, pcp)) in pcpu_free()
315 WRITE_ONCE(pool_global.stats.cur_used, pool_global.stats.cur_used - ODEBUG_BATCH_SIZE); in pcpu_free()
357 while (pool_should_refill(&pool_global) && pool_count(&pool_to_free)) { in fill_pool_from_freelist()
360 pool_move_batch(&pool_global, &pool_to_free); in fill_pool_from_freelist()
396 if (!pool_must_refill(&pool_global) && atomic_read(&cpus_allocating)) in fill_pool()
400 while (pool_should_refill(&pool_global)) { in fill_pool()
407 if (!pool_push_batch(&pool_global, &head)) in fill_pool()
445 cur = READ_ONCE(pool_global.stats.cur_used) * ODEBUG_FREE_WORK_MAX; in calc_usage()
503 if (pool_move_batch(&pool_global, &pool_to_free)) { in free_obj_work()
702 if (likely(!pool_should_refill(&pool_global))) in debug_objects_fill_pool()
708 if (likely(!pool_should_refill(&pool_global))) in debug_objects_fill_pool()
1148 pool_used = READ_ONCE(pool_global.stats.cur_used); in debug_stats_show()
1156 seq_printf(m, "pool_free : %u\n", pool_count(&pool_global) + pcp_free); in debug_stats_show()
1158 seq_printf(m, "pool_min_free : %u\n", data_race(pool_global.stats.min_fill)); in debug_stats_show()
1160 seq_printf(m, "pool_max_used : %u\n", data_race(pool_global.stats.max_used)); in debug_stats_show()
1441 pool_push_batch(&pool_global, &objects); in debug_objects_replace_static_objects()
1462 hlist_for_each_entry_safe(obj, tmp, &pool_global.objects, node) { in debug_objects_replace_static_objects()
1501 pool_global.max_cnt += extras; in debug_objects_mem_init()
1502 pool_global.min_cnt += extras; in debug_objects_mem_init()