Searched refs:allocs (Results 1 – 5 of 5) sorted by relevance
2016 * atomic allocs. Fill up from the most packed so that atomic in pcpu_balance_populated() 2017 * allocs don't increase fragmentation. If atomic allocation in pcpu_balance_populated() 2019 * should prevent atomic allocs larger than PAGE_SIZE from keeping in pcpu_balance_populated() 2020 * failing indefinitely; however, large atomic allocs are not in pcpu_balance_populated() 2467 int upa, apl; /* units per alloc, allocs per line */ in pcpu_dump_alloc_info() 2877 int allocs = 0, wasted = 0; in pcpu_build_alloc_info() 2884 allocs += this_allocs; in pcpu_build_alloc_info() 2897 if (allocs > last_allocs) in pcpu_build_alloc_info() 2899 last_allocs = allocs; in pcpu_build_alloc_info() 2864 int allocs = 0, wasted = 0; pcpu_build_alloc_info() local
119 unsigned long allocs; /* number of allocated neighs */117 unsigned long allocs; /* number of allocated neighs */ global() member
536 NEIGH_CACHE_STAT_INC(tbl, allocs); in neigh_alloc()2301 ndst.ndts_allocs += READ_ONCE(st->allocs); in neightbl_fill_info()3535 st->allocs, in neigh_stat_seq_show()
3072 struct wmi_sched_scheme_slot allocs[WMI_SCHED_MAX_ALLOCS_PER_CMD]; member
7004 allocs to different slabs, especially in hardened