H A D | zsmalloc.c | 40 * class->lock 141 * determined). NOTE: all those class sizes must be set as multiple of 191 * Size of objects stored in this class. Must be multiple 252 unsigned int class:CLASS_BITS + 1; 353 /* class->lock(which owns the handle) synchronizes races */ in record_obj() 448 /* Protected by class->lock */ in get_zspage_inuse() 499 return pool->size_class[zspage->class]; 504 * class maintains a list of zspages where each zspage is divided 507 * size class which has chunk size big enough to hold the given size. in get_size_class_index() 520 static inline void class_stat_add(struct size_class *class, in in class_stat_add() 250 unsigned int class:CLASS_BITS + 1; global() member 518 class_stat_add(struct size_class * class,int type,unsigned long cnt) class_stat_add() argument 524 class_stat_sub(struct size_class * class,int type,unsigned long cnt) class_stat_sub() argument 530 class_stat_read(struct size_class * class,int type) class_stat_read() argument 558 struct size_class *class; zs_stats_size_show() local 661 get_fullness_group(struct size_class * class,struct zspage * zspage) get_fullness_group() argument 688 insert_zspage(struct size_class * class,struct zspage * zspage,int fullness) insert_zspage() argument 701 remove_zspage(struct size_class * class,struct zspage * zspage) remove_zspage() argument 720 fix_fullness_group(struct size_class * class,struct zspage * zspage) fix_fullness_group() argument 840 __free_zspage(struct zs_pool * pool,struct size_class * class,struct zspage * zspage) __free_zspage() argument 867 free_zspage(struct zs_pool * pool,struct size_class * class,struct zspage * zspage) free_zspage() argument 888 init_zspage(struct size_class * class,struct zspage * zspage) init_zspage() argument 932 create_page_chain(struct size_class * class,struct zspage * zspage,struct page * pages[]) create_page_chain() argument 969 alloc_zspage(struct zs_pool * pool,struct size_class * class,gfp_t gfp) alloc_zspage() argument 1009 find_get_zspage(struct size_class * class) find_get_zspage() argument 1131 zspage_full(struct size_class * class,struct zspage * zspage) zspage_full() argument 1154 struct size_class *class; zs_lookup_class_index() local 1191 struct size_class *class; zs_map_object() local 1252 struct size_class *class; zs_unmap_object() local 1304 struct size_class *class; obj_malloc() local 1354 struct size_class *class; zs_malloc() local 1441 struct size_class *class; zs_free() local 1471 zs_object_copy(struct size_class * class,unsigned long dst,unsigned long src) zs_object_copy() argument 1545 find_alloced_obj(struct size_class * class,struct page * page,int * obj_idx) find_alloced_obj() argument 1578 struct size_class *class = pool->size_class[src_zspage->class]; migrate_zspage() local 1606 isolate_src_zspage(struct size_class * class) isolate_src_zspage() argument 1623 isolate_dst_zspage(struct size_class * class) isolate_dst_zspage() argument 1647 putback_zspage(struct size_class * class,struct zspage * zspage) putback_zspage() argument 1730 replace_sub_page(struct size_class * class,struct zspage * zspage,struct page * newpage,struct page * oldpage) replace_sub_page() argument 1768 struct size_class *class; zs_page_migrate() local 1862 struct size_class *class; async_free_zspage() local 1925 zs_can_compact(struct size_class * class) zs_can_compact() argument 1941 __zs_compact(struct zs_pool * pool,struct size_class * class) __zs_compact() argument 2005 struct size_class *class; zs_compact() local 2056 struct size_class *class; zs_shrinker_count() local 2152 struct size_class *class; zs_create_pool() local 2247 struct size_class *class = pool->size_class[i]; zs_destroy_pool() local [all...] |