Lines Matching full:cma
3 * CMA DebugFS Interface
10 #include <linux/cma.h>
16 #include "cma.h"
36 struct cma *cma = data; in cma_used_get() local
38 spin_lock_irq(&cma->lock); in cma_used_get()
39 *val = cma->count - cma->available_count; in cma_used_get()
40 spin_unlock_irq(&cma->lock); in cma_used_get()
48 struct cma *cma = data; in cma_maxchunk_get() local
55 spin_lock_irq(&cma->lock); in cma_maxchunk_get()
56 for (r = 0; r < cma->nranges; r++) { in cma_maxchunk_get()
57 cmr = &cma->ranges[r]; in cma_maxchunk_get()
58 bitmap_maxno = cma_bitmap_maxno(cma, cmr); in cma_maxchunk_get()
62 spin_unlock_irq(&cma->lock); in cma_maxchunk_get()
63 *val = (u64)maxchunk << cma->order_per_bit; in cma_maxchunk_get()
69 static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem) in cma_add_to_cma_mem_list() argument
71 spin_lock(&cma->mem_head_lock); in cma_add_to_cma_mem_list()
72 hlist_add_head(&mem->node, &cma->mem_head); in cma_add_to_cma_mem_list()
73 spin_unlock(&cma->mem_head_lock); in cma_add_to_cma_mem_list()
76 static struct cma_mem *cma_get_entry_from_list(struct cma *cma) in cma_get_entry_from_list() argument
80 spin_lock(&cma->mem_head_lock); in cma_get_entry_from_list()
81 if (!hlist_empty(&cma->mem_head)) { in cma_get_entry_from_list()
82 mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); in cma_get_entry_from_list()
85 spin_unlock(&cma->mem_head_lock); in cma_get_entry_from_list()
90 static int cma_free_mem(struct cma *cma, int count) in cma_free_mem() argument
95 mem = cma_get_entry_from_list(cma); in cma_free_mem()
100 cma_release(cma, mem->p, mem->n); in cma_free_mem()
103 } else if (cma->order_per_bit == 0) { in cma_free_mem()
104 cma_release(cma, mem->p, count); in cma_free_mem()
108 cma_add_to_cma_mem_list(cma, mem); in cma_free_mem()
110 pr_debug("cma: cannot release partial block when order_per_bit != 0\n"); in cma_free_mem()
111 cma_add_to_cma_mem_list(cma, mem); in cma_free_mem()
123 struct cma *cma = data; in cma_free_write() local
125 return cma_free_mem(cma, pages); in cma_free_write()
129 static int cma_alloc_mem(struct cma *cma, int count) in cma_alloc_mem() argument
138 p = cma_alloc(cma, count, 0, false); in cma_alloc_mem()
147 cma_add_to_cma_mem_list(cma, mem); in cma_alloc_mem()
155 struct cma *cma = data; in cma_alloc_write() local
157 return cma_alloc_mem(cma, pages); in cma_alloc_write()
161 static void cma_debugfs_add_one(struct cma *cma, struct dentry *root_dentry) in cma_debugfs_add_one() argument
168 tmp = debugfs_create_dir(cma->name, root_dentry); in cma_debugfs_add_one()
170 debugfs_create_file("alloc", 0200, tmp, cma, &cma_alloc_fops); in cma_debugfs_add_one()
171 debugfs_create_file("free", 0200, tmp, cma, &cma_free_fops); in cma_debugfs_add_one()
172 debugfs_create_file("count", 0444, tmp, &cma->count, &cma_debugfs_fops); in cma_debugfs_add_one()
174 &cma->order_per_bit, &cma_debugfs_fops); in cma_debugfs_add_one()
175 debugfs_create_file("used", 0444, tmp, cma, &cma_used_fops); in cma_debugfs_add_one()
176 debugfs_create_file("maxchunk", 0444, tmp, cma, &cma_maxchunk_fops); in cma_debugfs_add_one()
179 for (r = 0; r < cma->nranges; r++) { in cma_debugfs_add_one()
180 cmr = &cma->ranges[r]; in cma_debugfs_add_one()
187 DIV_ROUND_UP(cma_bitmap_maxno(cma, cmr), in cma_debugfs_add_one()
205 cma_debugfs_root = debugfs_create_dir("cma", NULL); in cma_debugfs_init()