1 /* 2 * CMA DebugFS Interface 3 * 4 * Copyright (c) 2015 Sasha Levin <sasha.levin@oracle.com> 5 */ 6 7 8 #include <linux/debugfs.h> 9 #include <linux/cma.h> 10 #include <linux/list.h> 11 #include <linux/kernel.h> 12 #include <linux/slab.h> 13 #include <linux/mm_types.h> 14 15 #include "cma.h" 16 17 struct cma_mem { 18 struct hlist_node node; 19 struct page *p; 20 unsigned long n; 21 }; 22 23 static struct dentry *cma_debugfs_root; 24 25 static int cma_debugfs_get(void *data, u64 *val) 26 { 27 unsigned long *p = data; 28 29 *val = *p; 30 31 return 0; 32 } 33 DEFINE_SIMPLE_ATTRIBUTE(cma_debugfs_fops, cma_debugfs_get, NULL, "%llu\n"); 34 35 static int cma_used_get(void *data, u64 *val) 36 { 37 struct cma *cma = data; 38 unsigned long used; 39 40 mutex_lock(&cma->lock); 41 /* pages counter is smaller than sizeof(int) */ 42 used = bitmap_weight(cma->bitmap, (int)cma_bitmap_maxno(cma)); 43 mutex_unlock(&cma->lock); 44 *val = (u64)used << cma->order_per_bit; 45 46 return 0; 47 } 48 DEFINE_SIMPLE_ATTRIBUTE(cma_used_fops, cma_used_get, NULL, "%llu\n"); 49 50 static int cma_maxchunk_get(void *data, u64 *val) 51 { 52 struct cma *cma = data; 53 unsigned long maxchunk = 0; 54 unsigned long start, end = 0; 55 unsigned long bitmap_maxno = cma_bitmap_maxno(cma); 56 57 mutex_lock(&cma->lock); 58 for (;;) { 59 start = find_next_zero_bit(cma->bitmap, bitmap_maxno, end); 60 if (start >= cma->count) 61 break; 62 end = find_next_bit(cma->bitmap, bitmap_maxno, start); 63 maxchunk = max(end - start, maxchunk); 64 } 65 mutex_unlock(&cma->lock); 66 *val = (u64)maxchunk << cma->order_per_bit; 67 68 return 0; 69 } 70 DEFINE_SIMPLE_ATTRIBUTE(cma_maxchunk_fops, cma_maxchunk_get, NULL, "%llu\n"); 71 72 static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem) 73 { 74 spin_lock(&cma->mem_head_lock); 75 hlist_add_head(&mem->node, &cma->mem_head); 76 spin_unlock(&cma->mem_head_lock); 77 } 78 79 static struct cma_mem *cma_get_entry_from_list(struct cma *cma) 80 { 81 struct cma_mem *mem = NULL; 82 83 spin_lock(&cma->mem_head_lock); 84 if (!hlist_empty(&cma->mem_head)) { 85 mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); 86 hlist_del_init(&mem->node); 87 } 88 spin_unlock(&cma->mem_head_lock); 89 90 return mem; 91 } 92 93 static int cma_free_mem(struct cma *cma, int count) 94 { 95 struct cma_mem *mem = NULL; 96 97 while (count) { 98 mem = cma_get_entry_from_list(cma); 99 if (mem == NULL) 100 return 0; 101 102 if (mem->n <= count) { 103 cma_release(cma, mem->p, mem->n); 104 count -= mem->n; 105 kfree(mem); 106 } else if (cma->order_per_bit == 0) { 107 cma_release(cma, mem->p, count); 108 mem->p += count; 109 mem->n -= count; 110 count = 0; 111 cma_add_to_cma_mem_list(cma, mem); 112 } else { 113 pr_debug("cma: cannot release partial block when order_per_bit != 0\n"); 114 cma_add_to_cma_mem_list(cma, mem); 115 break; 116 } 117 } 118 119 return 0; 120 121 } 122 123 static int cma_free_write(void *data, u64 val) 124 { 125 int pages = val; 126 struct cma *cma = data; 127 128 return cma_free_mem(cma, pages); 129 } 130 DEFINE_SIMPLE_ATTRIBUTE(cma_free_fops, NULL, cma_free_write, "%llu\n"); 131 132 static int cma_alloc_mem(struct cma *cma, int count) 133 { 134 struct cma_mem *mem; 135 struct page *p; 136 137 mem = kzalloc(sizeof(*mem), GFP_KERNEL); 138 if (!mem) 139 return -ENOMEM; 140 141 p = cma_alloc(cma, count, 0); 142 if (!p) { 143 kfree(mem); 144 return -ENOMEM; 145 } 146 147 mem->p = p; 148 mem->n = count; 149 150 cma_add_to_cma_mem_list(cma, mem); 151 152 return 0; 153 } 154 155 static int cma_alloc_write(void *data, u64 val) 156 { 157 int pages = val; 158 struct cma *cma = data; 159 160 return cma_alloc_mem(cma, pages); 161 } 162 DEFINE_SIMPLE_ATTRIBUTE(cma_alloc_fops, NULL, cma_alloc_write, "%llu\n"); 163 164 static void cma_debugfs_add_one(struct cma *cma, int idx) 165 { 166 struct dentry *tmp; 167 char name[16]; 168 int u32s; 169 170 sprintf(name, "cma-%d", idx); 171 172 tmp = debugfs_create_dir(name, cma_debugfs_root); 173 174 debugfs_create_file("alloc", S_IWUSR, tmp, cma, 175 &cma_alloc_fops); 176 177 debugfs_create_file("free", S_IWUSR, tmp, cma, 178 &cma_free_fops); 179 180 debugfs_create_file("base_pfn", S_IRUGO, tmp, 181 &cma->base_pfn, &cma_debugfs_fops); 182 debugfs_create_file("count", S_IRUGO, tmp, 183 &cma->count, &cma_debugfs_fops); 184 debugfs_create_file("order_per_bit", S_IRUGO, tmp, 185 &cma->order_per_bit, &cma_debugfs_fops); 186 debugfs_create_file("used", S_IRUGO, tmp, cma, &cma_used_fops); 187 debugfs_create_file("maxchunk", S_IRUGO, tmp, cma, &cma_maxchunk_fops); 188 189 u32s = DIV_ROUND_UP(cma_bitmap_maxno(cma), BITS_PER_BYTE * sizeof(u32)); 190 debugfs_create_u32_array("bitmap", S_IRUGO, tmp, (u32*)cma->bitmap, u32s); 191 } 192 193 static int __init cma_debugfs_init(void) 194 { 195 int i; 196 197 cma_debugfs_root = debugfs_create_dir("cma", NULL); 198 if (!cma_debugfs_root) 199 return -ENOMEM; 200 201 for (i = 0; i < cma_area_count; i++) 202 cma_debugfs_add_one(&cma_areas[i], i); 203 204 return 0; 205 } 206 late_initcall(cma_debugfs_init); 207