xref: /titanic_41/usr/src/uts/intel/io/drm/radeon_mem.c (revision 40db2e2b777b79f3dd0d6d9629593a07f86b9c0a)
1 
2 /*
3  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
4  * Use is subject to license terms.
5  */
6 /* radeon_mem.c -- Simple GART/fb memory manager for radeon -*- linux-c -*- */
7 /*
8  * Copyright (C) The Weather Channel, Inc.  2002.  All Rights Reserved.
9  *
10  * The Weather Channel (TM) funded Tungsten Graphics to develop the
11  * initial release of the Radeon 8500 driver under the XFree86 license.
12  * This notice must be preserved.
13  *
14  * Permission is hereby granted, free of charge, to any person obtaining a
15  * copy of this software and associated documentation files (the "Software"),
16  * to deal in the Software without restriction, including without limitation
17  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
18  * and/or sell copies of the Software, and to permit persons to whom the
19  * Software is furnished to do so, subject to the following conditions:
20  *
21  * The above copyright notice and this permission notice (including the next
22  * paragraph) shall be included in all copies or substantial portions of the
23  * Software.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
26  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
27  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
28  * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
29  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
30  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
31  * DEALINGS IN THE SOFTWARE.
32  *
33  * Authors:
34  *    Keith Whitwell <keith@tungstengraphics.com>
35  */
36 
37 #pragma ident	"%Z%%M%	%I%	%E% SMI"
38 
39 #include "drmP.h"
40 #include "drm.h"
41 #include "radeon_drm.h"
42 #include "radeon_drv.h"
43 #include "radeon_io32.h"
44 
45 /*
46  * Very simple allocator for GART memory, working on a static range
47  * already mapped into each client's address space.
48  */
49 
50 static struct mem_block *
51 split_block(struct mem_block *p, int start, int size, drm_file_t *filp)
52 {
53 	/* Maybe cut off the start of an existing block */
54 	if (start > p->start) {
55 		struct mem_block *newblock =
56 		    drm_alloc(sizeof (*newblock), DRM_MEM_BUFS);
57 		if (!newblock)
58 			goto out;
59 		newblock->start = start;
60 		newblock->size = p->size - (start - p->start);
61 		newblock->filp = NULL;
62 		newblock->next = p->next;
63 		newblock->prev = p;
64 		p->next->prev = newblock;
65 		p->next = newblock;
66 		p->size -= newblock->size;
67 		p = newblock;
68 	}
69 
70 	/* Maybe cut off the end of an existing block */
71 	if (size < p->size) {
72 		struct mem_block *newblock =
73 		    drm_alloc(sizeof (*newblock), DRM_MEM_BUFS);
74 		if (!newblock)
75 			goto out;
76 		newblock->start = start + size;
77 		newblock->size = p->size - size;
78 		newblock->filp = NULL;
79 		newblock->next = p->next;
80 		newblock->prev = p;
81 		p->next->prev = newblock;
82 		p->next = newblock;
83 		p->size = size;
84 	}
85 
86 out:
87 	/* Our block is in the middle */
88 	p->filp = filp;
89 	return (p);
90 }
91 
92 static struct mem_block *
93 alloc_block(struct mem_block *heap, int size, int align2, drm_file_t *filp)
94 {
95 	struct mem_block *p;
96 	int mask = (1 << align2) - 1;
97 
98 	for (p = heap->next; p != heap; p = p->next) {
99 		int start = (p->start + mask) & ~mask;
100 		if (p->filp == 0 && start + size <= p->start + p->size)
101 			return (split_block(p, start, size, filp));
102 	}
103 
104 	return (NULL);
105 }
106 
107 static struct mem_block *
108 find_block(struct mem_block *heap, int start)
109 {
110 	struct mem_block *p;
111 
112 	for (p = heap->next; p != heap; p = p->next)
113 		if (p->start == start)
114 			return (p);
115 
116 	return (NULL);
117 }
118 
119 static void
120 free_block(struct mem_block *p)
121 {
122 	p->filp = NULL;
123 
124 	/*
125 	 * Assumes a single contiguous range.  Needs a special filp in
126 	 * 'heap' to stop it being subsumed.
127 	 */
128 	if (p->next->filp == 0) {
129 		struct mem_block *q = p->next;
130 		p->size += q->size;
131 		p->next = q->next;
132 		p->next->prev = p;
133 		drm_free(q, sizeof (*q), DRM_MEM_BUFS);
134 	}
135 
136 	if (p->prev->filp == 0) {
137 		struct mem_block *q = p->prev;
138 		q->size += p->size;
139 		q->next = p->next;
140 		q->next->prev = q;
141 		drm_free(p, sizeof (*q), DRM_MEM_BUFS);
142 	}
143 }
144 
145 /*
146  * Initialize.  How to check for an uninitialized heap?
147  */
148 static int
149 init_heap(struct mem_block **heap, int start, int size)
150 {
151 	struct mem_block *blocks = drm_alloc(sizeof (*blocks), DRM_MEM_BUFS);
152 
153 	if (!blocks)
154 		return (ENOMEM);
155 
156 	*heap = drm_alloc(sizeof (**heap), DRM_MEM_BUFS);
157 	if (!*heap) {
158 		drm_free(blocks, sizeof (*blocks), DRM_MEM_BUFS);
159 		return (ENOMEM);
160 	}
161 
162 	blocks->start = start;
163 	blocks->size = size;
164 	blocks->filp = NULL;
165 	blocks->next = blocks->prev = *heap;
166 
167 	(void) memset(*heap, 0, sizeof (**heap));
168 	(*heap)->filp = (drm_file_t *)-1;
169 	(*heap)->next = (*heap)->prev = blocks;
170 	return (0);
171 }
172 
173 /*
174  * Free all blocks associated with the releasing file.
175  */
176 void
177 radeon_mem_release(drm_file_t *filp, struct mem_block *heap)
178 {
179 	struct mem_block *p;
180 
181 	if (!heap || !heap->next)
182 		return;
183 
184 	for (p = heap->next; p != heap; p = p->next) {
185 		if (p->filp == filp)
186 			p->filp = NULL;
187 	}
188 
189 	/*
190 	 * Assumes a single contiguous range.  Needs a special filp in
191 	 * 'heap' to stop it being subsumed.
192 	 */
193 	for (p = heap->next; p != heap; p = p->next) {
194 		while (p->filp == 0 && p->next->filp == 0) {
195 			struct mem_block *q = p->next;
196 			p->size += q->size;
197 			p->next = q->next;
198 			p->next->prev = p;
199 			drm_free(q, sizeof (*q), DRM_MEM_DRIVER);
200 		}
201 	}
202 }
203 
204 /*
205  * Shutdown.
206  */
207 void
208 radeon_mem_takedown(struct mem_block **heap)
209 {
210 	struct mem_block *p;
211 
212 	if (!*heap)
213 		return;
214 
215 	for (p = (*heap)->next; p != *heap; ) {
216 		struct mem_block *q = p;
217 		p = p->next;
218 		drm_free(q, sizeof (*q), DRM_MEM_DRIVER);
219 	}
220 
221 	drm_free(*heap, sizeof (**heap), DRM_MEM_DRIVER);
222 	*heap = NULL;
223 }
224 
225 /* IOCTL HANDLERS */
226 
227 static struct mem_block **
228 get_heap(drm_radeon_private_t *dev_priv, int region)
229 {
230 	switch (region) {
231 	case RADEON_MEM_REGION_GART:
232 		return (&dev_priv->gart_heap);
233 	case RADEON_MEM_REGION_FB:
234 		return (&dev_priv->fb_heap);
235 	default:
236 		return (NULL);
237 	}
238 }
239 
240 /*ARGSUSED*/
241 int
242 radeon_mem_alloc(DRM_IOCTL_ARGS)
243 {
244 	DRM_DEVICE;
245 	drm_radeon_private_t *dev_priv = dev->dev_private;
246 	drm_radeon_mem_alloc_t alloc;
247 	struct mem_block *block, **heap;
248 
249 	if (!dev_priv) {
250 		DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
251 		return (EINVAL);
252 	}
253 
254 #ifdef _MULTI_DATAMODEL
255 	if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) {
256 		drm_radeon_mem_alloc_32_t alloc32;
257 
258 		DRM_COPYFROM_WITH_RETURN(&alloc32, (void *) data,
259 		    sizeof (alloc32));
260 		alloc.region = alloc32.region;
261 		alloc.alignment = alloc32.alignment;
262 		alloc.size = alloc32.size;
263 		alloc.region_offset = (void *)(uintptr_t)alloc32.region_offset;
264 	} else {
265 #endif
266 		DRM_COPYFROM_WITH_RETURN(&alloc, (void *) data, sizeof (alloc));
267 #ifdef _MULTI_DATAMODEL
268 	}
269 #endif
270 
271 	heap = get_heap(dev_priv, alloc.region);
272 	if (!heap || !*heap)
273 		return (EFAULT);
274 
275 	/*
276 	 * Make things easier on ourselves: all allocations at least
277 	 * 4k aligned.
278 	 */
279 	if (alloc.alignment < 12)
280 		alloc.alignment = 12;
281 
282 	block = alloc_block(*heap, alloc.size, alloc.alignment, fpriv);
283 
284 	if (!block)
285 		return (ENOMEM);
286 
287 	if (DRM_COPY_TO_USER(alloc.region_offset, &block->start,
288 	    sizeof (int))) {
289 		DRM_ERROR("copy_to_user\n");
290 		return (EFAULT);
291 	}
292 
293 	return (0);
294 }
295 
296 /*ARGSUSED*/
297 int
298 radeon_mem_free(DRM_IOCTL_ARGS)
299 {
300 	DRM_DEVICE;
301 	drm_radeon_private_t *dev_priv = dev->dev_private;
302 	drm_radeon_mem_free_t memfree;
303 	struct mem_block *block, **heap;
304 
305 	if (!dev_priv) {
306 		DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
307 		return (EINVAL);
308 	}
309 
310 	DRM_COPYFROM_WITH_RETURN(&memfree, (void *) data, sizeof (memfree));
311 
312 	heap = get_heap(dev_priv, memfree.region);
313 	if (!heap || !*heap)
314 		return (EFAULT);
315 
316 	block = find_block(*heap, memfree.region_offset);
317 	if (!block)
318 		return (EFAULT);
319 
320 	if (block->filp != fpriv)
321 		return (EPERM);
322 
323 	free_block(block);
324 	return (0);
325 }
326 
327 /*ARGSUSED*/
328 int
329 radeon_mem_init_heap(DRM_IOCTL_ARGS)
330 {
331 	DRM_DEVICE;
332 	drm_radeon_private_t *dev_priv = dev->dev_private;
333 	drm_radeon_mem_init_heap_t initheap;
334 	struct mem_block **heap;
335 
336 	if (!dev_priv) {
337 		DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
338 		return (EINVAL);
339 	}
340 
341 	DRM_COPYFROM_WITH_RETURN(&initheap, (void *) data, sizeof (initheap));
342 
343 	heap = get_heap(dev_priv, initheap.region);
344 	if (!heap)
345 		return (EFAULT);
346 
347 	if (*heap) {
348 		DRM_ERROR("heap already initialized?");
349 		return (EFAULT);
350 	}
351 
352 	return (init_heap(heap, initheap.start, initheap.size));
353 }
354