1*e57b9183Scg149915
2*e57b9183Scg149915 /*
3*e57b9183Scg149915 * Copyright 2008 Sun Microsystems, Inc. All rights reserved.
4*e57b9183Scg149915 * Use is subject to license terms.
5*e57b9183Scg149915 */
6*e57b9183Scg149915 /* radeon_mem.c -- Simple GART/fb memory manager for radeon -*- linux-c -*- */
7*e57b9183Scg149915 /*
8*e57b9183Scg149915 * Copyright (C) The Weather Channel, Inc. 2002. All Rights Reserved.
9*e57b9183Scg149915 *
10*e57b9183Scg149915 * The Weather Channel (TM) funded Tungsten Graphics to develop the
11*e57b9183Scg149915 * initial release of the Radeon 8500 driver under the XFree86 license.
12*e57b9183Scg149915 * This notice must be preserved.
13*e57b9183Scg149915 *
14*e57b9183Scg149915 * Permission is hereby granted, free of charge, to any person obtaining a
15*e57b9183Scg149915 * copy of this software and associated documentation files (the "Software"),
16*e57b9183Scg149915 * to deal in the Software without restriction, including without limitation
17*e57b9183Scg149915 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
18*e57b9183Scg149915 * and/or sell copies of the Software, and to permit persons to whom the
19*e57b9183Scg149915 * Software is furnished to do so, subject to the following conditions:
20*e57b9183Scg149915 *
21*e57b9183Scg149915 * The above copyright notice and this permission notice (including the next
22*e57b9183Scg149915 * paragraph) shall be included in all copies or substantial portions of the
23*e57b9183Scg149915 * Software.
24*e57b9183Scg149915 *
25*e57b9183Scg149915 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
26*e57b9183Scg149915 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
27*e57b9183Scg149915 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
28*e57b9183Scg149915 * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
29*e57b9183Scg149915 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
30*e57b9183Scg149915 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
31*e57b9183Scg149915 * DEALINGS IN THE SOFTWARE.
32*e57b9183Scg149915 *
33*e57b9183Scg149915 * Authors:
34*e57b9183Scg149915 * Keith Whitwell <keith@tungstengraphics.com>
35*e57b9183Scg149915 */
36*e57b9183Scg149915
37*e57b9183Scg149915 #pragma ident "%Z%%M% %I% %E% SMI"
38*e57b9183Scg149915
39*e57b9183Scg149915 #include "drmP.h"
40*e57b9183Scg149915 #include "drm.h"
41*e57b9183Scg149915 #include "radeon_drm.h"
42*e57b9183Scg149915 #include "radeon_drv.h"
43*e57b9183Scg149915 #include "radeon_io32.h"
44*e57b9183Scg149915
45*e57b9183Scg149915 /*
46*e57b9183Scg149915 * Very simple allocator for GART memory, working on a static range
47*e57b9183Scg149915 * already mapped into each client's address space.
48*e57b9183Scg149915 */
49*e57b9183Scg149915
50*e57b9183Scg149915 static struct mem_block *
split_block(struct mem_block * p,int start,int size,drm_file_t * filp)51*e57b9183Scg149915 split_block(struct mem_block *p, int start, int size, drm_file_t *filp)
52*e57b9183Scg149915 {
53*e57b9183Scg149915 /* Maybe cut off the start of an existing block */
54*e57b9183Scg149915 if (start > p->start) {
55*e57b9183Scg149915 struct mem_block *newblock =
56*e57b9183Scg149915 drm_alloc(sizeof (*newblock), DRM_MEM_BUFS);
57*e57b9183Scg149915 if (!newblock)
58*e57b9183Scg149915 goto out;
59*e57b9183Scg149915 newblock->start = start;
60*e57b9183Scg149915 newblock->size = p->size - (start - p->start);
61*e57b9183Scg149915 newblock->filp = NULL;
62*e57b9183Scg149915 newblock->next = p->next;
63*e57b9183Scg149915 newblock->prev = p;
64*e57b9183Scg149915 p->next->prev = newblock;
65*e57b9183Scg149915 p->next = newblock;
66*e57b9183Scg149915 p->size -= newblock->size;
67*e57b9183Scg149915 p = newblock;
68*e57b9183Scg149915 }
69*e57b9183Scg149915
70*e57b9183Scg149915 /* Maybe cut off the end of an existing block */
71*e57b9183Scg149915 if (size < p->size) {
72*e57b9183Scg149915 struct mem_block *newblock =
73*e57b9183Scg149915 drm_alloc(sizeof (*newblock), DRM_MEM_BUFS);
74*e57b9183Scg149915 if (!newblock)
75*e57b9183Scg149915 goto out;
76*e57b9183Scg149915 newblock->start = start + size;
77*e57b9183Scg149915 newblock->size = p->size - size;
78*e57b9183Scg149915 newblock->filp = NULL;
79*e57b9183Scg149915 newblock->next = p->next;
80*e57b9183Scg149915 newblock->prev = p;
81*e57b9183Scg149915 p->next->prev = newblock;
82*e57b9183Scg149915 p->next = newblock;
83*e57b9183Scg149915 p->size = size;
84*e57b9183Scg149915 }
85*e57b9183Scg149915
86*e57b9183Scg149915 out:
87*e57b9183Scg149915 /* Our block is in the middle */
88*e57b9183Scg149915 p->filp = filp;
89*e57b9183Scg149915 return (p);
90*e57b9183Scg149915 }
91*e57b9183Scg149915
92*e57b9183Scg149915 static struct mem_block *
alloc_block(struct mem_block * heap,int size,int align2,drm_file_t * filp)93*e57b9183Scg149915 alloc_block(struct mem_block *heap, int size, int align2, drm_file_t *filp)
94*e57b9183Scg149915 {
95*e57b9183Scg149915 struct mem_block *p;
96*e57b9183Scg149915 int mask = (1 << align2) - 1;
97*e57b9183Scg149915
98*e57b9183Scg149915 for (p = heap->next; p != heap; p = p->next) {
99*e57b9183Scg149915 int start = (p->start + mask) & ~mask;
100*e57b9183Scg149915 if (p->filp == 0 && start + size <= p->start + p->size)
101*e57b9183Scg149915 return (split_block(p, start, size, filp));
102*e57b9183Scg149915 }
103*e57b9183Scg149915
104*e57b9183Scg149915 return (NULL);
105*e57b9183Scg149915 }
106*e57b9183Scg149915
107*e57b9183Scg149915 static struct mem_block *
find_block(struct mem_block * heap,int start)108*e57b9183Scg149915 find_block(struct mem_block *heap, int start)
109*e57b9183Scg149915 {
110*e57b9183Scg149915 struct mem_block *p;
111*e57b9183Scg149915
112*e57b9183Scg149915 for (p = heap->next; p != heap; p = p->next)
113*e57b9183Scg149915 if (p->start == start)
114*e57b9183Scg149915 return (p);
115*e57b9183Scg149915
116*e57b9183Scg149915 return (NULL);
117*e57b9183Scg149915 }
118*e57b9183Scg149915
119*e57b9183Scg149915 static void
free_block(struct mem_block * p)120*e57b9183Scg149915 free_block(struct mem_block *p)
121*e57b9183Scg149915 {
122*e57b9183Scg149915 p->filp = NULL;
123*e57b9183Scg149915
124*e57b9183Scg149915 /*
125*e57b9183Scg149915 * Assumes a single contiguous range. Needs a special filp in
126*e57b9183Scg149915 * 'heap' to stop it being subsumed.
127*e57b9183Scg149915 */
128*e57b9183Scg149915 if (p->next->filp == 0) {
129*e57b9183Scg149915 struct mem_block *q = p->next;
130*e57b9183Scg149915 p->size += q->size;
131*e57b9183Scg149915 p->next = q->next;
132*e57b9183Scg149915 p->next->prev = p;
133*e57b9183Scg149915 drm_free(q, sizeof (*q), DRM_MEM_BUFS);
134*e57b9183Scg149915 }
135*e57b9183Scg149915
136*e57b9183Scg149915 if (p->prev->filp == 0) {
137*e57b9183Scg149915 struct mem_block *q = p->prev;
138*e57b9183Scg149915 q->size += p->size;
139*e57b9183Scg149915 q->next = p->next;
140*e57b9183Scg149915 q->next->prev = q;
141*e57b9183Scg149915 drm_free(p, sizeof (*q), DRM_MEM_BUFS);
142*e57b9183Scg149915 }
143*e57b9183Scg149915 }
144*e57b9183Scg149915
145*e57b9183Scg149915 /*
146*e57b9183Scg149915 * Initialize. How to check for an uninitialized heap?
147*e57b9183Scg149915 */
148*e57b9183Scg149915 static int
init_heap(struct mem_block ** heap,int start,int size)149*e57b9183Scg149915 init_heap(struct mem_block **heap, int start, int size)
150*e57b9183Scg149915 {
151*e57b9183Scg149915 struct mem_block *blocks = drm_alloc(sizeof (*blocks), DRM_MEM_BUFS);
152*e57b9183Scg149915
153*e57b9183Scg149915 if (!blocks)
154*e57b9183Scg149915 return (ENOMEM);
155*e57b9183Scg149915
156*e57b9183Scg149915 *heap = drm_alloc(sizeof (**heap), DRM_MEM_BUFS);
157*e57b9183Scg149915 if (!*heap) {
158*e57b9183Scg149915 drm_free(blocks, sizeof (*blocks), DRM_MEM_BUFS);
159*e57b9183Scg149915 return (ENOMEM);
160*e57b9183Scg149915 }
161*e57b9183Scg149915
162*e57b9183Scg149915 blocks->start = start;
163*e57b9183Scg149915 blocks->size = size;
164*e57b9183Scg149915 blocks->filp = NULL;
165*e57b9183Scg149915 blocks->next = blocks->prev = *heap;
166*e57b9183Scg149915
167*e57b9183Scg149915 (void) memset(*heap, 0, sizeof (**heap));
168*e57b9183Scg149915 (*heap)->filp = (drm_file_t *)-1;
169*e57b9183Scg149915 (*heap)->next = (*heap)->prev = blocks;
170*e57b9183Scg149915 return (0);
171*e57b9183Scg149915 }
172*e57b9183Scg149915
173*e57b9183Scg149915 /*
174*e57b9183Scg149915 * Free all blocks associated with the releasing file.
175*e57b9183Scg149915 */
176*e57b9183Scg149915 void
radeon_mem_release(drm_file_t * filp,struct mem_block * heap)177*e57b9183Scg149915 radeon_mem_release(drm_file_t *filp, struct mem_block *heap)
178*e57b9183Scg149915 {
179*e57b9183Scg149915 struct mem_block *p;
180*e57b9183Scg149915
181*e57b9183Scg149915 if (!heap || !heap->next)
182*e57b9183Scg149915 return;
183*e57b9183Scg149915
184*e57b9183Scg149915 for (p = heap->next; p != heap; p = p->next) {
185*e57b9183Scg149915 if (p->filp == filp)
186*e57b9183Scg149915 p->filp = NULL;
187*e57b9183Scg149915 }
188*e57b9183Scg149915
189*e57b9183Scg149915 /*
190*e57b9183Scg149915 * Assumes a single contiguous range. Needs a special filp in
191*e57b9183Scg149915 * 'heap' to stop it being subsumed.
192*e57b9183Scg149915 */
193*e57b9183Scg149915 for (p = heap->next; p != heap; p = p->next) {
194*e57b9183Scg149915 while (p->filp == 0 && p->next->filp == 0) {
195*e57b9183Scg149915 struct mem_block *q = p->next;
196*e57b9183Scg149915 p->size += q->size;
197*e57b9183Scg149915 p->next = q->next;
198*e57b9183Scg149915 p->next->prev = p;
199*e57b9183Scg149915 drm_free(q, sizeof (*q), DRM_MEM_DRIVER);
200*e57b9183Scg149915 }
201*e57b9183Scg149915 }
202*e57b9183Scg149915 }
203*e57b9183Scg149915
204*e57b9183Scg149915 /*
205*e57b9183Scg149915 * Shutdown.
206*e57b9183Scg149915 */
207*e57b9183Scg149915 void
radeon_mem_takedown(struct mem_block ** heap)208*e57b9183Scg149915 radeon_mem_takedown(struct mem_block **heap)
209*e57b9183Scg149915 {
210*e57b9183Scg149915 struct mem_block *p;
211*e57b9183Scg149915
212*e57b9183Scg149915 if (!*heap)
213*e57b9183Scg149915 return;
214*e57b9183Scg149915
215*e57b9183Scg149915 for (p = (*heap)->next; p != *heap; ) {
216*e57b9183Scg149915 struct mem_block *q = p;
217*e57b9183Scg149915 p = p->next;
218*e57b9183Scg149915 drm_free(q, sizeof (*q), DRM_MEM_DRIVER);
219*e57b9183Scg149915 }
220*e57b9183Scg149915
221*e57b9183Scg149915 drm_free(*heap, sizeof (**heap), DRM_MEM_DRIVER);
222*e57b9183Scg149915 *heap = NULL;
223*e57b9183Scg149915 }
224*e57b9183Scg149915
225*e57b9183Scg149915 /* IOCTL HANDLERS */
226*e57b9183Scg149915
227*e57b9183Scg149915 static struct mem_block **
get_heap(drm_radeon_private_t * dev_priv,int region)228*e57b9183Scg149915 get_heap(drm_radeon_private_t *dev_priv, int region)
229*e57b9183Scg149915 {
230*e57b9183Scg149915 switch (region) {
231*e57b9183Scg149915 case RADEON_MEM_REGION_GART:
232*e57b9183Scg149915 return (&dev_priv->gart_heap);
233*e57b9183Scg149915 case RADEON_MEM_REGION_FB:
234*e57b9183Scg149915 return (&dev_priv->fb_heap);
235*e57b9183Scg149915 default:
236*e57b9183Scg149915 return (NULL);
237*e57b9183Scg149915 }
238*e57b9183Scg149915 }
239*e57b9183Scg149915
240*e57b9183Scg149915 /*ARGSUSED*/
241*e57b9183Scg149915 int
radeon_mem_alloc(DRM_IOCTL_ARGS)242*e57b9183Scg149915 radeon_mem_alloc(DRM_IOCTL_ARGS)
243*e57b9183Scg149915 {
244*e57b9183Scg149915 DRM_DEVICE;
245*e57b9183Scg149915 drm_radeon_private_t *dev_priv = dev->dev_private;
246*e57b9183Scg149915 drm_radeon_mem_alloc_t alloc;
247*e57b9183Scg149915 struct mem_block *block, **heap;
248*e57b9183Scg149915
249*e57b9183Scg149915 if (!dev_priv) {
250*e57b9183Scg149915 DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
251*e57b9183Scg149915 return (EINVAL);
252*e57b9183Scg149915 }
253*e57b9183Scg149915
254*e57b9183Scg149915 #ifdef _MULTI_DATAMODEL
255*e57b9183Scg149915 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) {
256*e57b9183Scg149915 drm_radeon_mem_alloc_32_t alloc32;
257*e57b9183Scg149915
258*e57b9183Scg149915 DRM_COPYFROM_WITH_RETURN(&alloc32, (void *) data,
259*e57b9183Scg149915 sizeof (alloc32));
260*e57b9183Scg149915 alloc.region = alloc32.region;
261*e57b9183Scg149915 alloc.alignment = alloc32.alignment;
262*e57b9183Scg149915 alloc.size = alloc32.size;
263*e57b9183Scg149915 alloc.region_offset = (void *)(uintptr_t)alloc32.region_offset;
264*e57b9183Scg149915 } else {
265*e57b9183Scg149915 #endif
266*e57b9183Scg149915 DRM_COPYFROM_WITH_RETURN(&alloc, (void *) data, sizeof (alloc));
267*e57b9183Scg149915 #ifdef _MULTI_DATAMODEL
268*e57b9183Scg149915 }
269*e57b9183Scg149915 #endif
270*e57b9183Scg149915
271*e57b9183Scg149915 heap = get_heap(dev_priv, alloc.region);
272*e57b9183Scg149915 if (!heap || !*heap)
273*e57b9183Scg149915 return (EFAULT);
274*e57b9183Scg149915
275*e57b9183Scg149915 /*
276*e57b9183Scg149915 * Make things easier on ourselves: all allocations at least
277*e57b9183Scg149915 * 4k aligned.
278*e57b9183Scg149915 */
279*e57b9183Scg149915 if (alloc.alignment < 12)
280*e57b9183Scg149915 alloc.alignment = 12;
281*e57b9183Scg149915
282*e57b9183Scg149915 block = alloc_block(*heap, alloc.size, alloc.alignment, fpriv);
283*e57b9183Scg149915
284*e57b9183Scg149915 if (!block)
285*e57b9183Scg149915 return (ENOMEM);
286*e57b9183Scg149915
287*e57b9183Scg149915 if (DRM_COPY_TO_USER(alloc.region_offset, &block->start,
288*e57b9183Scg149915 sizeof (int))) {
289*e57b9183Scg149915 DRM_ERROR("copy_to_user\n");
290*e57b9183Scg149915 return (EFAULT);
291*e57b9183Scg149915 }
292*e57b9183Scg149915
293*e57b9183Scg149915 return (0);
294*e57b9183Scg149915 }
295*e57b9183Scg149915
296*e57b9183Scg149915 /*ARGSUSED*/
297*e57b9183Scg149915 int
radeon_mem_free(DRM_IOCTL_ARGS)298*e57b9183Scg149915 radeon_mem_free(DRM_IOCTL_ARGS)
299*e57b9183Scg149915 {
300*e57b9183Scg149915 DRM_DEVICE;
301*e57b9183Scg149915 drm_radeon_private_t *dev_priv = dev->dev_private;
302*e57b9183Scg149915 drm_radeon_mem_free_t memfree;
303*e57b9183Scg149915 struct mem_block *block, **heap;
304*e57b9183Scg149915
305*e57b9183Scg149915 if (!dev_priv) {
306*e57b9183Scg149915 DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
307*e57b9183Scg149915 return (EINVAL);
308*e57b9183Scg149915 }
309*e57b9183Scg149915
310*e57b9183Scg149915 DRM_COPYFROM_WITH_RETURN(&memfree, (void *) data, sizeof (memfree));
311*e57b9183Scg149915
312*e57b9183Scg149915 heap = get_heap(dev_priv, memfree.region);
313*e57b9183Scg149915 if (!heap || !*heap)
314*e57b9183Scg149915 return (EFAULT);
315*e57b9183Scg149915
316*e57b9183Scg149915 block = find_block(*heap, memfree.region_offset);
317*e57b9183Scg149915 if (!block)
318*e57b9183Scg149915 return (EFAULT);
319*e57b9183Scg149915
320*e57b9183Scg149915 if (block->filp != fpriv)
321*e57b9183Scg149915 return (EPERM);
322*e57b9183Scg149915
323*e57b9183Scg149915 free_block(block);
324*e57b9183Scg149915 return (0);
325*e57b9183Scg149915 }
326*e57b9183Scg149915
327*e57b9183Scg149915 /*ARGSUSED*/
328*e57b9183Scg149915 int
radeon_mem_init_heap(DRM_IOCTL_ARGS)329*e57b9183Scg149915 radeon_mem_init_heap(DRM_IOCTL_ARGS)
330*e57b9183Scg149915 {
331*e57b9183Scg149915 DRM_DEVICE;
332*e57b9183Scg149915 drm_radeon_private_t *dev_priv = dev->dev_private;
333*e57b9183Scg149915 drm_radeon_mem_init_heap_t initheap;
334*e57b9183Scg149915 struct mem_block **heap;
335*e57b9183Scg149915
336*e57b9183Scg149915 if (!dev_priv) {
337*e57b9183Scg149915 DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
338*e57b9183Scg149915 return (EINVAL);
339*e57b9183Scg149915 }
340*e57b9183Scg149915
341*e57b9183Scg149915 DRM_COPYFROM_WITH_RETURN(&initheap, (void *) data, sizeof (initheap));
342*e57b9183Scg149915
343*e57b9183Scg149915 heap = get_heap(dev_priv, initheap.region);
344*e57b9183Scg149915 if (!heap)
345*e57b9183Scg149915 return (EFAULT);
346*e57b9183Scg149915
347*e57b9183Scg149915 if (*heap) {
348*e57b9183Scg149915 DRM_ERROR("heap already initialized?");
349*e57b9183Scg149915 return (EFAULT);
350*e57b9183Scg149915 }
351*e57b9183Scg149915
352*e57b9183Scg149915 return (init_heap(heap, initheap.start, initheap.size));
353*e57b9183Scg149915 }
354