xref: /titanic_44/usr/src/uts/intel/io/drm/radeon_mem.c (revision e57b9183811d515e3bbcd1a104516f0102fde114)
1*e57b9183Scg149915 
2*e57b9183Scg149915 /*
3*e57b9183Scg149915  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
4*e57b9183Scg149915  * Use is subject to license terms.
5*e57b9183Scg149915  */
6*e57b9183Scg149915 /* radeon_mem.c -- Simple GART/fb memory manager for radeon -*- linux-c -*- */
7*e57b9183Scg149915 /*
8*e57b9183Scg149915  * Copyright (C) The Weather Channel, Inc.  2002.  All Rights Reserved.
9*e57b9183Scg149915  *
10*e57b9183Scg149915  * The Weather Channel (TM) funded Tungsten Graphics to develop the
11*e57b9183Scg149915  * initial release of the Radeon 8500 driver under the XFree86 license.
12*e57b9183Scg149915  * This notice must be preserved.
13*e57b9183Scg149915  *
14*e57b9183Scg149915  * Permission is hereby granted, free of charge, to any person obtaining a
15*e57b9183Scg149915  * copy of this software and associated documentation files (the "Software"),
16*e57b9183Scg149915  * to deal in the Software without restriction, including without limitation
17*e57b9183Scg149915  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
18*e57b9183Scg149915  * and/or sell copies of the Software, and to permit persons to whom the
19*e57b9183Scg149915  * Software is furnished to do so, subject to the following conditions:
20*e57b9183Scg149915  *
21*e57b9183Scg149915  * The above copyright notice and this permission notice (including the next
22*e57b9183Scg149915  * paragraph) shall be included in all copies or substantial portions of the
23*e57b9183Scg149915  * Software.
24*e57b9183Scg149915  *
25*e57b9183Scg149915  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
26*e57b9183Scg149915  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
27*e57b9183Scg149915  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
28*e57b9183Scg149915  * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
29*e57b9183Scg149915  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
30*e57b9183Scg149915  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
31*e57b9183Scg149915  * DEALINGS IN THE SOFTWARE.
32*e57b9183Scg149915  *
33*e57b9183Scg149915  * Authors:
34*e57b9183Scg149915  *    Keith Whitwell <keith@tungstengraphics.com>
35*e57b9183Scg149915  */
36*e57b9183Scg149915 
37*e57b9183Scg149915 #pragma ident	"%Z%%M%	%I%	%E% SMI"
38*e57b9183Scg149915 
39*e57b9183Scg149915 #include "drmP.h"
40*e57b9183Scg149915 #include "drm.h"
41*e57b9183Scg149915 #include "radeon_drm.h"
42*e57b9183Scg149915 #include "radeon_drv.h"
43*e57b9183Scg149915 #include "radeon_io32.h"
44*e57b9183Scg149915 
45*e57b9183Scg149915 /*
46*e57b9183Scg149915  * Very simple allocator for GART memory, working on a static range
47*e57b9183Scg149915  * already mapped into each client's address space.
48*e57b9183Scg149915  */
49*e57b9183Scg149915 
50*e57b9183Scg149915 static struct mem_block *
split_block(struct mem_block * p,int start,int size,drm_file_t * filp)51*e57b9183Scg149915 split_block(struct mem_block *p, int start, int size, drm_file_t *filp)
52*e57b9183Scg149915 {
53*e57b9183Scg149915 	/* Maybe cut off the start of an existing block */
54*e57b9183Scg149915 	if (start > p->start) {
55*e57b9183Scg149915 		struct mem_block *newblock =
56*e57b9183Scg149915 		    drm_alloc(sizeof (*newblock), DRM_MEM_BUFS);
57*e57b9183Scg149915 		if (!newblock)
58*e57b9183Scg149915 			goto out;
59*e57b9183Scg149915 		newblock->start = start;
60*e57b9183Scg149915 		newblock->size = p->size - (start - p->start);
61*e57b9183Scg149915 		newblock->filp = NULL;
62*e57b9183Scg149915 		newblock->next = p->next;
63*e57b9183Scg149915 		newblock->prev = p;
64*e57b9183Scg149915 		p->next->prev = newblock;
65*e57b9183Scg149915 		p->next = newblock;
66*e57b9183Scg149915 		p->size -= newblock->size;
67*e57b9183Scg149915 		p = newblock;
68*e57b9183Scg149915 	}
69*e57b9183Scg149915 
70*e57b9183Scg149915 	/* Maybe cut off the end of an existing block */
71*e57b9183Scg149915 	if (size < p->size) {
72*e57b9183Scg149915 		struct mem_block *newblock =
73*e57b9183Scg149915 		    drm_alloc(sizeof (*newblock), DRM_MEM_BUFS);
74*e57b9183Scg149915 		if (!newblock)
75*e57b9183Scg149915 			goto out;
76*e57b9183Scg149915 		newblock->start = start + size;
77*e57b9183Scg149915 		newblock->size = p->size - size;
78*e57b9183Scg149915 		newblock->filp = NULL;
79*e57b9183Scg149915 		newblock->next = p->next;
80*e57b9183Scg149915 		newblock->prev = p;
81*e57b9183Scg149915 		p->next->prev = newblock;
82*e57b9183Scg149915 		p->next = newblock;
83*e57b9183Scg149915 		p->size = size;
84*e57b9183Scg149915 	}
85*e57b9183Scg149915 
86*e57b9183Scg149915 out:
87*e57b9183Scg149915 	/* Our block is in the middle */
88*e57b9183Scg149915 	p->filp = filp;
89*e57b9183Scg149915 	return (p);
90*e57b9183Scg149915 }
91*e57b9183Scg149915 
92*e57b9183Scg149915 static struct mem_block *
alloc_block(struct mem_block * heap,int size,int align2,drm_file_t * filp)93*e57b9183Scg149915 alloc_block(struct mem_block *heap, int size, int align2, drm_file_t *filp)
94*e57b9183Scg149915 {
95*e57b9183Scg149915 	struct mem_block *p;
96*e57b9183Scg149915 	int mask = (1 << align2) - 1;
97*e57b9183Scg149915 
98*e57b9183Scg149915 	for (p = heap->next; p != heap; p = p->next) {
99*e57b9183Scg149915 		int start = (p->start + mask) & ~mask;
100*e57b9183Scg149915 		if (p->filp == 0 && start + size <= p->start + p->size)
101*e57b9183Scg149915 			return (split_block(p, start, size, filp));
102*e57b9183Scg149915 	}
103*e57b9183Scg149915 
104*e57b9183Scg149915 	return (NULL);
105*e57b9183Scg149915 }
106*e57b9183Scg149915 
107*e57b9183Scg149915 static struct mem_block *
find_block(struct mem_block * heap,int start)108*e57b9183Scg149915 find_block(struct mem_block *heap, int start)
109*e57b9183Scg149915 {
110*e57b9183Scg149915 	struct mem_block *p;
111*e57b9183Scg149915 
112*e57b9183Scg149915 	for (p = heap->next; p != heap; p = p->next)
113*e57b9183Scg149915 		if (p->start == start)
114*e57b9183Scg149915 			return (p);
115*e57b9183Scg149915 
116*e57b9183Scg149915 	return (NULL);
117*e57b9183Scg149915 }
118*e57b9183Scg149915 
119*e57b9183Scg149915 static void
free_block(struct mem_block * p)120*e57b9183Scg149915 free_block(struct mem_block *p)
121*e57b9183Scg149915 {
122*e57b9183Scg149915 	p->filp = NULL;
123*e57b9183Scg149915 
124*e57b9183Scg149915 	/*
125*e57b9183Scg149915 	 * Assumes a single contiguous range.  Needs a special filp in
126*e57b9183Scg149915 	 * 'heap' to stop it being subsumed.
127*e57b9183Scg149915 	 */
128*e57b9183Scg149915 	if (p->next->filp == 0) {
129*e57b9183Scg149915 		struct mem_block *q = p->next;
130*e57b9183Scg149915 		p->size += q->size;
131*e57b9183Scg149915 		p->next = q->next;
132*e57b9183Scg149915 		p->next->prev = p;
133*e57b9183Scg149915 		drm_free(q, sizeof (*q), DRM_MEM_BUFS);
134*e57b9183Scg149915 	}
135*e57b9183Scg149915 
136*e57b9183Scg149915 	if (p->prev->filp == 0) {
137*e57b9183Scg149915 		struct mem_block *q = p->prev;
138*e57b9183Scg149915 		q->size += p->size;
139*e57b9183Scg149915 		q->next = p->next;
140*e57b9183Scg149915 		q->next->prev = q;
141*e57b9183Scg149915 		drm_free(p, sizeof (*q), DRM_MEM_BUFS);
142*e57b9183Scg149915 	}
143*e57b9183Scg149915 }
144*e57b9183Scg149915 
145*e57b9183Scg149915 /*
146*e57b9183Scg149915  * Initialize.  How to check for an uninitialized heap?
147*e57b9183Scg149915  */
148*e57b9183Scg149915 static int
init_heap(struct mem_block ** heap,int start,int size)149*e57b9183Scg149915 init_heap(struct mem_block **heap, int start, int size)
150*e57b9183Scg149915 {
151*e57b9183Scg149915 	struct mem_block *blocks = drm_alloc(sizeof (*blocks), DRM_MEM_BUFS);
152*e57b9183Scg149915 
153*e57b9183Scg149915 	if (!blocks)
154*e57b9183Scg149915 		return (ENOMEM);
155*e57b9183Scg149915 
156*e57b9183Scg149915 	*heap = drm_alloc(sizeof (**heap), DRM_MEM_BUFS);
157*e57b9183Scg149915 	if (!*heap) {
158*e57b9183Scg149915 		drm_free(blocks, sizeof (*blocks), DRM_MEM_BUFS);
159*e57b9183Scg149915 		return (ENOMEM);
160*e57b9183Scg149915 	}
161*e57b9183Scg149915 
162*e57b9183Scg149915 	blocks->start = start;
163*e57b9183Scg149915 	blocks->size = size;
164*e57b9183Scg149915 	blocks->filp = NULL;
165*e57b9183Scg149915 	blocks->next = blocks->prev = *heap;
166*e57b9183Scg149915 
167*e57b9183Scg149915 	(void) memset(*heap, 0, sizeof (**heap));
168*e57b9183Scg149915 	(*heap)->filp = (drm_file_t *)-1;
169*e57b9183Scg149915 	(*heap)->next = (*heap)->prev = blocks;
170*e57b9183Scg149915 	return (0);
171*e57b9183Scg149915 }
172*e57b9183Scg149915 
173*e57b9183Scg149915 /*
174*e57b9183Scg149915  * Free all blocks associated with the releasing file.
175*e57b9183Scg149915  */
176*e57b9183Scg149915 void
radeon_mem_release(drm_file_t * filp,struct mem_block * heap)177*e57b9183Scg149915 radeon_mem_release(drm_file_t *filp, struct mem_block *heap)
178*e57b9183Scg149915 {
179*e57b9183Scg149915 	struct mem_block *p;
180*e57b9183Scg149915 
181*e57b9183Scg149915 	if (!heap || !heap->next)
182*e57b9183Scg149915 		return;
183*e57b9183Scg149915 
184*e57b9183Scg149915 	for (p = heap->next; p != heap; p = p->next) {
185*e57b9183Scg149915 		if (p->filp == filp)
186*e57b9183Scg149915 			p->filp = NULL;
187*e57b9183Scg149915 	}
188*e57b9183Scg149915 
189*e57b9183Scg149915 	/*
190*e57b9183Scg149915 	 * Assumes a single contiguous range.  Needs a special filp in
191*e57b9183Scg149915 	 * 'heap' to stop it being subsumed.
192*e57b9183Scg149915 	 */
193*e57b9183Scg149915 	for (p = heap->next; p != heap; p = p->next) {
194*e57b9183Scg149915 		while (p->filp == 0 && p->next->filp == 0) {
195*e57b9183Scg149915 			struct mem_block *q = p->next;
196*e57b9183Scg149915 			p->size += q->size;
197*e57b9183Scg149915 			p->next = q->next;
198*e57b9183Scg149915 			p->next->prev = p;
199*e57b9183Scg149915 			drm_free(q, sizeof (*q), DRM_MEM_DRIVER);
200*e57b9183Scg149915 		}
201*e57b9183Scg149915 	}
202*e57b9183Scg149915 }
203*e57b9183Scg149915 
204*e57b9183Scg149915 /*
205*e57b9183Scg149915  * Shutdown.
206*e57b9183Scg149915  */
207*e57b9183Scg149915 void
radeon_mem_takedown(struct mem_block ** heap)208*e57b9183Scg149915 radeon_mem_takedown(struct mem_block **heap)
209*e57b9183Scg149915 {
210*e57b9183Scg149915 	struct mem_block *p;
211*e57b9183Scg149915 
212*e57b9183Scg149915 	if (!*heap)
213*e57b9183Scg149915 		return;
214*e57b9183Scg149915 
215*e57b9183Scg149915 	for (p = (*heap)->next; p != *heap; ) {
216*e57b9183Scg149915 		struct mem_block *q = p;
217*e57b9183Scg149915 		p = p->next;
218*e57b9183Scg149915 		drm_free(q, sizeof (*q), DRM_MEM_DRIVER);
219*e57b9183Scg149915 	}
220*e57b9183Scg149915 
221*e57b9183Scg149915 	drm_free(*heap, sizeof (**heap), DRM_MEM_DRIVER);
222*e57b9183Scg149915 	*heap = NULL;
223*e57b9183Scg149915 }
224*e57b9183Scg149915 
225*e57b9183Scg149915 /* IOCTL HANDLERS */
226*e57b9183Scg149915 
227*e57b9183Scg149915 static struct mem_block **
get_heap(drm_radeon_private_t * dev_priv,int region)228*e57b9183Scg149915 get_heap(drm_radeon_private_t *dev_priv, int region)
229*e57b9183Scg149915 {
230*e57b9183Scg149915 	switch (region) {
231*e57b9183Scg149915 	case RADEON_MEM_REGION_GART:
232*e57b9183Scg149915 		return (&dev_priv->gart_heap);
233*e57b9183Scg149915 	case RADEON_MEM_REGION_FB:
234*e57b9183Scg149915 		return (&dev_priv->fb_heap);
235*e57b9183Scg149915 	default:
236*e57b9183Scg149915 		return (NULL);
237*e57b9183Scg149915 	}
238*e57b9183Scg149915 }
239*e57b9183Scg149915 
240*e57b9183Scg149915 /*ARGSUSED*/
241*e57b9183Scg149915 int
radeon_mem_alloc(DRM_IOCTL_ARGS)242*e57b9183Scg149915 radeon_mem_alloc(DRM_IOCTL_ARGS)
243*e57b9183Scg149915 {
244*e57b9183Scg149915 	DRM_DEVICE;
245*e57b9183Scg149915 	drm_radeon_private_t *dev_priv = dev->dev_private;
246*e57b9183Scg149915 	drm_radeon_mem_alloc_t alloc;
247*e57b9183Scg149915 	struct mem_block *block, **heap;
248*e57b9183Scg149915 
249*e57b9183Scg149915 	if (!dev_priv) {
250*e57b9183Scg149915 		DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
251*e57b9183Scg149915 		return (EINVAL);
252*e57b9183Scg149915 	}
253*e57b9183Scg149915 
254*e57b9183Scg149915 #ifdef _MULTI_DATAMODEL
255*e57b9183Scg149915 	if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) {
256*e57b9183Scg149915 		drm_radeon_mem_alloc_32_t alloc32;
257*e57b9183Scg149915 
258*e57b9183Scg149915 		DRM_COPYFROM_WITH_RETURN(&alloc32, (void *) data,
259*e57b9183Scg149915 		    sizeof (alloc32));
260*e57b9183Scg149915 		alloc.region = alloc32.region;
261*e57b9183Scg149915 		alloc.alignment = alloc32.alignment;
262*e57b9183Scg149915 		alloc.size = alloc32.size;
263*e57b9183Scg149915 		alloc.region_offset = (void *)(uintptr_t)alloc32.region_offset;
264*e57b9183Scg149915 	} else {
265*e57b9183Scg149915 #endif
266*e57b9183Scg149915 		DRM_COPYFROM_WITH_RETURN(&alloc, (void *) data, sizeof (alloc));
267*e57b9183Scg149915 #ifdef _MULTI_DATAMODEL
268*e57b9183Scg149915 	}
269*e57b9183Scg149915 #endif
270*e57b9183Scg149915 
271*e57b9183Scg149915 	heap = get_heap(dev_priv, alloc.region);
272*e57b9183Scg149915 	if (!heap || !*heap)
273*e57b9183Scg149915 		return (EFAULT);
274*e57b9183Scg149915 
275*e57b9183Scg149915 	/*
276*e57b9183Scg149915 	 * Make things easier on ourselves: all allocations at least
277*e57b9183Scg149915 	 * 4k aligned.
278*e57b9183Scg149915 	 */
279*e57b9183Scg149915 	if (alloc.alignment < 12)
280*e57b9183Scg149915 		alloc.alignment = 12;
281*e57b9183Scg149915 
282*e57b9183Scg149915 	block = alloc_block(*heap, alloc.size, alloc.alignment, fpriv);
283*e57b9183Scg149915 
284*e57b9183Scg149915 	if (!block)
285*e57b9183Scg149915 		return (ENOMEM);
286*e57b9183Scg149915 
287*e57b9183Scg149915 	if (DRM_COPY_TO_USER(alloc.region_offset, &block->start,
288*e57b9183Scg149915 	    sizeof (int))) {
289*e57b9183Scg149915 		DRM_ERROR("copy_to_user\n");
290*e57b9183Scg149915 		return (EFAULT);
291*e57b9183Scg149915 	}
292*e57b9183Scg149915 
293*e57b9183Scg149915 	return (0);
294*e57b9183Scg149915 }
295*e57b9183Scg149915 
296*e57b9183Scg149915 /*ARGSUSED*/
297*e57b9183Scg149915 int
radeon_mem_free(DRM_IOCTL_ARGS)298*e57b9183Scg149915 radeon_mem_free(DRM_IOCTL_ARGS)
299*e57b9183Scg149915 {
300*e57b9183Scg149915 	DRM_DEVICE;
301*e57b9183Scg149915 	drm_radeon_private_t *dev_priv = dev->dev_private;
302*e57b9183Scg149915 	drm_radeon_mem_free_t memfree;
303*e57b9183Scg149915 	struct mem_block *block, **heap;
304*e57b9183Scg149915 
305*e57b9183Scg149915 	if (!dev_priv) {
306*e57b9183Scg149915 		DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
307*e57b9183Scg149915 		return (EINVAL);
308*e57b9183Scg149915 	}
309*e57b9183Scg149915 
310*e57b9183Scg149915 	DRM_COPYFROM_WITH_RETURN(&memfree, (void *) data, sizeof (memfree));
311*e57b9183Scg149915 
312*e57b9183Scg149915 	heap = get_heap(dev_priv, memfree.region);
313*e57b9183Scg149915 	if (!heap || !*heap)
314*e57b9183Scg149915 		return (EFAULT);
315*e57b9183Scg149915 
316*e57b9183Scg149915 	block = find_block(*heap, memfree.region_offset);
317*e57b9183Scg149915 	if (!block)
318*e57b9183Scg149915 		return (EFAULT);
319*e57b9183Scg149915 
320*e57b9183Scg149915 	if (block->filp != fpriv)
321*e57b9183Scg149915 		return (EPERM);
322*e57b9183Scg149915 
323*e57b9183Scg149915 	free_block(block);
324*e57b9183Scg149915 	return (0);
325*e57b9183Scg149915 }
326*e57b9183Scg149915 
327*e57b9183Scg149915 /*ARGSUSED*/
328*e57b9183Scg149915 int
radeon_mem_init_heap(DRM_IOCTL_ARGS)329*e57b9183Scg149915 radeon_mem_init_heap(DRM_IOCTL_ARGS)
330*e57b9183Scg149915 {
331*e57b9183Scg149915 	DRM_DEVICE;
332*e57b9183Scg149915 	drm_radeon_private_t *dev_priv = dev->dev_private;
333*e57b9183Scg149915 	drm_radeon_mem_init_heap_t initheap;
334*e57b9183Scg149915 	struct mem_block **heap;
335*e57b9183Scg149915 
336*e57b9183Scg149915 	if (!dev_priv) {
337*e57b9183Scg149915 		DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
338*e57b9183Scg149915 		return (EINVAL);
339*e57b9183Scg149915 	}
340*e57b9183Scg149915 
341*e57b9183Scg149915 	DRM_COPYFROM_WITH_RETURN(&initheap, (void *) data, sizeof (initheap));
342*e57b9183Scg149915 
343*e57b9183Scg149915 	heap = get_heap(dev_priv, initheap.region);
344*e57b9183Scg149915 	if (!heap)
345*e57b9183Scg149915 		return (EFAULT);
346*e57b9183Scg149915 
347*e57b9183Scg149915 	if (*heap) {
348*e57b9183Scg149915 		DRM_ERROR("heap already initialized?");
349*e57b9183Scg149915 		return (EFAULT);
350*e57b9183Scg149915 	}
351*e57b9183Scg149915 
352*e57b9183Scg149915 	return (init_heap(heap, initheap.start, initheap.size));
353*e57b9183Scg149915 }
354