xref: /titanic_44/usr/src/uts/sun4/io/px/px_fdvma.c (revision 8eea8e29cc4374d1ee24c25a07f45af132db3499)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 /*
23  * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 /*
30  * Internal PCI Fast DVMA implementation
31  */
32 #include <sys/types.h>
33 #include <sys/kmem.h>
34 #include <sys/async.h>
35 #include <sys/sysmacros.h>
36 #include <sys/sunddi.h>
37 #include <sys/ddi_impldefs.h>
38 #include <sys/dvma.h>
39 #include "px_obj.h"
40 
41 /*LINTLIBRARY*/
42 
43 static struct dvma_ops fdvma_ops;
44 typedef struct fast_dvma fdvma_t;
45 
46 /*
47  * The following routines are used to implement the sun4u fast dvma
48  * routines on this bus.
49  */
50 
51 /*ARGSUSED*/
52 static void
53 px_fdvma_load(ddi_dma_handle_t h, caddr_t a, uint_t len, uint_t index,
54 	ddi_dma_cookie_t *cp)
55 {
56 	ddi_dma_impl_t *mp = (ddi_dma_impl_t *)h;
57 	fdvma_t *fdvma_p = (fdvma_t *)mp->dmai_fdvma;
58 	px_t *px_p = (px_t *)fdvma_p->softsp;
59 	px_mmu_t *mmu_p = px_p->px_mmu_p;
60 	dev_info_t *dip = px_p->px_dip;
61 	px_dvma_addr_t dvma_addr, dvma_pg;
62 	uint32_t offset;
63 	size_t npages, pg_index;
64 	uint64_t attr;
65 
66 	offset = (uint32_t)a & MMU_PAGE_OFFSET;
67 	npages = MMU_BTOPR(len + offset);
68 	if (!npages)
69 		return;
70 
71 	/* make sure we don't exceed reserved boundary */
72 	DBG(DBG_FAST_DVMA, dip, "load index=%x: %p+%x ", index, a, len);
73 	if (index + npages > mp->dmai_ndvmapages) {
74 		cmn_err(px_panic_on_fatal_errors ? CE_PANIC : CE_WARN,
75 			"%s%d: kaddr_load index(%x)+pgs(%x) exceeds limit\n",
76 			ddi_driver_name(dip), ddi_get_instance(dip),
77 			index, npages);
78 		return;
79 	}
80 	fdvma_p->pagecnt[index] = npages;
81 
82 	dvma_addr = mp->dmai_mapping + MMU_PTOB(index);
83 	dvma_pg = MMU_BTOP(dvma_addr);
84 	pg_index = dvma_pg - mmu_p->dvma_base_pg;
85 
86 	/* construct the dma cookie to be returned */
87 	MAKE_DMA_COOKIE(cp, dvma_addr | offset, len);
88 	DBG(DBG_FAST_DVMA | DBG_CONT, dip, "cookie: %x+%x\n",
89 		cp->dmac_address, cp->dmac_size);
90 
91 	attr = PX_GET_TTE_ATTR(mp->dmai_rflags);
92 
93 	if (px_lib_iommu_map(dip, PCI_TSBID(0, pg_index), npages, attr,
94 	    (void *)a, 0, MMU_MAP_BUF) != DDI_SUCCESS) {
95 		cmn_err(CE_WARN, "%s%d: kaddr_load can't get "
96 		    "page frame for vaddr %x", ddi_driver_name(dip),
97 		    ddi_get_instance(dip), (int)a);
98 	}
99 }
100 
101 /*ARGSUSED*/
102 static void
103 px_fdvma_unload(ddi_dma_handle_t h, uint_t index, uint_t sync_flag)
104 {
105 	ddi_dma_impl_t *mp = (ddi_dma_impl_t *)h;
106 	fdvma_t *fdvma_p = (fdvma_t *)mp->dmai_fdvma;
107 	px_t *px_p = (px_t *)fdvma_p->softsp;
108 	size_t npages = fdvma_p->pagecnt[index];
109 	px_dvma_addr_t dvma_pg = MMU_BTOP(mp->dmai_mapping + MMU_PTOB(index));
110 
111 	DBG(DBG_FAST_DVMA, px_p->px_dip,
112 		"unload index=%x sync_flag=%x %x+%x+%x\n", index, sync_flag,
113 		mp->dmai_mapping, MMU_PTOB(index), MMU_PTOB(npages));
114 
115 	px_mmu_unmap_pages(px_p->px_mmu_p, dvma_pg, npages);
116 	fdvma_p->pagecnt[index] = 0;
117 }
118 
119 /*ARGSUSED*/
120 static void
121 px_fdvma_sync(ddi_dma_handle_t h, uint_t index, uint_t sync_flag)
122 {
123 	ddi_dma_impl_t *mp = (ddi_dma_impl_t *)h;
124 	fdvma_t *fdvma_p = (fdvma_t *)mp->dmai_fdvma;
125 	px_t *px_p = (px_t *)fdvma_p->softsp;
126 	size_t npg = fdvma_p->pagecnt[index];
127 
128 	DBG(DBG_FAST_DVMA, px_p->px_dip,
129 		"sync index=%x sync_flag=%x %x+%x+%x\n", index, sync_flag,
130 		mp->dmai_mapping, MMU_PTOB(index), MMU_PTOB(npg));
131 }
132 
133 int
134 px_fdvma_reserve(dev_info_t *dip, dev_info_t *rdip, px_t *px_p,
135 	ddi_dma_req_t *dmareq, ddi_dma_handle_t *handlep)
136 {
137 	fdvma_t *fdvma_p;
138 	px_dvma_addr_t dvma_pg;
139 	px_mmu_t *mmu_p = px_p->px_mmu_p;
140 	size_t npages;
141 	ddi_dma_impl_t *mp;
142 	ddi_dma_lim_t *lim_p = dmareq->dmar_limits;
143 	ulong_t hi = lim_p->dlim_addr_hi;
144 	ulong_t lo = lim_p->dlim_addr_lo;
145 	size_t counter_max = (lim_p->dlim_cntr_max + 1) & MMU_PAGE_MASK;
146 
147 	if (px_disable_fdvma)
148 		return (DDI_FAILURE);
149 
150 	DBG(DBG_DMA_CTL, dip, "DDI_DMA_RESERVE: rdip=%s%d\n",
151 		ddi_driver_name(rdip), ddi_get_instance(rdip));
152 
153 	/*
154 	 * Check the limit structure.
155 	 */
156 	if ((lo >= hi) || (hi < mmu_p->mmu_dvma_base))
157 		return (DDI_DMA_BADLIMITS);
158 
159 	/*
160 	 * Check the size of the request.
161 	 */
162 	npages = dmareq->dmar_object.dmao_size;
163 	if (npages > mmu_p->mmu_dvma_reserve)
164 		return (DDI_DMA_NORESOURCES);
165 
166 	/*
167 	 * Allocate the dma handle.
168 	 */
169 	mp = kmem_zalloc(sizeof (px_dma_hdl_t), KM_SLEEP);
170 
171 	/*
172 	 * Get entries from dvma space map.
173 	 * (vmem_t *vmp,
174 	 *	size_t size, size_t align, size_t phase,
175 	 *	size_t nocross, void *minaddr, void *maxaddr, int vmflag)
176 	 */
177 	dvma_pg = MMU_BTOP((ulong_t)vmem_xalloc(mmu_p->mmu_dvma_map,
178 		MMU_PTOB(npages), MMU_PAGE_SIZE, 0,
179 		counter_max, (void *)lo, (void *)(hi + 1),
180 		dmareq->dmar_fp == DDI_DMA_SLEEP ? VM_SLEEP : VM_NOSLEEP));
181 	if (dvma_pg == 0) {
182 		kmem_free(mp, sizeof (px_dma_hdl_t));
183 		return (DDI_DMA_NOMAPPING);
184 	}
185 	mmu_p->mmu_dvma_reserve -= npages;
186 
187 	/*
188 	 * Create the fast dvma request structure.
189 	 */
190 	fdvma_p = kmem_alloc(sizeof (fdvma_t), KM_SLEEP);
191 	fdvma_p->pagecnt = kmem_alloc(npages * sizeof (uint_t), KM_SLEEP);
192 	fdvma_p->ops = &fdvma_ops;
193 	fdvma_p->softsp = (caddr_t)px_p;
194 	fdvma_p->sync_flag = NULL;
195 
196 	/*
197 	 * Initialize the handle.
198 	 */
199 	mp->dmai_rdip = rdip;
200 	mp->dmai_rflags = DMP_BYPASSNEXUS | DDI_DMA_READ | DMP_NOSYNC;
201 	mp->dmai_minxfer = dmareq->dmar_limits->dlim_minxfer;
202 	mp->dmai_burstsizes = dmareq->dmar_limits->dlim_burstsizes;
203 	mp->dmai_mapping = MMU_PTOB(dvma_pg);
204 	mp->dmai_ndvmapages = npages;
205 	mp->dmai_size = npages * MMU_PAGE_SIZE;
206 	mp->dmai_nwin = 0;
207 	mp->dmai_fdvma = (caddr_t)fdvma_p;
208 	DBG(DBG_DMA_CTL, dip,
209 		"DDI_DMA_RESERVE: mp=%p dvma=%x npages=%x private=%p\n",
210 		mp, mp->dmai_mapping, npages, fdvma_p);
211 	*handlep = (ddi_dma_handle_t)mp;
212 	return (DDI_SUCCESS);
213 }
214 
215 int
216 px_fdvma_release(dev_info_t *dip, px_t *px_p, ddi_dma_impl_t *mp)
217 {
218 	px_mmu_t *mmu_p = px_p->px_mmu_p;
219 	size_t npages;
220 	fdvma_t *fdvma_p = (fdvma_t *)mp->dmai_fdvma;
221 
222 	if (px_disable_fdvma)
223 		return (DDI_FAILURE);
224 
225 	/* validate fdvma handle */
226 	if (!(mp->dmai_rflags & DMP_BYPASSNEXUS)) {
227 		DBG(DBG_DMA_CTL, dip, "DDI_DMA_RELEASE: not fast dma\n");
228 		return (DDI_FAILURE);
229 	}
230 
231 	/* flush all reserved dvma addresses from mmu */
232 	px_mmu_unmap_window(mmu_p, mp);
233 
234 	npages = mp->dmai_ndvmapages;
235 	vmem_xfree(mmu_p->mmu_dvma_map, (void *)mp->dmai_mapping,
236 		MMU_PTOB(npages));
237 
238 	mmu_p->mmu_dvma_reserve += npages;
239 	mp->dmai_ndvmapages = 0;
240 
241 	/* see if there is anyone waiting for dvma space */
242 	if (mmu_p->mmu_dvma_clid != 0) {
243 		DBG(DBG_DMA_CTL, dip, "run dvma callback\n");
244 		ddi_run_callback(&mmu_p->mmu_dvma_clid);
245 	}
246 
247 	/* free data structures */
248 	kmem_free(fdvma_p->pagecnt, npages * sizeof (uint_t));
249 	kmem_free(fdvma_p, sizeof (fdvma_t));
250 	kmem_free(mp, sizeof (px_dma_hdl_t));
251 
252 	/* see if there is anyone waiting for kmem */
253 	if (px_kmem_clid != 0) {
254 		DBG(DBG_DMA_CTL, dip, "run handle callback\n");
255 		ddi_run_callback(&px_kmem_clid);
256 	}
257 	return (DDI_SUCCESS);
258 }
259 
260 static struct dvma_ops fdvma_ops = {
261 	DVMAO_REV,
262 	px_fdvma_load,
263 	px_fdvma_unload,
264 	px_fdvma_sync
265 };
266