xref: /titanic_41/usr/src/uts/sfmmu/vm/xhat_sfmmu.c (revision 843e19887f64dde75055cf8842fc4db2171eff45)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 /*
23  * Copyright 2004 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 
30 #include <sys/types.h>
31 #include <sys/cmn_err.h>
32 #include <sys/mman.h>
33 #include <vm/hat_sfmmu.h>
34 #include <vm/xhat.h>
35 #include <vm/xhat_sfmmu.h>
36 #include <vm/page.h>
37 #include <vm/as.h>
38 
39 
40 
41 /*
42  * Allocates a block that includes both struct xhat and
43  * provider-specific data.
44  */
45 struct xhat_hme_blk *
46 xhat_alloc_xhatblk(struct xhat *xhat)
47 {
48 	struct xhat_hme_blk *xblk;
49 	xblk_cache_t	*xblkcache = xhat->xhat_provider->xblkcache;
50 
51 
52 
53 	mutex_enter(&xblkcache->lock);
54 	if (xblkcache->free_blks) {
55 		xblk = (struct xhat_hme_blk *)
56 		    sfmmu_hmetohblk(xblkcache->free_blks);
57 
58 		/*
59 		 * Since we are always walking the list in the
60 		 * forward direction, we don't update prev pointers
61 		 */
62 		xblkcache->free_blks = xblk->xblk_hme[0].hme_next;
63 		mutex_exit(&xblkcache->lock);
64 	} else {
65 		mutex_exit(&xblkcache->lock);
66 		xblk = kmem_cache_alloc(xblkcache->cache, KM_SLEEP);
67 	}
68 
69 	return (xblk);
70 }
71 
72 
73 /*
74  * Return the block to free_blks pool. The memory will
75  * be freed in the reclaim routine.
76  */
77 void
78 xhat_free_xhatblk(struct xhat_hme_blk *xblk)
79 {
80 	xblk_cache_t	*xblkcache = xblk->xhat_hme_blk_hat->
81 	    xhat_provider->xblkcache;
82 
83 
84 	mutex_enter(&xblkcache->lock);
85 	xblk->xblk_hme[0].hme_next = xblkcache->free_blks;
86 	xblkcache->free_blks = &xblk->xblk_hme[0];
87 	mutex_exit(&xblkcache->lock);
88 }
89 
90 
91 /*
92  * Ran by kmem reaper thread. Also called when
93  * provider unregisters
94  */
95 void
96 xhat_xblkcache_reclaim(void *arg)
97 {
98 	xhat_provider_t *provider = (xhat_provider_t *)arg;
99 	struct sf_hment	*sfhme;
100 	struct xhat_hme_blk	*xblk;
101 	xblk_cache_t	*xblkcache;
102 
103 	if (provider == NULL)
104 		cmn_err(CE_PANIC, "xhat_xblkcache_reclaim() is passed NULL");
105 
106 	xblkcache = provider->xblkcache;
107 
108 
109 	while (xblkcache->free_blks != NULL) {
110 
111 		/*
112 		 * Put free blocks on a separate list
113 		 * and free free_blks pointer.
114 		 */
115 		mutex_enter(&xblkcache->lock);
116 		sfhme = xblkcache->free_blks;
117 		xblkcache->free_blks = NULL;
118 		mutex_exit(&xblkcache->lock);
119 
120 		while (sfhme != NULL) {
121 			xblk = (struct xhat_hme_blk *)sfmmu_hmetohblk(sfhme);
122 			ASSERT(xblk->xhat_hme_blk_misc.xhat_bit == 1);
123 			sfhme = sfhme->hme_next;
124 			kmem_cache_free(xblkcache->cache, xblk);
125 		}
126 	}
127 }
128 
129 
130 
131 
132 /*
133  * Insert the xhat block (or, more precisely, the sf_hment)
134  * into page's p_mapping list.
135  */
136 pfn_t
137 xhat_insert_xhatblk(page_t *pp, struct xhat *xhat, void **blk)
138 {
139 	kmutex_t *pml;
140 	pfn_t pfn;
141 	struct xhat_hme_blk *xblk;
142 
143 
144 
145 	xblk = xhat_alloc_xhatblk(xhat);
146 	if (xblk == NULL)
147 		return (0);
148 
149 	/* Add a "user" to the XHAT */
150 	xhat_hat_hold(xhat);
151 
152 	xblk->xhat_hme_blk_hat = xhat;
153 	xblk->xhat_hme_blk_misc.xhat_bit = 1;
154 
155 	pml = sfmmu_mlist_enter(pp);
156 
157 
158 	/* Insert at the head of p_mapping list */
159 	xblk->xblk_hme[0].hme_prev = NULL;
160 	xblk->xblk_hme[0].hme_next = pp->p_mapping;
161 	xblk->xblk_hme[0].hme_page = pp;
162 
163 	/* Only one tte per xhat_hme_blk, at least for now */
164 	xblk->xblk_hme[0].hme_tte.tte_hmenum = 0;
165 
166 	if (pp->p_mapping) {
167 		((struct sf_hment *)(pp->p_mapping))->hme_prev =
168 		    &(xblk->xblk_hme[0]);
169 		ASSERT(pp->p_share > 0);
170 	} else	{
171 		/* EMPTY */
172 		ASSERT(pp->p_share == 0);
173 	}
174 	pp->p_mapping = &(xblk->xblk_hme[0]);
175 
176 	/*
177 	 * Update number of mappings.
178 	 */
179 	pp->p_share++;
180 	pfn = pp->p_pagenum;
181 
182 	sfmmu_mlist_exit(pml);
183 
184 	*blk = XBLK2PROVBLK(xblk);
185 
186 	return (pfn);
187 }
188 
189 
190 /*
191  * mlist_locked indicates whether the mapping list
192  * is locked. If provider did not lock it himself, the
193  * only time it is locked in HAT layer is in
194  * hat_pageunload().
195  */
196 int
197 xhat_delete_xhatblk(void *blk, int mlist_locked)
198 {
199 	struct xhat_hme_blk *xblk = PROVBLK2XBLK(blk);
200 	page_t *pp = xblk->xblk_hme[0].hme_page;
201 	kmutex_t *pml;
202 
203 
204 	ASSERT(pp != NULL);
205 	ASSERT(pp->p_share > 0);
206 
207 	if (!mlist_locked)
208 		pml = sfmmu_mlist_enter(pp);
209 	else
210 		ASSERT(sfmmu_mlist_held(pp));
211 
212 	pp->p_share--;
213 
214 	if (xblk->xblk_hme[0].hme_prev) {
215 		ASSERT(pp->p_mapping != &(xblk->xblk_hme[0]));
216 		ASSERT(xblk->xblk_hme[0].hme_prev->hme_page == pp);
217 		xblk->xblk_hme[0].hme_prev->hme_next =
218 		    xblk->xblk_hme[0].hme_next;
219 	} else {
220 		ASSERT(pp->p_mapping == &(xblk->xblk_hme[0]));
221 		pp->p_mapping = xblk->xblk_hme[0].hme_next;
222 		ASSERT((pp->p_mapping == NULL) ?
223 			(pp->p_share == 0) : 1);
224 	}
225 
226 	if (xblk->xblk_hme->hme_next) {
227 		ASSERT(xblk->xblk_hme[0].hme_next->hme_page == pp);
228 		xblk->xblk_hme[0].hme_next->hme_prev =
229 		    xblk->xblk_hme[0].hme_prev;
230 	}
231 
232 	if (!mlist_locked)
233 		sfmmu_mlist_exit(pml);
234 
235 	xhat_hat_rele(xblk->xhat_hme_blk_hat);
236 	xhat_free_xhatblk(xblk);
237 
238 
239 	return (0);
240 }
241