xref: /titanic_52/usr/src/uts/common/vm/seg.h (revision d91236fe104c7ea63142e053b22a39c8a30d304b)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 /*	Copyright (c) 1984, 1986, 1987, 1988, 1989 AT&T	*/
27 /*	  All Rights Reserved  	*/
28 
29 /*
30  * University Copyright- Copyright (c) 1982, 1986, 1988
31  * The Regents of the University of California
32  * All Rights Reserved
33  *
34  * University Acknowledgment- Portions of this document are derived from
35  * software developed by the University of California, Berkeley, and its
36  * contributors.
37  */
38 
39 #ifndef	_VM_SEG_H
40 #define	_VM_SEG_H
41 
42 #pragma ident	"%Z%%M%	%I%	%E% SMI"
43 
44 #include <sys/vnode.h>
45 #include <sys/avl.h>
46 #include <vm/seg_enum.h>
47 #include <vm/faultcode.h>
48 #include <vm/hat.h>
49 
50 #ifdef	__cplusplus
51 extern "C" {
52 #endif
53 
54 /*
55  * VM - Segments.
56  */
57 
58 struct anon_map;
59 
60 /*
61  * kstat statistics for segment advise
62  */
63 typedef struct {
64 	kstat_named_t MADV_FREE_hit;
65 	kstat_named_t MADV_FREE_miss;
66 } segadvstat_t;
67 
68 /*
69  * memory object ids
70  */
71 typedef struct memid { u_longlong_t val[2]; } memid_t;
72 
73 /*
74  * An address space contains a set of segments, managed by drivers.
75  * Drivers support mapped devices, sharing, copy-on-write, etc.
76  *
77  * The seg structure contains a lock to prevent races, the base virtual
78  * address and size of the segment, a back pointer to the containing
79  * address space, pointers to maintain an AVL tree of segments in the
80  * same address space, and procedure and data hooks for the driver.
81  * The AVL tree of segments for the address space is sorted by
82  * ascending base addresses and overlapping segments are not allowed.
83  *
84  * After a segment is created, faults may occur on pages of the segment.
85  * When a fault occurs, the fault handling code must get the desired
86  * object and set up the hardware translation to the object.  For some
87  * objects, the fault handling code also implements copy-on-write.
88  *
89  * When the hat wants to unload a translation, it can call the unload
90  * routine which is responsible for processing reference and modify bits.
91  *
92  * Each segment is protected by it's containing address space lock.  To
93  * access any field in the segment structure, the "as" must be locked.
94  * If a segment field is to be modified, the address space lock must be
95  * write locked.
96  */
97 
98 typedef struct pcache_link {
99 	struct pcache_link	*p_lnext;
100 	struct pcache_link	*p_lprev;
101 } pcache_link_t;
102 
103 typedef struct seg {
104 	caddr_t	s_base;			/* base virtual address */
105 	size_t	s_size;			/* size in bytes */
106 	uint_t	s_szc;			/* max page size code */
107 	uint_t	s_flags;		/* flags for segment, see below */
108 	struct	as *s_as;		/* containing address space */
109 	avl_node_t s_tree;		/* AVL tree links to segs in this as */
110 	struct	seg_ops *s_ops;		/* ops vector: see below */
111 	void *s_data;			/* private data for instance */
112 	kmutex_t s_pmtx;		/* protects seg's pcache list */
113 	pcache_link_t s_phead;		/* head of seg's pcache list */
114 } seg_t;
115 
116 #define	S_PURGE		(0x01)		/* seg should be purged in as_gap() */
117 
118 struct	seg_ops {
119 	int	(*dup)(struct seg *, struct seg *);
120 	int	(*unmap)(struct seg *, caddr_t, size_t);
121 	void	(*free)(struct seg *);
122 	faultcode_t (*fault)(struct hat *, struct seg *, caddr_t, size_t,
123 	    enum fault_type, enum seg_rw);
124 	faultcode_t (*faulta)(struct seg *, caddr_t);
125 	int	(*setprot)(struct seg *, caddr_t, size_t, uint_t);
126 	int	(*checkprot)(struct seg *, caddr_t, size_t, uint_t);
127 	int	(*kluster)(struct seg *, caddr_t, ssize_t);
128 	size_t	(*swapout)(struct seg *);
129 	int	(*sync)(struct seg *, caddr_t, size_t, int, uint_t);
130 	size_t	(*incore)(struct seg *, caddr_t, size_t, char *);
131 	int	(*lockop)(struct seg *, caddr_t, size_t, int, int, ulong_t *,
132 			size_t);
133 	int	(*getprot)(struct seg *, caddr_t, size_t, uint_t *);
134 	u_offset_t	(*getoffset)(struct seg *, caddr_t);
135 	int	(*gettype)(struct seg *, caddr_t);
136 	int	(*getvp)(struct seg *, caddr_t, struct vnode **);
137 	int	(*advise)(struct seg *, caddr_t, size_t, uint_t);
138 	void	(*dump)(struct seg *);
139 	int	(*pagelock)(struct seg *, caddr_t, size_t, struct page ***,
140 			enum lock_type, enum seg_rw);
141 	int	(*setpagesize)(struct seg *, caddr_t, size_t, uint_t);
142 	int	(*getmemid)(struct seg *, caddr_t, memid_t *);
143 	struct lgrp_mem_policy_info	*(*getpolicy)(struct seg *, caddr_t);
144 	int	(*capable)(struct seg *, segcapability_t);
145 };
146 
147 #ifdef _KERNEL
148 
149 /*
150  * Generic segment operations
151  */
152 extern	void	seg_init(void);
153 extern	struct	seg *seg_alloc(struct as *as, caddr_t base, size_t size);
154 extern	int	seg_attach(struct as *as, caddr_t base, size_t size,
155 			struct seg *seg);
156 extern	void	seg_unmap(struct seg *seg);
157 extern	void	seg_free(struct seg *seg);
158 
159 /*
160  * functions for pagelock cache support
161  */
162 typedef	int (*seg_preclaim_cbfunc_t)(void *, caddr_t, size_t,
163     struct page **, enum seg_rw, int);
164 
165 extern	struct	page **seg_plookup(struct seg *seg, struct anon_map *amp,
166     caddr_t addr, size_t len, enum seg_rw rw, uint_t flags);
167 extern	void	seg_pinactive(struct seg *seg, struct anon_map *amp,
168     caddr_t addr, size_t len, struct page **pp, enum seg_rw rw,
169     uint_t flags, seg_preclaim_cbfunc_t callback);
170 
171 extern	void	seg_ppurge(struct seg *seg, struct anon_map *amp,
172     uint_t flags);
173 extern	void	seg_ppurge_wiredpp(struct page **pp);
174 
175 extern	int	seg_pinsert_check(struct seg *seg, struct anon_map *amp,
176     caddr_t addr, size_t len, uint_t flags);
177 extern	int	seg_pinsert(struct seg *seg, struct anon_map *amp,
178     caddr_t addr, size_t len, size_t wlen, struct page **pp, enum seg_rw rw,
179     uint_t flags, seg_preclaim_cbfunc_t callback);
180 
181 extern	void	seg_pasync_thread(void);
182 extern	void	seg_preap(void);
183 extern	int	seg_p_disable(void);
184 extern	void	seg_p_enable(void);
185 
186 extern	segadvstat_t	segadvstat;
187 
188 /*
189  * Flags for pagelock cache support.
190  * Flags argument is passed as uint_t to pcache routines.  upper 16 bits of
191  * the flags argument are reserved for alignment page shift when SEGP_PSHIFT
192  * is set.
193  */
194 #define	SEGP_FORCE_WIRED	0x1	/* skip check against seg_pwindow */
195 #define	SEGP_AMP		0x2	/* anon map's pcache entry */
196 #define	SEGP_PSHIFT		0x4	/* addr pgsz shift for hash function */
197 
198 /*
199  * Return values for seg_pinsert and seg_pinsert_check functions.
200  */
201 #define	SEGP_SUCCESS		0	/* seg_pinsert() succeeded */
202 #define	SEGP_FAIL		1	/* seg_pinsert() failed */
203 
204 /* Page status bits for segop_incore */
205 #define	SEG_PAGE_INCORE		0x01	/* VA has a page backing it */
206 #define	SEG_PAGE_LOCKED		0x02	/* VA has a page that is locked */
207 #define	SEG_PAGE_HASCOW		0x04	/* VA has a page with a copy-on-write */
208 #define	SEG_PAGE_SOFTLOCK	0x08	/* VA has a page with softlock held */
209 #define	SEG_PAGE_VNODEBACKED	0x10	/* Segment is backed by a vnode */
210 #define	SEG_PAGE_ANON		0x20	/* VA has an anonymous page */
211 #define	SEG_PAGE_VNODE		0x40	/* VA has a vnode page backing it */
212 
213 #define	SEGOP_DUP(s, n)		    (*(s)->s_ops->dup)((s), (n))
214 #define	SEGOP_UNMAP(s, a, l)	    (*(s)->s_ops->unmap)((s), (a), (l))
215 #define	SEGOP_FREE(s)		    (*(s)->s_ops->free)((s))
216 #define	SEGOP_FAULT(h, s, a, l, t, rw) \
217 		(*(s)->s_ops->fault)((h), (s), (a), (l), (t), (rw))
218 #define	SEGOP_FAULTA(s, a)	    (*(s)->s_ops->faulta)((s), (a))
219 #define	SEGOP_SETPROT(s, a, l, p)   (*(s)->s_ops->setprot)((s), (a), (l), (p))
220 #define	SEGOP_CHECKPROT(s, a, l, p) (*(s)->s_ops->checkprot)((s), (a), (l), (p))
221 #define	SEGOP_KLUSTER(s, a, d)	    (*(s)->s_ops->kluster)((s), (a), (d))
222 #define	SEGOP_SWAPOUT(s)	    (*(s)->s_ops->swapout)((s))
223 #define	SEGOP_SYNC(s, a, l, atr, f) \
224 		(*(s)->s_ops->sync)((s), (a), (l), (atr), (f))
225 #define	SEGOP_INCORE(s, a, l, v)    (*(s)->s_ops->incore)((s), (a), (l), (v))
226 #define	SEGOP_LOCKOP(s, a, l, atr, op, b, p) \
227 		(*(s)->s_ops->lockop)((s), (a), (l), (atr), (op), (b), (p))
228 #define	SEGOP_GETPROT(s, a, l, p)   (*(s)->s_ops->getprot)((s), (a), (l), (p))
229 #define	SEGOP_GETOFFSET(s, a)	    (*(s)->s_ops->getoffset)((s), (a))
230 #define	SEGOP_GETTYPE(s, a)	    (*(s)->s_ops->gettype)((s), (a))
231 #define	SEGOP_GETVP(s, a, vpp)	    (*(s)->s_ops->getvp)((s), (a), (vpp))
232 #define	SEGOP_ADVISE(s, a, l, b)    (*(s)->s_ops->advise)((s), (a), (l), (b))
233 #define	SEGOP_DUMP(s)		    (*(s)->s_ops->dump)((s))
234 #define	SEGOP_PAGELOCK(s, a, l, p, t, rw) \
235 		(*(s)->s_ops->pagelock)((s), (a), (l), (p), (t), (rw))
236 #define	SEGOP_SETPAGESIZE(s, a, l, szc) \
237 		(*(s)->s_ops->setpagesize)((s), (a), (l), (szc))
238 #define	SEGOP_GETMEMID(s, a, mp)    (*(s)->s_ops->getmemid)((s), (a), (mp))
239 #define	SEGOP_GETPOLICY(s, a)	    (*(s)->s_ops->getpolicy)((s), (a))
240 #define	SEGOP_CAPABLE(s, c)	    (*(s)->s_ops->capable)((s), (c))
241 
242 #define	seg_page(seg, addr) \
243 	(((uintptr_t)((addr) - (seg)->s_base)) >> PAGESHIFT)
244 
245 #define	seg_pages(seg) \
246 	(((uintptr_t)((seg)->s_size + PAGEOFFSET)) >> PAGESHIFT)
247 
248 #define	IE_NOMEM	-1	/* internal to seg layer */
249 #define	IE_RETRY	-2	/* internal to seg layer */
250 #define	IE_REATTACH	-3	/* internal to seg layer */
251 
252 /* Delay/retry factors for seg_p_mem_config_pre_del */
253 #define	SEGP_PREDEL_DELAY_FACTOR	4
254 /*
255  * As a workaround to being unable to purge the pagelock
256  * cache during a DR delete memory operation, we use
257  * a stall threshold that is twice the maximum seen
258  * during testing.  This workaround will be removed
259  * when a suitable fix is found.
260  */
261 #define	SEGP_STALL_SECONDS	25
262 #define	SEGP_STALL_THRESHOLD \
263 	(SEGP_STALL_SECONDS * SEGP_PREDEL_DELAY_FACTOR)
264 
265 #ifdef VMDEBUG
266 
267 uint_t	seg_page(struct seg *, caddr_t);
268 uint_t	seg_pages(struct seg *);
269 
270 #endif	/* VMDEBUG */
271 
272 boolean_t	seg_can_change_zones(struct seg *);
273 size_t		seg_swresv(struct seg *);
274 
275 #endif	/* _KERNEL */
276 
277 #ifdef	__cplusplus
278 }
279 #endif
280 
281 #endif	/* _VM_SEG_H */
282