xref: /illumos-gate/usr/src/uts/i86xpv/vm/seg_mf.c (revision 66582b606a8194f7f3ba5b3a3a6dca5b0d346361)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  * Copyright 2018 Joyent, Inc.
26  */
27 
28 /*
29  * Machine frame segment driver.  This segment driver allows dom0 processes to
30  * map pages of other domains or Xen (e.g. during save/restore).  ioctl()s on
31  * the privcmd driver provide the MFN values backing each mapping, and we map
32  * them into the process's address space at this time.  Demand-faulting is not
33  * supported by this driver due to the requirements upon some of the ioctl()s.
34  */
35 
36 
37 #include <sys/types.h>
38 #include <sys/systm.h>
39 #include <sys/vmsystm.h>
40 #include <sys/mman.h>
41 #include <sys/errno.h>
42 #include <sys/kmem.h>
43 #include <sys/cmn_err.h>
44 #include <sys/vnode.h>
45 #include <sys/conf.h>
46 #include <sys/debug.h>
47 #include <sys/lgrp.h>
48 #include <sys/hypervisor.h>
49 
50 #include <vm/page.h>
51 #include <vm/hat.h>
52 #include <vm/as.h>
53 #include <vm/seg.h>
54 
55 #include <vm/hat_pte.h>
56 #include <vm/hat_i86.h>
57 #include <vm/seg_mf.h>
58 
59 #include <sys/fs/snode.h>
60 
61 #define	VTOCVP(vp)	(VTOS(vp)->s_commonvp)
62 
63 typedef struct segmf_mfn_s {
64 	mfn_t		m_mfn;
65 } segmf_mfn_t;
66 
67 /* g_flags */
68 #define	SEGMF_GFLAGS_WR		0x1
69 #define	SEGMF_GFLAGS_MAPPED	0x2
70 typedef struct segmf_gref_s {
71 	uint64_t	g_ptep;
72 	grant_ref_t	g_gref;
73 	uint32_t	g_flags;
74 	grant_handle_t	g_handle;
75 } segmf_gref_t;
76 
77 typedef union segmf_mu_u {
78 	segmf_mfn_t	m;
79 	segmf_gref_t	g;
80 } segmf_mu_t;
81 
82 typedef enum {
83 	SEGMF_MAP_EMPTY = 0,
84 	SEGMF_MAP_MFN,
85 	SEGMF_MAP_GREF
86 } segmf_map_type_t;
87 
88 typedef struct segmf_map_s {
89 	segmf_map_type_t	t_type;
90 	segmf_mu_t		u;
91 } segmf_map_t;
92 
93 struct segmf_data {
94 	kmutex_t	lock;
95 	struct vnode	*vp;
96 	uchar_t		prot;
97 	uchar_t		maxprot;
98 	size_t		softlockcnt;
99 	domid_t		domid;
100 	segmf_map_t	*map;
101 };
102 
103 static struct seg_ops segmf_ops;
104 
105 static int segmf_fault_gref_range(struct seg *seg, caddr_t addr, size_t len);
106 
107 static struct segmf_data *
108 segmf_data_zalloc(struct seg *seg)
109 {
110 	struct segmf_data *data = kmem_zalloc(sizeof (*data), KM_SLEEP);
111 
112 	mutex_init(&data->lock, "segmf.lock", MUTEX_DEFAULT, NULL);
113 	seg->s_ops = &segmf_ops;
114 	seg->s_data = data;
115 	return (data);
116 }
117 
118 int
119 segmf_create(struct seg **segpp, void *args)
120 {
121 	struct seg *seg = *segpp;
122 	struct segmf_crargs *a = args;
123 	struct segmf_data *data;
124 	struct as *as = seg->s_as;
125 	pgcnt_t i, npages = seg_pages(seg);
126 	int error;
127 
128 	hat_map(as->a_hat, seg->s_base, seg->s_size, HAT_MAP);
129 
130 	data = segmf_data_zalloc(seg);
131 	data->vp = specfind(a->dev, VCHR);
132 	data->prot = a->prot;
133 	data->maxprot = a->maxprot;
134 
135 	data->map = kmem_alloc(npages * sizeof (segmf_map_t), KM_SLEEP);
136 	for (i = 0; i < npages; i++) {
137 		data->map[i].t_type = SEGMF_MAP_EMPTY;
138 	}
139 
140 	error = VOP_ADDMAP(VTOCVP(data->vp), 0, as, seg->s_base, seg->s_size,
141 	    data->prot, data->maxprot, MAP_SHARED, CRED(), NULL);
142 
143 	if (error != 0)
144 		hat_unload(as->a_hat,
145 		    seg->s_base, seg->s_size, HAT_UNLOAD_UNMAP);
146 	return (error);
147 }
148 
149 /*
150  * Duplicate a seg and return new segment in newseg.
151  */
152 static int
153 segmf_dup(struct seg *seg, struct seg *newseg)
154 {
155 	struct segmf_data *data = seg->s_data;
156 	struct segmf_data *ndata;
157 	pgcnt_t npages = seg_pages(newseg);
158 	size_t sz;
159 
160 	ndata = segmf_data_zalloc(newseg);
161 
162 	VN_HOLD(data->vp);
163 	ndata->vp = data->vp;
164 	ndata->prot = data->prot;
165 	ndata->maxprot = data->maxprot;
166 	ndata->domid = data->domid;
167 
168 	sz = npages * sizeof (segmf_map_t);
169 	ndata->map = kmem_alloc(sz, KM_SLEEP);
170 	bcopy(data->map, ndata->map, sz);
171 
172 	return (VOP_ADDMAP(VTOCVP(ndata->vp), 0, newseg->s_as,
173 	    newseg->s_base, newseg->s_size, ndata->prot, ndata->maxprot,
174 	    MAP_SHARED, CRED(), NULL));
175 }
176 
177 /*
178  * We only support unmapping the whole segment, and we automatically unlock
179  * what we previously soft-locked.
180  */
181 static int
182 segmf_unmap(struct seg *seg, caddr_t addr, size_t len)
183 {
184 	struct segmf_data *data = seg->s_data;
185 	offset_t off;
186 
187 	if (addr < seg->s_base || addr + len > seg->s_base + seg->s_size ||
188 	    (len & PAGEOFFSET) || ((uintptr_t)addr & PAGEOFFSET))
189 		panic("segmf_unmap");
190 
191 	if (addr != seg->s_base || len != seg->s_size)
192 		return (ENOTSUP);
193 
194 	hat_unload(seg->s_as->a_hat, addr, len,
195 	    HAT_UNLOAD_UNMAP | HAT_UNLOAD_UNLOCK);
196 
197 	off = (offset_t)seg_page(seg, addr);
198 
199 	ASSERT(data->vp != NULL);
200 
201 	(void) VOP_DELMAP(VTOCVP(data->vp), off, seg->s_as, addr, len,
202 	    data->prot, data->maxprot, MAP_SHARED, CRED(), NULL);
203 
204 	seg_free(seg);
205 	return (0);
206 }
207 
208 static void
209 segmf_free(struct seg *seg)
210 {
211 	struct segmf_data *data = seg->s_data;
212 	pgcnt_t npages = seg_pages(seg);
213 
214 	kmem_free(data->map, npages * sizeof (segmf_map_t));
215 	VN_RELE(data->vp);
216 	mutex_destroy(&data->lock);
217 	kmem_free(data, sizeof (*data));
218 }
219 
220 static int segmf_faultpage_debug = 0;
221 /*ARGSUSED*/
222 static int
223 segmf_faultpage(struct hat *hat, struct seg *seg, caddr_t addr,
224     enum fault_type type, uint_t prot)
225 {
226 	struct segmf_data *data = seg->s_data;
227 	uint_t hat_flags = HAT_LOAD_NOCONSIST;
228 	mfn_t mfn;
229 	x86pte_t pte;
230 	segmf_map_t *map;
231 	uint_t idx;
232 
233 
234 	idx = seg_page(seg, addr);
235 	map = &data->map[idx];
236 	ASSERT(map->t_type == SEGMF_MAP_MFN);
237 
238 	mfn = map->u.m.m_mfn;
239 
240 	if (type == F_SOFTLOCK) {
241 		mutex_enter(&freemem_lock);
242 		data->softlockcnt++;
243 		mutex_exit(&freemem_lock);
244 		hat_flags |= HAT_LOAD_LOCK;
245 	} else
246 		hat_flags |= HAT_LOAD;
247 
248 	if (segmf_faultpage_debug > 0) {
249 		uprintf("segmf_faultpage: addr %p domid %x mfn %lx prot %x\n",
250 		    (void *)addr, data->domid, mfn, prot);
251 		segmf_faultpage_debug--;
252 	}
253 
254 	/*
255 	 * Ask the HAT to load a throwaway mapping to page zero, then
256 	 * overwrite it with our foreign domain mapping. It gets removed
257 	 * later via hat_unload()
258 	 */
259 	hat_devload(hat, addr, MMU_PAGESIZE, (pfn_t)0,
260 	    PROT_READ | HAT_UNORDERED_OK, hat_flags);
261 
262 	pte = mmu_ptob((x86pte_t)mfn) | PT_VALID | PT_USER | PT_FOREIGN;
263 	if (prot & PROT_WRITE)
264 		pte |= PT_WRITABLE;
265 
266 	if (HYPERVISOR_update_va_mapping_otherdomain((uintptr_t)addr, pte,
267 	    UVMF_INVLPG | UVMF_ALL, data->domid) != 0) {
268 		hat_flags = HAT_UNLOAD_UNMAP;
269 
270 		if (type == F_SOFTLOCK) {
271 			hat_flags |= HAT_UNLOAD_UNLOCK;
272 			mutex_enter(&freemem_lock);
273 			data->softlockcnt--;
274 			mutex_exit(&freemem_lock);
275 		}
276 
277 		hat_unload(hat, addr, MMU_PAGESIZE, hat_flags);
278 		return (FC_MAKE_ERR(EFAULT));
279 	}
280 
281 	return (0);
282 }
283 
284 static int
285 seg_rw_to_prot(enum seg_rw rw)
286 {
287 	switch (rw) {
288 	case S_READ:
289 		return (PROT_READ);
290 	case S_WRITE:
291 		return (PROT_WRITE);
292 	case S_EXEC:
293 		return (PROT_EXEC);
294 	case S_OTHER:
295 	default:
296 		break;
297 	}
298 	return (PROT_READ | PROT_WRITE | PROT_EXEC);
299 }
300 
301 static void
302 segmf_softunlock(struct hat *hat, struct seg *seg, caddr_t addr, size_t len)
303 {
304 	struct segmf_data *data = seg->s_data;
305 
306 	hat_unlock(hat, addr, len);
307 
308 	mutex_enter(&freemem_lock);
309 	ASSERT(data->softlockcnt >= btopr(len));
310 	data->softlockcnt -= btopr(len);
311 	mutex_exit(&freemem_lock);
312 
313 	if (data->softlockcnt == 0) {
314 		struct as *as = seg->s_as;
315 
316 		if (AS_ISUNMAPWAIT(as)) {
317 			mutex_enter(&as->a_contents);
318 			if (AS_ISUNMAPWAIT(as)) {
319 				AS_CLRUNMAPWAIT(as);
320 				cv_broadcast(&as->a_cv);
321 			}
322 			mutex_exit(&as->a_contents);
323 		}
324 	}
325 }
326 
327 static int
328 segmf_fault_range(struct hat *hat, struct seg *seg, caddr_t addr, size_t len,
329     enum fault_type type, enum seg_rw rw)
330 {
331 	struct segmf_data *data = seg->s_data;
332 	int error = 0;
333 	caddr_t a;
334 
335 	if ((data->prot & seg_rw_to_prot(rw)) == 0)
336 		return (FC_PROT);
337 
338 	/* loop over the address range handling each fault */
339 
340 	for (a = addr; a < addr + len; a += PAGESIZE) {
341 		error = segmf_faultpage(hat, seg, a, type, data->prot);
342 		if (error != 0)
343 			break;
344 	}
345 
346 	if (error != 0 && type == F_SOFTLOCK) {
347 		size_t done = (size_t)(a - addr);
348 
349 		/*
350 		 * Undo what's been done so far.
351 		 */
352 		if (done > 0)
353 			segmf_softunlock(hat, seg, addr, done);
354 	}
355 
356 	return (error);
357 }
358 
359 /*
360  * We never demand-fault for seg_mf.
361  */
362 /*ARGSUSED*/
363 static int
364 segmf_fault(struct hat *hat, struct seg *seg, caddr_t addr, size_t len,
365     enum fault_type type, enum seg_rw rw)
366 {
367 	return (FC_MAKE_ERR(EFAULT));
368 }
369 
370 /*ARGSUSED*/
371 static int
372 segmf_faulta(struct seg *seg, caddr_t addr)
373 {
374 	return (0);
375 }
376 
377 /*ARGSUSED*/
378 static int
379 segmf_setprot(struct seg *seg, caddr_t addr, size_t len, uint_t prot)
380 {
381 	return (EINVAL);
382 }
383 
384 /*ARGSUSED*/
385 static int
386 segmf_checkprot(struct seg *seg, caddr_t addr, size_t len, uint_t prot)
387 {
388 	return (EINVAL);
389 }
390 
391 /*ARGSUSED*/
392 static int
393 segmf_kluster(struct seg *seg, caddr_t addr, ssize_t delta)
394 {
395 	return (-1);
396 }
397 
398 /*ARGSUSED*/
399 static int
400 segmf_sync(struct seg *seg, caddr_t addr, size_t len, int attr, uint_t flags)
401 {
402 	return (0);
403 }
404 
405 /*
406  * XXPV	Hmm.  Should we say that mf mapping are "in core?"
407  */
408 
409 /*ARGSUSED*/
410 static size_t
411 segmf_incore(struct seg *seg, caddr_t addr, size_t len, char *vec)
412 {
413 	size_t v;
414 
415 	for (v = 0, len = (len + PAGEOFFSET) & PAGEMASK; len;
416 	    len -= PAGESIZE, v += PAGESIZE)
417 		*vec++ = 1;
418 	return (v);
419 }
420 
421 /*ARGSUSED*/
422 static int
423 segmf_lockop(struct seg *seg, caddr_t addr,
424     size_t len, int attr, int op, ulong_t *lockmap, size_t pos)
425 {
426 	return (0);
427 }
428 
429 static int
430 segmf_getprot(struct seg *seg, caddr_t addr, size_t len, uint_t *protv)
431 {
432 	struct segmf_data *data = seg->s_data;
433 	pgcnt_t pgno = seg_page(seg, addr + len) - seg_page(seg, addr) + 1;
434 
435 	if (pgno != 0) {
436 		do
437 			protv[--pgno] = data->prot;
438 		while (pgno != 0)
439 			;
440 	}
441 	return (0);
442 }
443 
444 static u_offset_t
445 segmf_getoffset(struct seg *seg, caddr_t addr)
446 {
447 	return (addr - seg->s_base);
448 }
449 
450 /*ARGSUSED*/
451 static int
452 segmf_gettype(struct seg *seg, caddr_t addr)
453 {
454 	return (MAP_SHARED);
455 }
456 
457 /*ARGSUSED1*/
458 static int
459 segmf_getvp(struct seg *seg, caddr_t addr, struct vnode **vpp)
460 {
461 	struct segmf_data *data = seg->s_data;
462 
463 	*vpp = VTOCVP(data->vp);
464 	return (0);
465 }
466 
467 /*ARGSUSED*/
468 static int
469 segmf_advise(struct seg *seg, caddr_t addr, size_t len, uint_t behav)
470 {
471 	return (0);
472 }
473 
474 /*ARGSUSED*/
475 static void
476 segmf_dump(struct seg *seg)
477 {}
478 
479 /*ARGSUSED*/
480 static int
481 segmf_pagelock(struct seg *seg, caddr_t addr, size_t len,
482     struct page ***ppp, enum lock_type type, enum seg_rw rw)
483 {
484 	return (ENOTSUP);
485 }
486 
487 /*ARGSUSED*/
488 static int
489 segmf_setpagesize(struct seg *seg, caddr_t addr, size_t len, uint_t szc)
490 {
491 	return (ENOTSUP);
492 }
493 
494 static int
495 segmf_getmemid(struct seg *seg, caddr_t addr, memid_t *memid)
496 {
497 	struct segmf_data *data = seg->s_data;
498 
499 	memid->val[0] = (uintptr_t)VTOCVP(data->vp);
500 	memid->val[1] = (uintptr_t)seg_page(seg, addr);
501 	return (0);
502 }
503 
504 /*ARGSUSED*/
505 static lgrp_mem_policy_info_t *
506 segmf_getpolicy(struct seg *seg, caddr_t addr)
507 {
508 	return (NULL);
509 }
510 
511 /*ARGSUSED*/
512 static int
513 segmf_capable(struct seg *seg, segcapability_t capability)
514 {
515 	return (0);
516 }
517 
518 /*
519  * Add a set of contiguous foreign MFNs to the segment. soft-locking them.  The
520  * pre-faulting is necessary due to live migration; in particular we must
521  * return an error in response to IOCTL_PRIVCMD_MMAPBATCH rather than faulting
522  * later on a bad MFN.  Whilst this isn't necessary for the other MMAP
523  * ioctl()s, we lock them too, as they should be transitory.
524  */
525 int
526 segmf_add_mfns(struct seg *seg, caddr_t addr, mfn_t mfn,
527     pgcnt_t pgcnt, domid_t domid)
528 {
529 	struct segmf_data *data = seg->s_data;
530 	pgcnt_t base;
531 	faultcode_t fc;
532 	pgcnt_t i;
533 	int error = 0;
534 
535 	if (seg->s_ops != &segmf_ops)
536 		return (EINVAL);
537 
538 	/*
539 	 * Don't mess with dom0.
540 	 *
541 	 * Only allow the domid to be set once for the segment.
542 	 * After that attempts to add mappings to this segment for
543 	 * other domains explicitly fails.
544 	 */
545 
546 	if (domid == 0 || domid == DOMID_SELF)
547 		return (EACCES);
548 
549 	mutex_enter(&data->lock);
550 
551 	if (data->domid == 0)
552 		data->domid = domid;
553 
554 	if (data->domid != domid) {
555 		error = EINVAL;
556 		goto out;
557 	}
558 
559 	base = seg_page(seg, addr);
560 
561 	for (i = 0; i < pgcnt; i++) {
562 		data->map[base + i].t_type = SEGMF_MAP_MFN;
563 		data->map[base + i].u.m.m_mfn = mfn++;
564 	}
565 
566 	fc = segmf_fault_range(seg->s_as->a_hat, seg, addr,
567 	    pgcnt * MMU_PAGESIZE, F_SOFTLOCK, S_OTHER);
568 
569 	if (fc != 0) {
570 		error = fc_decode(fc);
571 		for (i = 0; i < pgcnt; i++) {
572 			data->map[base + i].t_type = SEGMF_MAP_EMPTY;
573 		}
574 	}
575 
576 out:
577 	mutex_exit(&data->lock);
578 	return (error);
579 }
580 
581 int
582 segmf_add_grefs(struct seg *seg, caddr_t addr, uint_t flags,
583     grant_ref_t *grefs, uint_t cnt, domid_t domid)
584 {
585 	struct segmf_data *data;
586 	segmf_map_t *map;
587 	faultcode_t fc;
588 	uint_t idx;
589 	uint_t i;
590 	int e;
591 
592 	if (seg->s_ops != &segmf_ops)
593 		return (EINVAL);
594 
595 	/*
596 	 * Don't mess with dom0.
597 	 *
598 	 * Only allow the domid to be set once for the segment.
599 	 * After that attempts to add mappings to this segment for
600 	 * other domains explicitly fails.
601 	 */
602 
603 	if (domid == 0 || domid == DOMID_SELF)
604 		return (EACCES);
605 
606 	data = seg->s_data;
607 	idx = seg_page(seg, addr);
608 	map = &data->map[idx];
609 	e = 0;
610 
611 	mutex_enter(&data->lock);
612 
613 	if (data->domid == 0)
614 		data->domid = domid;
615 
616 	if (data->domid != domid) {
617 		e = EINVAL;
618 		goto out;
619 	}
620 
621 	/* store away the grefs passed in then fault in the pages */
622 	for (i = 0; i < cnt; i++) {
623 		map[i].t_type = SEGMF_MAP_GREF;
624 		map[i].u.g.g_gref = grefs[i];
625 		map[i].u.g.g_handle = 0;
626 		map[i].u.g.g_flags = 0;
627 		if (flags & SEGMF_GREF_WR) {
628 			map[i].u.g.g_flags |= SEGMF_GFLAGS_WR;
629 		}
630 	}
631 	fc = segmf_fault_gref_range(seg, addr, cnt);
632 	if (fc != 0) {
633 		e = fc_decode(fc);
634 		for (i = 0; i < cnt; i++) {
635 			data->map[i].t_type = SEGMF_MAP_EMPTY;
636 		}
637 	}
638 
639 out:
640 	mutex_exit(&data->lock);
641 	return (e);
642 }
643 
644 int
645 segmf_release_grefs(struct seg *seg, caddr_t addr, uint_t cnt)
646 {
647 	gnttab_unmap_grant_ref_t mapop[SEGMF_MAX_GREFS];
648 	struct segmf_data *data;
649 	segmf_map_t *map;
650 	uint_t idx;
651 	long e;
652 	int i;
653 	int n;
654 
655 
656 	if (cnt > SEGMF_MAX_GREFS) {
657 		return (-1);
658 	}
659 
660 	idx = seg_page(seg, addr);
661 	data = seg->s_data;
662 	map = &data->map[idx];
663 
664 	bzero(mapop, sizeof (gnttab_unmap_grant_ref_t) * cnt);
665 
666 	/*
667 	 * for each entry which isn't empty and is currently mapped,
668 	 * set it up for an unmap then mark them empty.
669 	 */
670 	n = 0;
671 	for (i = 0; i < cnt; i++) {
672 		ASSERT(map[i].t_type != SEGMF_MAP_MFN);
673 		if ((map[i].t_type == SEGMF_MAP_GREF) &&
674 		    (map[i].u.g.g_flags & SEGMF_GFLAGS_MAPPED)) {
675 			mapop[n].handle = map[i].u.g.g_handle;
676 			mapop[n].host_addr = map[i].u.g.g_ptep;
677 			mapop[n].dev_bus_addr = 0;
678 			n++;
679 		}
680 		map[i].t_type = SEGMF_MAP_EMPTY;
681 	}
682 
683 	/* if there's nothing to unmap, just return */
684 	if (n == 0) {
685 		return (0);
686 	}
687 
688 	e = HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &mapop, n);
689 	if (e != 0) {
690 		return (-1);
691 	}
692 
693 	return (0);
694 }
695 
696 
697 void
698 segmf_add_gref_pte(struct seg *seg, caddr_t addr, uint64_t pte_ma)
699 {
700 	struct segmf_data *data;
701 	uint_t idx;
702 
703 	idx = seg_page(seg, addr);
704 	data = seg->s_data;
705 
706 	data->map[idx].u.g.g_ptep = pte_ma;
707 }
708 
709 
710 static int
711 segmf_fault_gref_range(struct seg *seg, caddr_t addr, size_t cnt)
712 {
713 	gnttab_map_grant_ref_t mapop[SEGMF_MAX_GREFS];
714 	struct segmf_data *data;
715 	segmf_map_t *map;
716 	uint_t idx;
717 	int e;
718 	int i;
719 
720 
721 	if (cnt > SEGMF_MAX_GREFS) {
722 		return (-1);
723 	}
724 
725 	data = seg->s_data;
726 	idx = seg_page(seg, addr);
727 	map = &data->map[idx];
728 
729 	bzero(mapop, sizeof (gnttab_map_grant_ref_t) * cnt);
730 
731 	ASSERT(map->t_type == SEGMF_MAP_GREF);
732 
733 	/*
734 	 * map in each page passed in into the user apps AS. We do this by
735 	 * passing the MA of the actual pte of the mapping to the hypervisor.
736 	 */
737 	for (i = 0; i < cnt; i++) {
738 		mapop[i].host_addr = map[i].u.g.g_ptep;
739 		mapop[i].dom = data->domid;
740 		mapop[i].ref = map[i].u.g.g_gref;
741 		mapop[i].flags = GNTMAP_host_map | GNTMAP_application_map |
742 		    GNTMAP_contains_pte;
743 		if (!(map[i].u.g.g_flags & SEGMF_GFLAGS_WR)) {
744 			mapop[i].flags |= GNTMAP_readonly;
745 		}
746 	}
747 	e = xen_map_gref(GNTTABOP_map_grant_ref, mapop, cnt, B_TRUE);
748 	if ((e != 0) || (mapop[0].status != GNTST_okay)) {
749 		return (FC_MAKE_ERR(EFAULT));
750 	}
751 
752 	/* save handle for segmf_release_grefs() and mark it as mapped */
753 	for (i = 0; i < cnt; i++) {
754 		ASSERT(mapop[i].status == GNTST_okay);
755 		map[i].u.g.g_handle = mapop[i].handle;
756 		map[i].u.g.g_flags |= SEGMF_GFLAGS_MAPPED;
757 	}
758 
759 	return (0);
760 }
761 
762 static struct seg_ops segmf_ops = {
763 	segmf_dup,
764 	segmf_unmap,
765 	segmf_free,
766 	segmf_fault,
767 	segmf_faulta,
768 	segmf_setprot,
769 	segmf_checkprot,
770 	(int (*)())segmf_kluster,
771 	(size_t (*)(struct seg *))NULL,	/* swapout */
772 	segmf_sync,
773 	segmf_incore,
774 	segmf_lockop,
775 	segmf_getprot,
776 	segmf_getoffset,
777 	segmf_gettype,
778 	segmf_getvp,
779 	segmf_advise,
780 	segmf_dump,
781 	segmf_pagelock,
782 	segmf_setpagesize,
783 	segmf_getmemid,
784 	segmf_getpolicy,
785 	segmf_capable,
786 	seg_inherit_notsup
787 };
788