xref: /freebsd/sys/dev/vmware/vmci/vmci_kernel_if.c (revision 87b759f0fa1f7554d50ce640c40138512bbded44)
1 /*-
2  * Copyright (c) 2018 VMware, Inc.
3  *
4  * SPDX-License-Identifier: (BSD-2-Clause OR GPL-2.0)
5  */
6 
7 /* This file implements defines and helper functions. */
8 
9 #include <sys/cdefs.h>
10 #include <sys/malloc.h>
11 #include <sys/proc.h>
12 #include <sys/uio.h>
13 
14 #include <machine/bus.h>
15 
16 #include "vmci.h"
17 #include "vmci_defs.h"
18 #include "vmci_kernel_defs.h"
19 #include "vmci_kernel_if.h"
20 #include "vmci_queue.h"
21 
22 struct vmci_queue_kernel_if {
23 	size_t			num_pages;	/* Num pages incl. header. */
24 	struct vmci_dma_alloc	*dmas;		/* For dma alloc. */
25 };
26 
27 /*
28  *------------------------------------------------------------------------------
29  *
30  * vmci_init_lock
31  *
32  *     Initializes the lock. Must be called before use.
33  *
34  * Results:
35  *     Always VMCI_SUCCESS.
36  *
37  * Side effects:
38  *     Thread can block.
39  *
40  *------------------------------------------------------------------------------
41  */
42 
43 int
44 vmci_init_lock(vmci_lock *lock, char *name)
45 {
46 
47 	mtx_init(lock, name, NULL, MTX_DEF | MTX_NOWITNESS);
48 	return (VMCI_SUCCESS);
49 }
50 
51 /*
52  *------------------------------------------------------------------------------
53  *
54  * vmci_cleanup_lock
55  *
56  *     Cleanup the lock. Must be called before deallocating lock.
57  *
58  * Results:
59  *     None
60  *
61  * Side effects:
62  *     Deletes kernel lock state
63  *
64  *------------------------------------------------------------------------------
65  */
66 
67 void
68 vmci_cleanup_lock(vmci_lock *lock)
69 {
70 
71 	if (mtx_initialized(lock))
72 		mtx_destroy(lock);
73 }
74 
75 /*
76  *------------------------------------------------------------------------------
77  *
78  * vmci_grab_lock
79  *
80  *     Grabs the given lock.
81  *
82  * Results:
83  *      None
84  *
85  * Side effects:
86  *      Thread can block.
87  *
88  *------------------------------------------------------------------------------
89  */
90 
91 void
92 vmci_grab_lock(vmci_lock *lock)
93 {
94 
95 	mtx_lock(lock);
96 }
97 
98 /*
99  *------------------------------------------------------------------------------
100  *
101  * vmci_release_lock
102  *
103  *     Releases the given lock.
104  *
105  * Results:
106  *     None
107  *
108  * Side effects:
109  *     A thread blocked on this lock may wake up.
110  *
111  *------------------------------------------------------------------------------
112  */
113 
114 void
115 vmci_release_lock(vmci_lock *lock)
116 {
117 
118 	mtx_unlock(lock);
119 }
120 
121 /*
122  *------------------------------------------------------------------------------
123  *
124  * vmci_grab_lock_bh
125  *
126  *     Grabs the given lock.
127  *
128  * Results:
129  *     None
130  *
131  * Side effects:
132  *     None.
133  *
134  *------------------------------------------------------------------------------
135  */
136 
137 void
138 vmci_grab_lock_bh(vmci_lock *lock)
139 {
140 
141 	mtx_lock(lock);
142 }
143 
144 /*
145  *------------------------------------------------------------------------------
146  *
147  * vmci_release_lock_bh
148  *
149  *     Releases the given lock.
150  *
151  * Results:
152  *     None
153  *
154  * Side effects:
155  *     None.
156  *
157  *------------------------------------------------------------------------------
158  */
159 
160 void
161 vmci_release_lock_bh(vmci_lock *lock)
162 {
163 
164 	mtx_unlock(lock);
165 }
166 
167 /*
168  *------------------------------------------------------------------------------
169  *
170  * vmci_initialized_lock
171  *
172  *     Returns whether a lock has been initialized.
173  *
174  * Results:
175  *     Return 1 if initialized or 0 if unininitialized.
176  *
177  * Side effects:
178  *     None
179  *
180  *------------------------------------------------------------------------------
181  */
182 
183 int
184 vmci_initialized_lock(vmci_lock *lock)
185 {
186 
187 	return mtx_initialized(lock);
188 }
189 
190 /*
191  *------------------------------------------------------------------------------
192  *
193  * vmci_alloc_kernel_mem
194  *
195  *     Allocate physically contiguous memory for the VMCI driver.
196  *
197  * Results:
198  *     The address allocated or NULL on error.
199  *
200  *
201  * Side effects:
202  *     Memory may be allocated.
203  *
204  *------------------------------------------------------------------------------
205  */
206 
207 void *
208 vmci_alloc_kernel_mem(size_t size, int flags)
209 {
210 	void *ptr;
211 
212 	if ((flags & VMCI_MEMORY_ATOMIC) != 0)
213 		ptr = contigmalloc(size, M_DEVBUF, M_NOWAIT, 0, 0xFFFFFFFF,
214 		    8, 1024 * 1024);
215 	else
216 		ptr = contigmalloc(size, M_DEVBUF, M_WAITOK, 0, 0xFFFFFFFF,
217 		    8, 1024 * 1024);
218 
219 	return (ptr);
220 }
221 
222 /*
223  *------------------------------------------------------------------------------
224  *
225  * vmci_free_kernel_mem
226  *
227  *     Free kernel memory allocated for the VMCI driver.
228  *
229  * Results:
230  *     None.
231  *
232  * Side effects:
233  *     Memory is freed.
234  *
235  *------------------------------------------------------------------------------
236  */
237 
238 void
239 vmci_free_kernel_mem(void *ptr, size_t size __unused)
240 {
241 
242 	free(ptr, M_DEVBUF);
243 }
244 
245 /*
246  *------------------------------------------------------------------------------
247  *
248  * vmci_can_schedule_delayed_work --
249  *
250  *     Checks to see if the given platform supports delayed work callbacks.
251  *
252  * Results:
253  *     true if it does. false otherwise.
254  *
255  * Side effects:
256  *     None.
257  *
258  *------------------------------------------------------------------------------
259  */
260 
261 bool
262 vmci_can_schedule_delayed_work(void)
263 {
264 
265 	return (true);
266 }
267 
268 /*
269  *------------------------------------------------------------------------------
270  *
271  * vmci_schedule_delayed_work --
272  *
273  *     Schedule the specified callback.
274  *
275  * Results:
276  *     Zero on success, error code otherwise.
277  *
278  * Side effects:
279  *     None.
280  *
281  *------------------------------------------------------------------------------
282  */
283 
284 int
285 vmci_schedule_delayed_work(vmci_work_fn *work_fn, void *data)
286 {
287 
288 	return (vmci_schedule_delayed_work_fn(work_fn, data));
289 }
290 
291 /*
292  *------------------------------------------------------------------------------
293  *
294  * vmci_create_event --
295  *
296  * Results:
297  *     None.
298  *
299  * Side effects:
300  *     None.
301  *
302  *------------------------------------------------------------------------------
303  */
304 
305 void
306 vmci_create_event(vmci_event *event)
307 {
308 
309 	sema_init(event, 0, "vmci_event");
310 }
311 
312 /*
313  *------------------------------------------------------------------------------
314  *
315  * vmci_destroy_event --
316  *
317  * Results:
318  *     None.
319  *
320  * Side effects:
321  *     None.
322  *
323  *------------------------------------------------------------------------------
324  */
325 
326 void
327 vmci_destroy_event(vmci_event *event)
328 {
329 
330 	if (mtx_owned(&event->sema_mtx))
331 		sema_destroy(event);
332 }
333 
334 /*
335  *------------------------------------------------------------------------------
336  *
337  * vmci_signal_event --
338  *
339  * Results:
340  *     None.
341  *
342  * Side effects:
343  *     None.
344  *
345  *------------------------------------------------------------------------------
346  */
347 
348 void
349 vmci_signal_event(vmci_event *event)
350 {
351 
352 	sema_post(event);
353 }
354 
355 /*
356  *------------------------------------------------------------------------------
357  *
358  * vmci_wait_on_event --
359  *
360  * Results:
361  *     None.
362  *
363  * Side effects:
364  *     None.
365  *
366  *------------------------------------------------------------------------------
367  */
368 
369 void
370 vmci_wait_on_event(vmci_event *event, vmci_event_release_cb release_cb,
371     void *client_data)
372 {
373 
374 	release_cb(client_data);
375 	sema_wait(event);
376 }
377 
378 /*
379  *------------------------------------------------------------------------------
380  *
381  * vmci_mutex_init --
382  *
383  *     Initializes the mutex. Must be called before use.
384  *
385  * Results:
386  *     Success.
387  *
388  * Side effects:
389  *     None.
390  *
391  *------------------------------------------------------------------------------
392  */
393 
394 int
395 vmci_mutex_init(vmci_mutex *mutex, char *name)
396 {
397 
398 	mtx_init(mutex, name, NULL, MTX_DEF | MTX_NOWITNESS);
399 	return (VMCI_SUCCESS);
400 }
401 
402 /*
403  *------------------------------------------------------------------------------
404  *
405  * vmci_mutex_destroy --
406  *
407  *     Destroys the mutex.
408  *
409  * Results:
410  *     None.
411  *
412  * Side effects:
413  *     None.
414  *
415  *------------------------------------------------------------------------------
416  */
417 
418 void
419 vmci_mutex_destroy(vmci_mutex *mutex)
420 {
421 
422 	mtx_destroy(mutex);
423 }
424 
425 /*
426  *------------------------------------------------------------------------------
427  *
428  * vmci_mutex_acquire --
429  *
430  *     Acquires the mutex.
431  *
432  * Results:
433  *     None.
434  *
435  * Side effects:
436  *     Thread may block.
437  *
438  *------------------------------------------------------------------------------
439  */
440 
441 void
442 vmci_mutex_acquire(vmci_mutex *mutex)
443 {
444 
445 	mtx_lock(mutex);
446 }
447 
448 /*
449  *------------------------------------------------------------------------------
450  *
451  * vmci_mutex_release --
452  *
453  *     Releases the mutex.
454  *
455  * Results:
456  *     None.
457  *
458  * Side effects:
459  *     May wake up the thread blocking on this mutex.
460  *
461  *------------------------------------------------------------------------------
462  */
463 
464 void
465 vmci_mutex_release(vmci_mutex *mutex)
466 {
467 
468 	mtx_unlock(mutex);
469 }
470 
471 /*
472  *------------------------------------------------------------------------------
473  *
474  * vmci_mutex_initialized
475  *
476  *     Returns whether a mutex has been initialized.
477  *
478  * Results:
479  *     Return 1 if initialized or 0 if unininitialized.
480  *
481  * Side effects:
482  *     None
483  *
484  *------------------------------------------------------------------------------
485  */
486 
487 int
488 vmci_mutex_initialized(vmci_mutex *mutex)
489 {
490 
491 	return mtx_initialized(mutex);
492 }
493 /*
494  *------------------------------------------------------------------------------
495  *
496  * vmci_alloc_queue --
497  *
498  *     Allocates kernel queue pages of specified size with IOMMU mappings, plus
499  *     space for the queue structure/kernel interface and the queue header.
500  *
501  * Results:
502  *     Pointer to the queue on success, NULL otherwise.
503  *
504  * Side effects:
505  *     Memory is allocated.
506  *
507  *------------------------------------------------------------------------------
508  */
509 
510 void *
511 vmci_alloc_queue(uint64_t size, uint32_t flags)
512 {
513 	struct vmci_queue *queue;
514 	size_t i;
515 	const size_t num_pages = CEILING(size, PAGE_SIZE) + 1;
516 	const size_t dmas_size = num_pages * sizeof(struct vmci_dma_alloc);
517 	const size_t queue_size =
518 	    sizeof(*queue) + sizeof(*(queue->kernel_if)) + dmas_size;
519 
520 	/* Size should be enforced by vmci_qpair_alloc(), double-check here. */
521 	if (size > VMCI_MAX_GUEST_QP_MEMORY) {
522 		ASSERT(false);
523 		return (NULL);
524 	}
525 
526 	queue = malloc(queue_size, M_DEVBUF, M_NOWAIT);
527 	if (!queue)
528 		return (NULL);
529 
530 	queue->q_header = NULL;
531 	queue->saved_header = NULL;
532 	queue->kernel_if = (struct vmci_queue_kernel_if *)(queue + 1);
533 	queue->kernel_if->num_pages = num_pages;
534 	queue->kernel_if->dmas = (struct vmci_dma_alloc *)(queue->kernel_if +
535 	    1);
536 	for (i = 0; i < num_pages; i++) {
537 		vmci_dma_malloc(PAGE_SIZE, 1, &queue->kernel_if->dmas[i]);
538 		if (!queue->kernel_if->dmas[i].dma_vaddr) {
539 			/* Size excl. the header. */
540 			vmci_free_queue(queue, i * PAGE_SIZE);
541 			return (NULL);
542 		}
543 	}
544 
545 	/* Queue header is the first page. */
546 	queue->q_header = (void *)queue->kernel_if->dmas[0].dma_vaddr;
547 
548 	return ((void *)queue);
549 }
550 
551 /*
552  *------------------------------------------------------------------------------
553  *
554  * vmci_free_queue --
555  *
556  *     Frees kernel VA space for a given queue and its queue header, and frees
557  *     physical data pages.
558  *
559  * Results:
560  *     None.
561  *
562  * Side effects:
563  *     Memory is freed.
564  *
565  *------------------------------------------------------------------------------
566  */
567 
568 void
569 vmci_free_queue(void *q, uint64_t size)
570 {
571 	struct vmci_queue *queue = q;
572 
573 	if (queue) {
574 		const size_t num_pages = CEILING(size, PAGE_SIZE) + 1;
575 		uint64_t i;
576 
577 		/* Given size doesn't include header, so add in a page here. */
578 		for (i = 0; i < num_pages; i++)
579 			vmci_dma_free(&queue->kernel_if->dmas[i]);
580 		free(queue, M_DEVBUF);
581 	}
582 }
583 
584 /*
585  *------------------------------------------------------------------------------
586  *
587  * vmci_alloc_ppn_set --
588  *
589  *     Allocates two list of PPNs --- one for the pages in the produce queue,
590  *     and the other for the pages in the consume queue. Intializes the list of
591  *     PPNs with the page frame numbers of the KVA for the two queues (and the
592  *     queue headers).
593  *
594  * Results:
595  *     Success or failure.
596  *
597  * Side effects:
598  *     Memory may be allocated.
599  *
600  *-----------------------------------------------------------------------------
601  */
602 
603 int
604 vmci_alloc_ppn_set(void *prod_q, uint64_t num_produce_pages, void *cons_q,
605     uint64_t num_consume_pages, struct ppn_set *ppn_set)
606 {
607 	struct vmci_queue *consume_q = cons_q;
608 	struct vmci_queue *produce_q = prod_q;
609 	vmci_ppn_list consume_ppns;
610 	vmci_ppn_list produce_ppns;
611 	uint64_t i;
612 
613 	if (!produce_q || !num_produce_pages || !consume_q ||
614 	    !num_consume_pages || !ppn_set)
615 		return (VMCI_ERROR_INVALID_ARGS);
616 
617 	if (ppn_set->initialized)
618 		return (VMCI_ERROR_ALREADY_EXISTS);
619 
620 	produce_ppns =
621 	    vmci_alloc_kernel_mem(num_produce_pages * sizeof(*produce_ppns),
622 	    VMCI_MEMORY_NORMAL);
623 	if (!produce_ppns)
624 		return (VMCI_ERROR_NO_MEM);
625 
626 	consume_ppns =
627 	    vmci_alloc_kernel_mem(num_consume_pages * sizeof(*consume_ppns),
628 	    VMCI_MEMORY_NORMAL);
629 	if (!consume_ppns) {
630 		vmci_free_kernel_mem(produce_ppns,
631 		    num_produce_pages * sizeof(*produce_ppns));
632 		return (VMCI_ERROR_NO_MEM);
633 	}
634 
635 	for (i = 0; i < num_produce_pages; i++) {
636 		unsigned long pfn;
637 
638 		produce_ppns[i] =
639 		    pfn = produce_q->kernel_if->dmas[i].dma_paddr >> PAGE_SHIFT;
640 
641 		/*
642 		 * Fail allocation if PFN isn't supported by hypervisor.
643 		 */
644 
645 		if (sizeof(pfn) >
646 		    sizeof(*produce_ppns) && pfn != produce_ppns[i])
647 			goto ppn_error;
648 	}
649 	for (i = 0; i < num_consume_pages; i++) {
650 		unsigned long pfn;
651 
652 		consume_ppns[i] =
653 		    pfn = consume_q->kernel_if->dmas[i].dma_paddr >> PAGE_SHIFT;
654 
655 		/*
656 		 * Fail allocation if PFN isn't supported by hypervisor.
657 		 */
658 
659 		if (sizeof(pfn) >
660 		    sizeof(*consume_ppns) && pfn != consume_ppns[i])
661 			goto ppn_error;
662 	}
663 
664 	ppn_set->num_produce_pages = num_produce_pages;
665 	ppn_set->num_consume_pages = num_consume_pages;
666 	ppn_set->produce_ppns = produce_ppns;
667 	ppn_set->consume_ppns = consume_ppns;
668 	ppn_set->initialized = true;
669 	return (VMCI_SUCCESS);
670 
671 ppn_error:
672 	vmci_free_kernel_mem(produce_ppns, num_produce_pages *
673 	    sizeof(*produce_ppns));
674 	vmci_free_kernel_mem(consume_ppns, num_consume_pages *
675 	    sizeof(*consume_ppns));
676 	return (VMCI_ERROR_INVALID_ARGS);
677 }
678 
679 /*
680  *------------------------------------------------------------------------------
681  *
682  * vmci_free_ppn_set --
683  *
684  *     Frees the two list of PPNs for a queue pair.
685  *
686  * Results:
687  *     None.
688  *
689  * Side effects:
690  *     None.
691  *
692  *------------------------------------------------------------------------------
693  */
694 
695 void
696 vmci_free_ppn_set(struct ppn_set *ppn_set)
697 {
698 
699 	ASSERT(ppn_set);
700 	if (ppn_set->initialized) {
701 		/* Do not call these functions on NULL inputs. */
702 		ASSERT(ppn_set->produce_ppns && ppn_set->consume_ppns);
703 		vmci_free_kernel_mem(ppn_set->produce_ppns,
704 		    ppn_set->num_produce_pages *
705 		    sizeof(*ppn_set->produce_ppns));
706 		vmci_free_kernel_mem(ppn_set->consume_ppns,
707 		    ppn_set->num_consume_pages *
708 		    sizeof(*ppn_set->consume_ppns));
709 	}
710 	memset(ppn_set, 0, sizeof(*ppn_set));
711 }
712 
713 /*
714  *------------------------------------------------------------------------------
715  *
716  * vmci_populate_ppn_list --
717  *
718  *     Populates the list of PPNs in the hypercall structure with the PPNS
719  *     of the produce queue and the consume queue.
720  *
721  * Results:
722  *     VMCI_SUCCESS.
723  *
724  * Side effects:
725  *     None.
726  *
727  *------------------------------------------------------------------------------
728  */
729 
730 int
731 vmci_populate_ppn_list(uint8_t *call_buf, const struct ppn_set *ppn_set)
732 {
733 
734 	ASSERT(call_buf && ppn_set && ppn_set->initialized);
735 	memcpy(call_buf, ppn_set->produce_ppns,
736 	    ppn_set->num_produce_pages * sizeof(*ppn_set->produce_ppns));
737 	memcpy(call_buf + ppn_set->num_produce_pages *
738 	    sizeof(*ppn_set->produce_ppns), ppn_set->consume_ppns,
739 	    ppn_set->num_consume_pages * sizeof(*ppn_set->consume_ppns));
740 
741 	return (VMCI_SUCCESS);
742 }
743 
744 /*
745  *------------------------------------------------------------------------------
746  *
747  * vmci_memcpy_{to,from}iovec --
748  *
749  *     These helper routines will copy the specified bytes to/from memory that's
750  *     specified as a struct iovec.  The routines can not verify the correctness
751  *     of the struct iovec's contents.
752  *
753  * Results:
754  *      None.
755  *
756  * Side effects:
757  *      None.
758  *
759  *------------------------------------------------------------------------------
760  */
761 
762 static inline void
763 vmci_memcpy_toiovec(struct iovec *iov, uint8_t *src, size_t len)
764 {
765 
766 	while (len > 0) {
767 		if (iov->iov_len) {
768 			size_t to_copy = MIN(iov->iov_len, len);
769 			memcpy(iov->iov_base, src, to_copy);
770 			src += to_copy;
771 			len -= to_copy;
772 			iov->iov_base = (void *)((uintptr_t) iov->iov_base +
773 			    to_copy);
774 			iov->iov_len -= to_copy;
775 		}
776 		iov++;
777 	}
778 }
779 
780 static inline void
781 vmci_memcpy_fromiovec(uint8_t *dst, struct iovec *iov, size_t len)
782 {
783 
784 	while (len > 0) {
785 		if (iov->iov_len) {
786 			size_t to_copy = MIN(iov->iov_len, len);
787 			memcpy(dst, iov->iov_base, to_copy);
788 			dst += to_copy;
789 			len -= to_copy;
790 			iov->iov_base = (void *)((uintptr_t) iov->iov_base +
791 			    to_copy);
792 			iov->iov_len -= to_copy;
793 		}
794 		iov++;
795 	}
796 }
797 
798 /*
799  *------------------------------------------------------------------------------
800  *
801  * __vmci_memcpy_to_queue --
802  *
803  *     Copies from a given buffer or iovector to a VMCI Queue. Assumes that
804  *     offset + size does not wrap around in the queue.
805  *
806  * Results:
807  *     Zero on success, negative error code on failure.
808  *
809  * Side effects:
810  *     None.
811  *
812  *------------------------------------------------------------------------------
813  */
814 
815 #pragma GCC diagnostic ignored "-Wcast-qual"
816 static int
817 __vmci_memcpy_to_queue(struct vmci_queue *queue, uint64_t queue_offset,
818     const void *src, size_t size, bool is_iovec)
819 {
820 	struct vmci_queue_kernel_if *kernel_if = queue->kernel_if;
821 	size_t bytes_copied = 0;
822 
823 	while (bytes_copied < size) {
824 		const uint64_t page_index =
825 		    (queue_offset + bytes_copied) / PAGE_SIZE;
826 		const size_t page_offset =
827 		    (queue_offset + bytes_copied) & (PAGE_SIZE - 1);
828 		void *va;
829 		size_t to_copy;
830 
831 		/* Skip header. */
832 		va = (void *)kernel_if->dmas[page_index + 1].dma_vaddr;
833 
834 		ASSERT(va);
835 		/*
836 		 * Fill up the page if we have enough payload, or else
837 		 * copy the remaining bytes.
838 		 */
839 		to_copy = MIN(PAGE_SIZE - page_offset, size - bytes_copied);
840 
841 		if (is_iovec) {
842 			struct iovec *iov = (struct iovec *)src;
843 
844 			/* The iovec will track bytes_copied internally. */
845 			vmci_memcpy_fromiovec((uint8_t *)va + page_offset,
846 			    iov, to_copy);
847 		} else
848 			memcpy((uint8_t *)va + page_offset,
849 			    (uint8_t *)src + bytes_copied, to_copy);
850 		bytes_copied += to_copy;
851 	}
852 
853 	return (VMCI_SUCCESS);
854 }
855 
856 /*
857  *------------------------------------------------------------------------------
858  *
859  * __vmci_memcpy_from_queue --
860  *
861  *     Copies to a given buffer or iovector from a VMCI Queue. Assumes that
862  *     offset + size does not wrap around in the queue.
863  *
864  * Results:
865  *     Zero on success, negative error code on failure.
866  *
867  * Side effects:
868  *     None.
869  *
870  *------------------------------------------------------------------------------
871  */
872 
873 static int
874 __vmci_memcpy_from_queue(void *dest, const struct vmci_queue *queue,
875     uint64_t queue_offset, size_t size, bool is_iovec)
876 {
877 	struct vmci_queue_kernel_if *kernel_if = queue->kernel_if;
878 	size_t bytes_copied = 0;
879 
880 	while (bytes_copied < size) {
881 		const uint64_t page_index =
882 		    (queue_offset + bytes_copied) / PAGE_SIZE;
883 		const size_t page_offset =
884 		    (queue_offset + bytes_copied) & (PAGE_SIZE - 1);
885 		void *va;
886 		size_t to_copy;
887 
888 		/* Skip header. */
889 		va = (void *)kernel_if->dmas[page_index + 1].dma_vaddr;
890 
891 		ASSERT(va);
892 		/*
893 		 * Fill up the page if we have enough payload, or else
894 		 * copy the remaining bytes.
895 		 */
896 		to_copy = MIN(PAGE_SIZE - page_offset, size - bytes_copied);
897 
898 		if (is_iovec) {
899 			struct iovec *iov = (struct iovec *)dest;
900 
901 			/* The iovec will track bytesCopied internally. */
902 			vmci_memcpy_toiovec(iov, (uint8_t *)va +
903 			    page_offset, to_copy);
904 		} else
905 			memcpy((uint8_t *)dest + bytes_copied,
906 			    (uint8_t *)va + page_offset, to_copy);
907 
908 		bytes_copied += to_copy;
909 	}
910 
911 	return (VMCI_SUCCESS);
912 }
913 
914 /*
915  *------------------------------------------------------------------------------
916  *
917  * vmci_memcpy_to_queue --
918  *
919  *     Copies from a given buffer to a VMCI Queue.
920  *
921  * Results:
922  *     Zero on success, negative error code on failure.
923  *
924  * Side effects:
925  *     None.
926  *
927  *------------------------------------------------------------------------------
928  */
929 
930 int
931 vmci_memcpy_to_queue(struct vmci_queue *queue, uint64_t queue_offset,
932     const void *src, size_t src_offset, size_t size, int buf_type,
933     bool can_block)
934 {
935 
936 	ASSERT(can_block);
937 
938 	return (__vmci_memcpy_to_queue(queue, queue_offset,
939 	    (uint8_t *)src + src_offset, size, false));
940 }
941 
942 /*
943  *------------------------------------------------------------------------------
944  *
945  * vmci_memcpy_from_queue --
946  *
947  *      Copies to a given buffer from a VMCI Queue.
948  *
949  * Results:
950  *      Zero on success, negative error code on failure.
951  *
952  * Side effects:
953  *      None.
954  *
955  *------------------------------------------------------------------------------
956  */
957 
958 int
959 vmci_memcpy_from_queue(void *dest, size_t dest_offset,
960     const struct vmci_queue *queue, uint64_t queue_offset, size_t size,
961     int buf_type, bool can_block)
962 {
963 
964 	ASSERT(can_block);
965 
966 	return (__vmci_memcpy_from_queue((uint8_t *)dest + dest_offset,
967 	    queue, queue_offset, size, false));
968 }
969 
970 /*
971  *------------------------------------------------------------------------------
972  *
973  * vmci_memcpy_to_queue_local --
974  *
975  *     Copies from a given buffer to a local VMCI queue. This is the
976  *     same as a regular copy.
977  *
978  * Results:
979  *     Zero on success, negative error code on failure.
980  *
981  * Side effects:
982  *     None.
983  *
984  *------------------------------------------------------------------------------
985  */
986 
987 int
988 vmci_memcpy_to_queue_local(struct vmci_queue *queue, uint64_t queue_offset,
989     const void *src, size_t src_offset, size_t size, int buf_type,
990     bool can_block)
991 {
992 
993 	ASSERT(can_block);
994 
995 	return (__vmci_memcpy_to_queue(queue, queue_offset,
996 	    (uint8_t *)src + src_offset, size, false));
997 }
998 
999 /*
1000  *------------------------------------------------------------------------------
1001  *
1002  * vmci_memcpy_from_queue_local --
1003  *
1004  *     Copies to a given buffer from a VMCI Queue.
1005  *
1006  * Results:
1007  *     Zero on success, negative error code on failure.
1008  *
1009  * Side effects:
1010  *     None.
1011  *
1012  *------------------------------------------------------------------------------
1013  */
1014 
1015 int
1016 vmci_memcpy_from_queue_local(void *dest, size_t dest_offset,
1017     const struct vmci_queue *queue, uint64_t queue_offset, size_t size,
1018     int buf_type, bool can_block)
1019 {
1020 
1021 	ASSERT(can_block);
1022 
1023 	return (__vmci_memcpy_from_queue((uint8_t *)dest + dest_offset,
1024 	    queue, queue_offset, size, false));
1025 }
1026 
1027 /*------------------------------------------------------------------------------
1028  *
1029  * vmci_memcpy_to_queue_v --
1030  *
1031  *     Copies from a given iovec from a VMCI Queue.
1032  *
1033  * Results:
1034  *     Zero on success, negative error code on failure.
1035  *
1036  * Side effects:
1037  *     None.
1038  *
1039  *------------------------------------------------------------------------------
1040  */
1041 
1042 int
1043 vmci_memcpy_to_queue_v(struct vmci_queue *queue, uint64_t queue_offset,
1044     const void *src, size_t src_offset, size_t size, int buf_type,
1045     bool can_block)
1046 {
1047 
1048 	ASSERT(can_block);
1049 
1050 	/*
1051 	 * We ignore src_offset because src is really a struct iovec * and will
1052 	 * maintain offset internally.
1053 	 */
1054 	return (__vmci_memcpy_to_queue(queue, queue_offset, src, size,
1055 	    true));
1056 }
1057 
1058 /*
1059  *------------------------------------------------------------------------------
1060  *
1061  * vmci_memcpy_from_queue_v --
1062  *
1063  *     Copies to a given iovec from a VMCI Queue.
1064  *
1065  * Results:
1066  *     Zero on success, negative error code on failure.
1067  *
1068  * Side effects:
1069  *     None.
1070  *
1071  *------------------------------------------------------------------------------
1072  */
1073 
1074 int
1075 vmci_memcpy_from_queue_v(void *dest, size_t dest_offset,
1076     const struct vmci_queue *queue, uint64_t queue_offset, size_t size,
1077     int buf_type, bool can_block)
1078 {
1079 
1080 	ASSERT(can_block);
1081 
1082 	/*
1083 	 * We ignore dest_offset because dest is really a struct iovec * and
1084 	 * will maintain offset internally.
1085 	 */
1086 	return (__vmci_memcpy_from_queue(dest, queue, queue_offset, size,
1087 	    true));
1088 }
1089 
1090 /*
1091  *------------------------------------------------------------------------------
1092  *
1093  * vmci_read_port_bytes --
1094  *
1095  *     Copy memory from an I/O port to kernel memory.
1096  *
1097  * Results:
1098  *     No results.
1099  *
1100  * Side effects:
1101  *     None.
1102  *
1103  *------------------------------------------------------------------------------
1104  */
1105 
1106 void
1107 vmci_read_port_bytes(vmci_io_handle handle, vmci_io_port port, uint8_t *buffer,
1108     size_t buffer_length)
1109 {
1110 
1111 	insb(port, buffer, buffer_length);
1112 }
1113