1 /* $OpenBSD: criov.c,v 1.9 2002/01/29 15:48:29 jason Exp $ */
2
3 /*-
4 * Copyright (c) 1999 Theo de Raadt
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 *
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
15 * 3. The name of the author may not be used to endorse or promote products
16 * derived from this software without specific prior written permission.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
19 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
21 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
23 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 */
29
30 #include <sys/param.h>
31 #include <sys/systm.h>
32 #include <sys/proc.h>
33 #include <sys/errno.h>
34 #include <sys/malloc.h>
35 #include <sys/kernel.h>
36 #include <sys/mbuf.h>
37 #include <sys/uio.h>
38 #include <sys/limits.h>
39 #include <sys/lock.h>
40 #include <sys/sdt.h>
41
42 #include <machine/vmparam.h>
43
44 #include <vm/vm.h>
45 #include <vm/vm_page.h>
46 #include <vm/pmap.h>
47
48 #include <opencrypto/cryptodev.h>
49
50 SDT_PROVIDER_DECLARE(opencrypto);
51
52 /*
53 * These macros are only for avoiding code duplication, as we need to skip
54 * given number of bytes in the same way in several functions below.
55 */
56 #define CUIO_SKIP() do { \
57 KASSERT(off >= 0, ("%s: off %d < 0", __func__, off)); \
58 KASSERT(len >= 0, ("%s: len %d < 0", __func__, len)); \
59 while (off > 0) { \
60 KASSERT(iol >= 0, ("%s: empty in skip", __func__)); \
61 if (off < iov->iov_len) \
62 break; \
63 off -= iov->iov_len; \
64 iol--; \
65 iov++; \
66 } \
67 } while (0)
68
69 #define CVM_PAGE_SKIP() do { \
70 KASSERT(off >= 0, ("%s: off %d < 0", __func__, off)); \
71 KASSERT(len >= 0, ("%s: len %d < 0", __func__, len)); \
72 while (off > 0) { \
73 if (off < PAGE_SIZE) \
74 break; \
75 processed += PAGE_SIZE - off; \
76 off -= PAGE_SIZE - off; \
77 pages++; \
78 } \
79 } while (0)
80
81 static void
cuio_copydata(struct uio * uio,int off,int len,caddr_t cp)82 cuio_copydata(struct uio* uio, int off, int len, caddr_t cp)
83 {
84 struct iovec *iov = uio->uio_iov;
85 int iol __diagused = uio->uio_iovcnt;
86 unsigned count;
87
88 CUIO_SKIP();
89 while (len > 0) {
90 KASSERT(iol >= 0, ("%s: empty", __func__));
91 count = min(iov->iov_len - off, len);
92 bcopy(((caddr_t)iov->iov_base) + off, cp, count);
93 len -= count;
94 cp += count;
95 off = 0;
96 iol--;
97 iov++;
98 }
99 }
100
101 static void
cuio_copyback(struct uio * uio,int off,int len,c_caddr_t cp)102 cuio_copyback(struct uio* uio, int off, int len, c_caddr_t cp)
103 {
104 struct iovec *iov = uio->uio_iov;
105 int iol __diagused = uio->uio_iovcnt;
106 unsigned count;
107
108 CUIO_SKIP();
109 while (len > 0) {
110 KASSERT(iol >= 0, ("%s: empty", __func__));
111 count = min(iov->iov_len - off, len);
112 bcopy(cp, ((caddr_t)iov->iov_base) + off, count);
113 len -= count;
114 cp += count;
115 off = 0;
116 iol--;
117 iov++;
118 }
119 }
120
121 /*
122 * Return the index and offset of location in iovec list.
123 */
124 static int
cuio_getptr(struct uio * uio,int loc,int * off)125 cuio_getptr(struct uio *uio, int loc, int *off)
126 {
127 int ind, len;
128
129 ind = 0;
130 while (loc >= 0 && ind < uio->uio_iovcnt) {
131 len = uio->uio_iov[ind].iov_len;
132 if (len > loc) {
133 *off = loc;
134 return (ind);
135 }
136 loc -= len;
137 ind++;
138 }
139
140 if (ind > 0 && loc == 0) {
141 ind--;
142 *off = uio->uio_iov[ind].iov_len;
143 return (ind);
144 }
145
146 return (-1);
147 }
148
149 #if CRYPTO_MAY_HAVE_VMPAGE
150 /*
151 * Apply function f to the data in a vm_page_t list starting "off" bytes from
152 * the beginning, continuing for "len" bytes.
153 */
154 static int
cvm_page_apply(vm_page_t * pages,int off,int len,int (* f)(void *,const void *,u_int),void * arg)155 cvm_page_apply(vm_page_t *pages, int off, int len,
156 int (*f)(void *, const void *, u_int), void *arg)
157 {
158 int processed __unused;
159 unsigned count;
160 int rval;
161
162 processed = 0;
163 CVM_PAGE_SKIP();
164 while (len > 0) {
165 char *kaddr = (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages));
166 count = min(PAGE_SIZE - off, len);
167 rval = (*f)(arg, kaddr + off, count);
168 if (rval)
169 return (rval);
170 len -= count;
171 processed += count;
172 off = 0;
173 pages++;
174 }
175 return (0);
176 }
177
178 static inline void *
cvm_page_contiguous_segment(vm_page_t * pages,size_t skip,int len)179 cvm_page_contiguous_segment(vm_page_t *pages, size_t skip, int len)
180 {
181 if ((skip + len - 1) / PAGE_SIZE > skip / PAGE_SIZE)
182 return (NULL);
183
184 pages += (skip / PAGE_SIZE);
185 skip -= rounddown(skip, PAGE_SIZE);
186 return (((char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages))) + skip);
187 }
188
189 /*
190 * Copy len bytes of data from the vm_page_t array, skipping the first off
191 * bytes, into the pointer cp. Return the number of bytes skipped and copied.
192 * Does not verify the length of the array.
193 */
194 static int
cvm_page_copyback(vm_page_t * pages,int off,int len,c_caddr_t cp)195 cvm_page_copyback(vm_page_t *pages, int off, int len, c_caddr_t cp)
196 {
197 int processed = 0;
198 unsigned count;
199
200 CVM_PAGE_SKIP();
201 while (len > 0) {
202 count = min(PAGE_SIZE - off, len);
203 bcopy(cp, (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages)) + off,
204 count);
205 len -= count;
206 cp += count;
207 processed += count;
208 off = 0;
209 pages++;
210 }
211 return (processed);
212 }
213
214 /*
215 * Copy len bytes of data from the pointer cp into the vm_page_t array,
216 * skipping the first off bytes, Return the number of bytes skipped and copied.
217 * Does not verify the length of the array.
218 */
219 static int
cvm_page_copydata(vm_page_t * pages,int off,int len,caddr_t cp)220 cvm_page_copydata(vm_page_t *pages, int off, int len, caddr_t cp)
221 {
222 int processed = 0;
223 unsigned count;
224
225 CVM_PAGE_SKIP();
226 while (len > 0) {
227 count = min(PAGE_SIZE - off, len);
228 bcopy(((char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(*pages)) + off), cp,
229 count);
230 len -= count;
231 cp += count;
232 processed += count;
233 off = 0;
234 pages++;
235 }
236 return processed;
237 }
238 #endif /* CRYPTO_MAY_HAVE_VMPAGE */
239
240 /*
241 * Given a starting page in an m_epg, determine the length of the
242 * current physically contiguous segment.
243 */
244 static __inline size_t
m_epg_pages_extent(struct mbuf * m,int idx,u_int pglen)245 m_epg_pages_extent(struct mbuf *m, int idx, u_int pglen)
246 {
247 size_t len;
248 u_int i;
249
250 len = pglen;
251 for (i = idx + 1; i < m->m_epg_npgs; i++) {
252 if (m->m_epg_pa[i - 1] + PAGE_SIZE != m->m_epg_pa[i])
253 break;
254 len += m_epg_pagelen(m, i, 0);
255 }
256 return (len);
257 }
258
259 static void *
m_epg_segment(struct mbuf * m,size_t offset,size_t * len)260 m_epg_segment(struct mbuf *m, size_t offset, size_t *len)
261 {
262 u_int i, pglen, pgoff;
263
264 offset += mtod(m, vm_offset_t);
265 if (offset < m->m_epg_hdrlen) {
266 *len = m->m_epg_hdrlen - offset;
267 return (m->m_epg_hdr + offset);
268 }
269 offset -= m->m_epg_hdrlen;
270 pgoff = m->m_epg_1st_off;
271 for (i = 0; i < m->m_epg_npgs; i++) {
272 pglen = m_epg_pagelen(m, i, pgoff);
273 if (offset < pglen) {
274 *len = m_epg_pages_extent(m, i, pglen) - offset;
275 return ((void *)PHYS_TO_DMAP(m->m_epg_pa[i] + pgoff +
276 offset));
277 }
278 offset -= pglen;
279 pgoff = 0;
280 }
281 KASSERT(offset <= m->m_epg_trllen, ("%s: offset beyond trailer",
282 __func__));
283 *len = m->m_epg_trllen - offset;
284 return (m->m_epg_trail + offset);
285 }
286
287 static __inline void *
m_epg_contiguous_subsegment(struct mbuf * m,size_t skip,size_t len)288 m_epg_contiguous_subsegment(struct mbuf *m, size_t skip, size_t len)
289 {
290 void *base;
291 size_t seglen;
292
293 base = m_epg_segment(m, skip, &seglen);
294 if (len > seglen)
295 return (NULL);
296 return (base);
297 }
298
299 void
crypto_cursor_init(struct crypto_buffer_cursor * cc,const struct crypto_buffer * cb)300 crypto_cursor_init(struct crypto_buffer_cursor *cc,
301 const struct crypto_buffer *cb)
302 {
303 memset(cc, 0, sizeof(*cc));
304 cc->cc_type = cb->cb_type;
305 switch (cc->cc_type) {
306 case CRYPTO_BUF_CONTIG:
307 cc->cc_buf = cb->cb_buf;
308 cc->cc_buf_len = cb->cb_buf_len;
309 break;
310 case CRYPTO_BUF_MBUF:
311 case CRYPTO_BUF_SINGLE_MBUF:
312 cc->cc_mbuf = cb->cb_mbuf;
313 break;
314 case CRYPTO_BUF_VMPAGE:
315 cc->cc_vmpage = cb->cb_vm_page;
316 cc->cc_buf_len = cb->cb_vm_page_len;
317 cc->cc_offset = cb->cb_vm_page_offset;
318 break;
319 case CRYPTO_BUF_UIO:
320 cc->cc_iov = cb->cb_uio->uio_iov;
321 cc->cc_buf_len = cb->cb_uio->uio_resid;
322 break;
323 default:
324 #ifdef INVARIANTS
325 panic("%s: invalid buffer type %d", __func__, cb->cb_type);
326 #endif
327 break;
328 }
329 }
330
331 SDT_PROBE_DEFINE2(opencrypto, criov, cursor_advance, vmpage, "struct crypto_buffer_cursor*", "size_t");
332
333 void
crypto_cursor_advance(struct crypto_buffer_cursor * cc,size_t amount)334 crypto_cursor_advance(struct crypto_buffer_cursor *cc, size_t amount)
335 {
336 size_t remain;
337
338 switch (cc->cc_type) {
339 case CRYPTO_BUF_CONTIG:
340 MPASS(cc->cc_buf_len >= amount);
341 cc->cc_buf += amount;
342 cc->cc_buf_len -= amount;
343 break;
344 case CRYPTO_BUF_MBUF:
345 for (;;) {
346 remain = cc->cc_mbuf->m_len - cc->cc_offset;
347 if (amount < remain) {
348 cc->cc_offset += amount;
349 break;
350 }
351 amount -= remain;
352 cc->cc_mbuf = cc->cc_mbuf->m_next;
353 cc->cc_offset = 0;
354 if (amount == 0)
355 break;
356 }
357 break;
358 case CRYPTO_BUF_SINGLE_MBUF:
359 MPASS(cc->cc_mbuf->m_len >= cc->cc_offset + amount);
360 cc->cc_offset += amount;
361 break;
362 case CRYPTO_BUF_VMPAGE:
363 for (;;) {
364 SDT_PROBE2(opencrypto, criov, cursor_advance, vmpage,
365 cc, amount);
366 remain = MIN(PAGE_SIZE - cc->cc_offset, cc->cc_buf_len);
367 if (amount < remain) {
368 cc->cc_buf_len -= amount;
369 cc->cc_offset += amount;
370 break;
371 }
372 cc->cc_buf_len -= remain;
373 amount -= remain;
374 cc->cc_vmpage++;
375 cc->cc_offset = 0;
376 if (amount == 0 || cc->cc_buf_len == 0)
377 break;
378 }
379 break;
380 case CRYPTO_BUF_UIO:
381 for (;;) {
382 remain = cc->cc_iov->iov_len - cc->cc_offset;
383 if (amount < remain) {
384 cc->cc_offset += amount;
385 break;
386 }
387 cc->cc_buf_len -= remain;
388 amount -= remain;
389 cc->cc_iov++;
390 cc->cc_offset = 0;
391 if (amount == 0)
392 break;
393 }
394 break;
395 default:
396 #ifdef INVARIANTS
397 panic("%s: invalid buffer type %d", __func__, cc->cc_type);
398 #endif
399 break;
400 }
401 }
402
403 void *
crypto_cursor_segment(struct crypto_buffer_cursor * cc,size_t * len)404 crypto_cursor_segment(struct crypto_buffer_cursor *cc, size_t *len)
405 {
406 switch (cc->cc_type) {
407 case CRYPTO_BUF_CONTIG:
408 case CRYPTO_BUF_UIO:
409 case CRYPTO_BUF_VMPAGE:
410 if (cc->cc_buf_len == 0) {
411 *len = 0;
412 return (NULL);
413 }
414 break;
415 case CRYPTO_BUF_MBUF:
416 case CRYPTO_BUF_SINGLE_MBUF:
417 if (cc->cc_mbuf == NULL) {
418 *len = 0;
419 return (NULL);
420 }
421 break;
422 default:
423 #ifdef INVARIANTS
424 panic("%s: invalid buffer type %d", __func__, cc->cc_type);
425 #endif
426 *len = 0;
427 return (NULL);
428 }
429
430 switch (cc->cc_type) {
431 case CRYPTO_BUF_CONTIG:
432 *len = cc->cc_buf_len;
433 return (cc->cc_buf);
434 case CRYPTO_BUF_MBUF:
435 case CRYPTO_BUF_SINGLE_MBUF:
436 if (cc->cc_mbuf->m_flags & M_EXTPG)
437 return (m_epg_segment(cc->cc_mbuf, cc->cc_offset, len));
438 *len = cc->cc_mbuf->m_len - cc->cc_offset;
439 return (mtod(cc->cc_mbuf, char *) + cc->cc_offset);
440 case CRYPTO_BUF_VMPAGE:
441 *len = PAGE_SIZE - cc->cc_offset;
442 return ((char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(
443 *cc->cc_vmpage)) + cc->cc_offset);
444 case CRYPTO_BUF_UIO:
445 *len = cc->cc_iov->iov_len - cc->cc_offset;
446 return ((char *)cc->cc_iov->iov_base + cc->cc_offset);
447 default:
448 __assert_unreachable();
449 }
450 }
451
452 void
crypto_cursor_copyback(struct crypto_buffer_cursor * cc,int size,const void * vsrc)453 crypto_cursor_copyback(struct crypto_buffer_cursor *cc, int size,
454 const void *vsrc)
455 {
456 size_t remain, todo;
457 const char *src;
458 char *dst;
459
460 src = vsrc;
461 switch (cc->cc_type) {
462 case CRYPTO_BUF_CONTIG:
463 MPASS(cc->cc_buf_len >= size);
464 memcpy(cc->cc_buf, src, size);
465 cc->cc_buf += size;
466 cc->cc_buf_len -= size;
467 break;
468 case CRYPTO_BUF_MBUF:
469 for (;;) {
470 /*
471 * This uses m_copyback() for individual
472 * mbufs so that cc_mbuf and cc_offset are
473 * updated.
474 */
475 remain = cc->cc_mbuf->m_len - cc->cc_offset;
476 todo = MIN(remain, size);
477 m_copyback(cc->cc_mbuf, cc->cc_offset, todo, src);
478 src += todo;
479 if (todo < remain) {
480 cc->cc_offset += todo;
481 break;
482 }
483 size -= todo;
484 cc->cc_mbuf = cc->cc_mbuf->m_next;
485 cc->cc_offset = 0;
486 if (size == 0)
487 break;
488 }
489 break;
490 case CRYPTO_BUF_SINGLE_MBUF:
491 MPASS(cc->cc_mbuf->m_len >= cc->cc_offset + size);
492 m_copyback(cc->cc_mbuf, cc->cc_offset, size, src);
493 cc->cc_offset += size;
494 break;
495 case CRYPTO_BUF_VMPAGE:
496 for (;;) {
497 dst = (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(
498 *cc->cc_vmpage)) + cc->cc_offset;
499 remain = MIN(PAGE_SIZE - cc->cc_offset, cc->cc_buf_len);
500 todo = MIN(remain, size);
501 memcpy(dst, src, todo);
502 src += todo;
503 cc->cc_buf_len -= todo;
504 if (todo < remain) {
505 cc->cc_offset += todo;
506 break;
507 }
508 size -= todo;
509 cc->cc_vmpage++;
510 cc->cc_offset = 0;
511 if (size == 0)
512 break;
513 }
514 break;
515 case CRYPTO_BUF_UIO:
516 for (;;) {
517 dst = (char *)cc->cc_iov->iov_base + cc->cc_offset;
518 remain = cc->cc_iov->iov_len - cc->cc_offset;
519 todo = MIN(remain, size);
520 memcpy(dst, src, todo);
521 src += todo;
522 cc->cc_buf_len -= todo;
523 if (todo < remain) {
524 cc->cc_offset += todo;
525 break;
526 }
527 size -= todo;
528 cc->cc_iov++;
529 cc->cc_offset = 0;
530 if (size == 0)
531 break;
532 }
533 break;
534 default:
535 #ifdef INVARIANTS
536 panic("%s: invalid buffer type %d", __func__, cc->cc_type);
537 #endif
538 break;
539 }
540 }
541
542 void
crypto_cursor_copydata(struct crypto_buffer_cursor * cc,int size,void * vdst)543 crypto_cursor_copydata(struct crypto_buffer_cursor *cc, int size, void *vdst)
544 {
545 size_t remain, todo;
546 const char *src;
547 char *dst;
548
549 dst = vdst;
550 switch (cc->cc_type) {
551 case CRYPTO_BUF_CONTIG:
552 MPASS(cc->cc_buf_len >= size);
553 memcpy(dst, cc->cc_buf, size);
554 cc->cc_buf += size;
555 cc->cc_buf_len -= size;
556 break;
557 case CRYPTO_BUF_MBUF:
558 for (;;) {
559 /*
560 * This uses m_copydata() for individual
561 * mbufs so that cc_mbuf and cc_offset are
562 * updated.
563 */
564 remain = cc->cc_mbuf->m_len - cc->cc_offset;
565 todo = MIN(remain, size);
566 m_copydata(cc->cc_mbuf, cc->cc_offset, todo, dst);
567 dst += todo;
568 if (todo < remain) {
569 cc->cc_offset += todo;
570 break;
571 }
572 size -= todo;
573 cc->cc_mbuf = cc->cc_mbuf->m_next;
574 cc->cc_offset = 0;
575 if (size == 0)
576 break;
577 }
578 break;
579 case CRYPTO_BUF_SINGLE_MBUF:
580 MPASS(cc->cc_mbuf->m_len >= cc->cc_offset + size);
581 m_copydata(cc->cc_mbuf, cc->cc_offset, size, dst);
582 cc->cc_offset += size;
583 break;
584 case CRYPTO_BUF_VMPAGE:
585 for (;;) {
586 src = (char *)PHYS_TO_DMAP(VM_PAGE_TO_PHYS(
587 *cc->cc_vmpage)) + cc->cc_offset;
588 remain = MIN(PAGE_SIZE - cc->cc_offset, cc->cc_buf_len);
589 todo = MIN(remain, size);
590 memcpy(dst, src, todo);
591 dst += todo;
592 cc->cc_buf_len -= todo;
593 if (todo < remain) {
594 cc->cc_offset += todo;
595 break;
596 }
597 size -= todo;
598 cc->cc_vmpage++;
599 cc->cc_offset = 0;
600 if (size == 0)
601 break;
602 }
603 break;
604 case CRYPTO_BUF_UIO:
605 for (;;) {
606 src = (const char *)cc->cc_iov->iov_base +
607 cc->cc_offset;
608 remain = cc->cc_iov->iov_len - cc->cc_offset;
609 todo = MIN(remain, size);
610 memcpy(dst, src, todo);
611 dst += todo;
612 cc->cc_buf_len -= todo;
613 if (todo < remain) {
614 cc->cc_offset += todo;
615 break;
616 }
617 size -= todo;
618 cc->cc_iov++;
619 cc->cc_offset = 0;
620 if (size == 0)
621 break;
622 }
623 break;
624 default:
625 #ifdef INVARIANTS
626 panic("%s: invalid buffer type %d", __func__, cc->cc_type);
627 #endif
628 break;
629 }
630 }
631
632 /*
633 * To avoid advancing 'cursor', make a local copy that gets advanced
634 * instead.
635 */
636 void
crypto_cursor_copydata_noadv(struct crypto_buffer_cursor * cc,int size,void * vdst)637 crypto_cursor_copydata_noadv(struct crypto_buffer_cursor *cc, int size,
638 void *vdst)
639 {
640 struct crypto_buffer_cursor copy;
641
642 copy = *cc;
643 crypto_cursor_copydata(©, size, vdst);
644 }
645
646 /*
647 * Apply function f to the data in an iovec list starting "off" bytes from
648 * the beginning, continuing for "len" bytes.
649 */
650 static int
cuio_apply(struct uio * uio,int off,int len,int (* f)(void *,const void *,u_int),void * arg)651 cuio_apply(struct uio *uio, int off, int len,
652 int (*f)(void *, const void *, u_int), void *arg)
653 {
654 struct iovec *iov = uio->uio_iov;
655 int iol __diagused = uio->uio_iovcnt;
656 unsigned count;
657 int rval;
658
659 CUIO_SKIP();
660 while (len > 0) {
661 KASSERT(iol >= 0, ("%s: empty", __func__));
662 count = min(iov->iov_len - off, len);
663 rval = (*f)(arg, ((caddr_t)iov->iov_base) + off, count);
664 if (rval)
665 return (rval);
666 len -= count;
667 off = 0;
668 iol--;
669 iov++;
670 }
671 return (0);
672 }
673
674 void
crypto_copyback(struct cryptop * crp,int off,int size,const void * src)675 crypto_copyback(struct cryptop *crp, int off, int size, const void *src)
676 {
677 struct crypto_buffer *cb;
678
679 if (crp->crp_obuf.cb_type != CRYPTO_BUF_NONE)
680 cb = &crp->crp_obuf;
681 else
682 cb = &crp->crp_buf;
683 switch (cb->cb_type) {
684 case CRYPTO_BUF_MBUF:
685 case CRYPTO_BUF_SINGLE_MBUF:
686 m_copyback(cb->cb_mbuf, off, size, src);
687 break;
688 #if CRYPTO_MAY_HAVE_VMPAGE
689 case CRYPTO_BUF_VMPAGE:
690 MPASS(size <= cb->cb_vm_page_len);
691 MPASS(size + off <=
692 cb->cb_vm_page_len + cb->cb_vm_page_offset);
693 cvm_page_copyback(cb->cb_vm_page,
694 off + cb->cb_vm_page_offset, size, src);
695 break;
696 #endif /* CRYPTO_MAY_HAVE_VMPAGE */
697 case CRYPTO_BUF_UIO:
698 cuio_copyback(cb->cb_uio, off, size, src);
699 break;
700 case CRYPTO_BUF_CONTIG:
701 MPASS(off + size <= cb->cb_buf_len);
702 bcopy(src, cb->cb_buf + off, size);
703 break;
704 default:
705 #ifdef INVARIANTS
706 panic("invalid crp buf type %d", cb->cb_type);
707 #endif
708 break;
709 }
710 }
711
712 void
crypto_copydata(struct cryptop * crp,int off,int size,void * dst)713 crypto_copydata(struct cryptop *crp, int off, int size, void *dst)
714 {
715
716 switch (crp->crp_buf.cb_type) {
717 case CRYPTO_BUF_MBUF:
718 case CRYPTO_BUF_SINGLE_MBUF:
719 m_copydata(crp->crp_buf.cb_mbuf, off, size, dst);
720 break;
721 #if CRYPTO_MAY_HAVE_VMPAGE
722 case CRYPTO_BUF_VMPAGE:
723 MPASS(size <= crp->crp_buf.cb_vm_page_len);
724 MPASS(size + off <= crp->crp_buf.cb_vm_page_len +
725 crp->crp_buf.cb_vm_page_offset);
726 cvm_page_copydata(crp->crp_buf.cb_vm_page,
727 off + crp->crp_buf.cb_vm_page_offset, size, dst);
728 break;
729 #endif /* CRYPTO_MAY_HAVE_VMPAGE */
730 case CRYPTO_BUF_UIO:
731 cuio_copydata(crp->crp_buf.cb_uio, off, size, dst);
732 break;
733 case CRYPTO_BUF_CONTIG:
734 MPASS(off + size <= crp->crp_buf.cb_buf_len);
735 bcopy(crp->crp_buf.cb_buf + off, dst, size);
736 break;
737 default:
738 #ifdef INVARIANTS
739 panic("invalid crp buf type %d", crp->crp_buf.cb_type);
740 #endif
741 break;
742 }
743 }
744
745 int
crypto_apply_buf(struct crypto_buffer * cb,int off,int len,int (* f)(void *,const void *,u_int),void * arg)746 crypto_apply_buf(struct crypto_buffer *cb, int off, int len,
747 int (*f)(void *, const void *, u_int), void *arg)
748 {
749 int error;
750
751 switch (cb->cb_type) {
752 case CRYPTO_BUF_MBUF:
753 case CRYPTO_BUF_SINGLE_MBUF:
754 error = m_apply(cb->cb_mbuf, off, len,
755 (int (*)(void *, void *, u_int))f, arg);
756 break;
757 case CRYPTO_BUF_UIO:
758 error = cuio_apply(cb->cb_uio, off, len, f, arg);
759 break;
760 #if CRYPTO_MAY_HAVE_VMPAGE
761 case CRYPTO_BUF_VMPAGE:
762 error = cvm_page_apply(cb->cb_vm_page,
763 off + cb->cb_vm_page_offset, len, f, arg);
764 break;
765 #endif /* CRYPTO_MAY_HAVE_VMPAGE */
766 case CRYPTO_BUF_CONTIG:
767 MPASS(off + len <= cb->cb_buf_len);
768 error = (*f)(arg, cb->cb_buf + off, len);
769 break;
770 default:
771 #ifdef INVARIANTS
772 panic("invalid crypto buf type %d", cb->cb_type);
773 #endif
774 error = 0;
775 break;
776 }
777 return (error);
778 }
779
780 int
crypto_apply(struct cryptop * crp,int off,int len,int (* f)(void *,const void *,u_int),void * arg)781 crypto_apply(struct cryptop *crp, int off, int len,
782 int (*f)(void *, const void *, u_int), void *arg)
783 {
784 return (crypto_apply_buf(&crp->crp_buf, off, len, f, arg));
785 }
786
787 static inline void *
m_contiguous_subsegment(struct mbuf * m,size_t skip,size_t len)788 m_contiguous_subsegment(struct mbuf *m, size_t skip, size_t len)
789 {
790 int rel_off;
791
792 MPASS(skip <= INT_MAX);
793
794 m = m_getptr(m, (int)skip, &rel_off);
795 if (m == NULL)
796 return (NULL);
797
798 MPASS(rel_off >= 0);
799 skip = rel_off;
800 if (skip + len > m->m_len)
801 return (NULL);
802
803 if (m->m_flags & M_EXTPG)
804 return (m_epg_contiguous_subsegment(m, skip, len));
805 return (mtod(m, char*) + skip);
806 }
807
808 static inline void *
cuio_contiguous_segment(struct uio * uio,size_t skip,size_t len)809 cuio_contiguous_segment(struct uio *uio, size_t skip, size_t len)
810 {
811 int rel_off, idx;
812
813 MPASS(skip <= INT_MAX);
814 idx = cuio_getptr(uio, (int)skip, &rel_off);
815 if (idx < 0)
816 return (NULL);
817
818 MPASS(rel_off >= 0);
819 skip = rel_off;
820 if (skip + len > uio->uio_iov[idx].iov_len)
821 return (NULL);
822 return ((char *)uio->uio_iov[idx].iov_base + skip);
823 }
824
825 void *
crypto_buffer_contiguous_subsegment(struct crypto_buffer * cb,size_t skip,size_t len)826 crypto_buffer_contiguous_subsegment(struct crypto_buffer *cb, size_t skip,
827 size_t len)
828 {
829
830 switch (cb->cb_type) {
831 case CRYPTO_BUF_MBUF:
832 case CRYPTO_BUF_SINGLE_MBUF:
833 return (m_contiguous_subsegment(cb->cb_mbuf, skip, len));
834 case CRYPTO_BUF_UIO:
835 return (cuio_contiguous_segment(cb->cb_uio, skip, len));
836 #if CRYPTO_MAY_HAVE_VMPAGE
837 case CRYPTO_BUF_VMPAGE:
838 MPASS(skip + len <= cb->cb_vm_page_len);
839 return (cvm_page_contiguous_segment(cb->cb_vm_page,
840 skip + cb->cb_vm_page_offset, len));
841 #endif /* CRYPTO_MAY_HAVE_VMPAGE */
842 case CRYPTO_BUF_CONTIG:
843 MPASS(skip + len <= cb->cb_buf_len);
844 return (cb->cb_buf + skip);
845 default:
846 #ifdef INVARIANTS
847 panic("invalid crp buf type %d", cb->cb_type);
848 #endif
849 return (NULL);
850 }
851 }
852
853 void *
crypto_contiguous_subsegment(struct cryptop * crp,size_t skip,size_t len)854 crypto_contiguous_subsegment(struct cryptop *crp, size_t skip, size_t len)
855 {
856 return (crypto_buffer_contiguous_subsegment(&crp->crp_buf, skip, len));
857 }
858