16810ad6fSSam Leffler /*- 26810ad6fSSam Leffler * Copyright (c) 2002-2006 Sam Leffler. All rights reserved. 36810ad6fSSam Leffler * 46810ad6fSSam Leffler * Redistribution and use in source and binary forms, with or without 56810ad6fSSam Leffler * modification, are permitted provided that the following conditions 66810ad6fSSam Leffler * are met: 76810ad6fSSam Leffler * 1. Redistributions of source code must retain the above copyright 86810ad6fSSam Leffler * notice, this list of conditions and the following disclaimer. 96810ad6fSSam Leffler * 2. Redistributions in binary form must reproduce the above copyright 106810ad6fSSam Leffler * notice, this list of conditions and the following disclaimer in the 116810ad6fSSam Leffler * documentation and/or other materials provided with the distribution. 126810ad6fSSam Leffler * 136810ad6fSSam Leffler * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 146810ad6fSSam Leffler * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 156810ad6fSSam Leffler * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 166810ad6fSSam Leffler * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 176810ad6fSSam Leffler * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 186810ad6fSSam Leffler * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 196810ad6fSSam Leffler * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 206810ad6fSSam Leffler * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 216810ad6fSSam Leffler * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 226810ad6fSSam Leffler * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 236810ad6fSSam Leffler */ 246810ad6fSSam Leffler 256810ad6fSSam Leffler #include <sys/cdefs.h> 266810ad6fSSam Leffler __FBSDID("$FreeBSD$"); 276810ad6fSSam Leffler 286810ad6fSSam Leffler /* 296810ad6fSSam Leffler * Cryptographic Subsystem. 306810ad6fSSam Leffler * 316810ad6fSSam Leffler * This code is derived from the Openbsd Cryptographic Framework (OCF) 326810ad6fSSam Leffler * that has the copyright shown below. Very little of the original 336810ad6fSSam Leffler * code remains. 346810ad6fSSam Leffler */ 356810ad6fSSam Leffler 3660727d8bSWarner Losh /*- 37091d81d1SSam Leffler * The author of this code is Angelos D. Keromytis (angelos@cis.upenn.edu) 38091d81d1SSam Leffler * 39091d81d1SSam Leffler * This code was written by Angelos D. Keromytis in Athens, Greece, in 40091d81d1SSam Leffler * February 2000. Network Security Technologies Inc. (NSTI) kindly 41091d81d1SSam Leffler * supported the development of this code. 42091d81d1SSam Leffler * 43091d81d1SSam Leffler * Copyright (c) 2000, 2001 Angelos D. Keromytis 44091d81d1SSam Leffler * 45091d81d1SSam Leffler * Permission to use, copy, and modify this software with or without fee 46091d81d1SSam Leffler * is hereby granted, provided that this entire notice is included in 47091d81d1SSam Leffler * all source code copies of any software which is or includes a copy or 48091d81d1SSam Leffler * modification of this software. 49091d81d1SSam Leffler * 50091d81d1SSam Leffler * THIS SOFTWARE IS BEING PROVIDED "AS IS", WITHOUT ANY EXPRESS OR 51091d81d1SSam Leffler * IMPLIED WARRANTY. IN PARTICULAR, NONE OF THE AUTHORS MAKES ANY 52091d81d1SSam Leffler * REPRESENTATION OR WARRANTY OF ANY KIND CONCERNING THE 53091d81d1SSam Leffler * MERCHANTABILITY OF THIS SOFTWARE OR ITS FITNESS FOR ANY PARTICULAR 54091d81d1SSam Leffler * PURPOSE. 55091d81d1SSam Leffler */ 562c446514SDavid E. O'Brien 57c0341432SJohn Baldwin #include "opt_compat.h" 586810ad6fSSam Leffler #include "opt_ddb.h" 596810ad6fSSam Leffler 60091d81d1SSam Leffler #include <sys/param.h> 61091d81d1SSam Leffler #include <sys/systm.h> 627290cb47SMark Johnston #include <sys/counter.h> 63091d81d1SSam Leffler #include <sys/kernel.h> 64091d81d1SSam Leffler #include <sys/kthread.h> 65ec5c0e5bSAllan Jude #include <sys/linker.h> 66091d81d1SSam Leffler #include <sys/lock.h> 675dba30f1SPoul-Henning Kamp #include <sys/module.h> 68091d81d1SSam Leffler #include <sys/mutex.h> 69091d81d1SSam Leffler #include <sys/malloc.h> 709c0e3d3aSJohn Baldwin #include <sys/mbuf.h> 71091d81d1SSam Leffler #include <sys/proc.h> 72c0341432SJohn Baldwin #include <sys/refcount.h> 73df21ad6eSBjoern A. Zeeb #include <sys/sdt.h> 7439bbca6fSFabien Thomas #include <sys/smp.h> 75091d81d1SSam Leffler #include <sys/sysctl.h> 7639bbca6fSFabien Thomas #include <sys/taskqueue.h> 779c0e3d3aSJohn Baldwin #include <sys/uio.h> 78091d81d1SSam Leffler 796810ad6fSSam Leffler #include <ddb/ddb.h> 806810ad6fSSam Leffler 81e6f6d0c9SAlan Somers #include <machine/vmparam.h> 82091d81d1SSam Leffler #include <vm/uma.h> 83e6f6d0c9SAlan Somers 84ec5c0e5bSAllan Jude #include <crypto/intake.h> 85091d81d1SSam Leffler #include <opencrypto/cryptodev.h> 86c0341432SJohn Baldwin #include <opencrypto/xform_auth.h> 87c0341432SJohn Baldwin #include <opencrypto/xform_enc.h> 88091d81d1SSam Leffler 896810ad6fSSam Leffler #include <sys/kobj.h> 906810ad6fSSam Leffler #include <sys/bus.h> 916810ad6fSSam Leffler #include "cryptodev_if.h" 926810ad6fSSam Leffler 936ed982a2SAndrew Turner #if defined(__i386__) || defined(__amd64__) || defined(__aarch64__) 9404c49e68SKonstantin Belousov #include <machine/pcb.h> 9504c49e68SKonstantin Belousov #endif 9604c49e68SKonstantin Belousov 97df21ad6eSBjoern A. Zeeb SDT_PROVIDER_DEFINE(opencrypto); 98df21ad6eSBjoern A. Zeeb 99091d81d1SSam Leffler /* 100091d81d1SSam Leffler * Crypto drivers register themselves by allocating a slot in the 101091d81d1SSam Leffler * crypto_drivers table with crypto_get_driverid() and then registering 102c0341432SJohn Baldwin * each asym algorithm they support with crypto_kregister(). 103091d81d1SSam Leffler */ 104091d81d1SSam Leffler static struct mtx crypto_drivers_mtx; /* lock on driver table */ 105091d81d1SSam Leffler #define CRYPTO_DRIVER_LOCK() mtx_lock(&crypto_drivers_mtx) 106091d81d1SSam Leffler #define CRYPTO_DRIVER_UNLOCK() mtx_unlock(&crypto_drivers_mtx) 1076810ad6fSSam Leffler #define CRYPTO_DRIVER_ASSERT() mtx_assert(&crypto_drivers_mtx, MA_OWNED) 1086810ad6fSSam Leffler 1096810ad6fSSam Leffler /* 1106810ad6fSSam Leffler * Crypto device/driver capabilities structure. 1116810ad6fSSam Leffler * 1126810ad6fSSam Leffler * Synchronization: 1136810ad6fSSam Leffler * (d) - protected by CRYPTO_DRIVER_LOCK() 1146810ad6fSSam Leffler * (q) - protected by CRYPTO_Q_LOCK() 1156810ad6fSSam Leffler * Not tagged fields are read-only. 1166810ad6fSSam Leffler */ 1176810ad6fSSam Leffler struct cryptocap { 118c0341432SJohn Baldwin device_t cc_dev; 119c0341432SJohn Baldwin uint32_t cc_hid; 120d3d79e96SJohn Baldwin uint32_t cc_sessions; /* (d) # of sessions */ 121d3d79e96SJohn Baldwin uint32_t cc_koperations; /* (d) # os asym operations */ 122d3d79e96SJohn Baldwin uint8_t cc_kalg[CRK_ALGORITHM_MAX + 1]; 1236810ad6fSSam Leffler 1246810ad6fSSam Leffler int cc_flags; /* (d) flags */ 1256810ad6fSSam Leffler #define CRYPTOCAP_F_CLEANUP 0x80000000 /* needs resource cleanup */ 1266810ad6fSSam Leffler int cc_qblocked; /* (q) symmetric q blocked */ 1276810ad6fSSam Leffler int cc_kqblocked; /* (q) asymmetric q blocked */ 1281b0909d5SConrad Meyer size_t cc_session_size; 129c0341432SJohn Baldwin volatile int cc_refs; 1306810ad6fSSam Leffler }; 131c0341432SJohn Baldwin 132c0341432SJohn Baldwin static struct cryptocap **crypto_drivers = NULL; 133c0341432SJohn Baldwin static int crypto_drivers_size = 0; 134c0341432SJohn Baldwin 135c0341432SJohn Baldwin struct crypto_session { 136c0341432SJohn Baldwin struct cryptocap *cap; 137c0341432SJohn Baldwin struct crypto_session_params csp; 13898d788c8SMark Johnston uint64_t id; 1398adcc757SMark Johnston /* Driver softc follows. */ 140c0341432SJohn Baldwin }; 141091d81d1SSam Leffler 142091d81d1SSam Leffler /* 143091d81d1SSam Leffler * There are two queues for crypto requests; one for symmetric (e.g. 144091d81d1SSam Leffler * cipher) operations and one for asymmetric (e.g. MOD)operations. 145091d81d1SSam Leffler * A single mutex is used to lock access to both queues. We could 146091d81d1SSam Leffler * have one per-queue but having one simplifies handling of block/unblock 147091d81d1SSam Leffler * operations. 148091d81d1SSam Leffler */ 1493a865c82SPawel Jakub Dawidek static int crp_sleep = 0; 15039bbca6fSFabien Thomas static TAILQ_HEAD(cryptop_q ,cryptop) crp_q; /* request queues */ 151091d81d1SSam Leffler static TAILQ_HEAD(,cryptkop) crp_kq; 152091d81d1SSam Leffler static struct mtx crypto_q_mtx; 153091d81d1SSam Leffler #define CRYPTO_Q_LOCK() mtx_lock(&crypto_q_mtx) 154091d81d1SSam Leffler #define CRYPTO_Q_UNLOCK() mtx_unlock(&crypto_q_mtx) 155091d81d1SSam Leffler 15633f3bad3SJohn Baldwin SYSCTL_NODE(_kern, OID_AUTO, crypto, CTLFLAG_RW, 0, 157c0341432SJohn Baldwin "In-kernel cryptography"); 158c0341432SJohn Baldwin 159091d81d1SSam Leffler /* 16039bbca6fSFabien Thomas * Taskqueue used to dispatch the crypto requests 16139bbca6fSFabien Thomas * that have the CRYPTO_F_ASYNC flag 162091d81d1SSam Leffler */ 16339bbca6fSFabien Thomas static struct taskqueue *crypto_tq; 16439bbca6fSFabien Thomas 16539bbca6fSFabien Thomas /* 16639bbca6fSFabien Thomas * Crypto seq numbers are operated on with modular arithmetic 16739bbca6fSFabien Thomas */ 16839bbca6fSFabien Thomas #define CRYPTO_SEQ_GT(a,b) ((int)((a)-(b)) > 0) 16939bbca6fSFabien Thomas 17039bbca6fSFabien Thomas struct crypto_ret_worker { 17139bbca6fSFabien Thomas struct mtx crypto_ret_mtx; 17239bbca6fSFabien Thomas 17339bbca6fSFabien Thomas TAILQ_HEAD(,cryptop) crp_ordered_ret_q; /* ordered callback queue for symetric jobs */ 17439bbca6fSFabien Thomas TAILQ_HEAD(,cryptop) crp_ret_q; /* callback queue for symetric jobs */ 17539bbca6fSFabien Thomas TAILQ_HEAD(,cryptkop) crp_ret_kq; /* callback queue for asym jobs */ 17639bbca6fSFabien Thomas 177d3d79e96SJohn Baldwin uint32_t reorder_ops; /* total ordered sym jobs received */ 178d3d79e96SJohn Baldwin uint32_t reorder_cur_seq; /* current sym job dispatched */ 17939bbca6fSFabien Thomas 18039bbca6fSFabien Thomas struct proc *cryptoretproc; 18139bbca6fSFabien Thomas }; 18239bbca6fSFabien Thomas static struct crypto_ret_worker *crypto_ret_workers = NULL; 18339bbca6fSFabien Thomas 18439bbca6fSFabien Thomas #define CRYPTO_RETW(i) (&crypto_ret_workers[i]) 18539bbca6fSFabien Thomas #define CRYPTO_RETW_ID(w) ((w) - crypto_ret_workers) 18639bbca6fSFabien Thomas #define FOREACH_CRYPTO_RETW(w) \ 18739bbca6fSFabien Thomas for (w = crypto_ret_workers; w < crypto_ret_workers + crypto_workers_num; ++w) 18839bbca6fSFabien Thomas 18939bbca6fSFabien Thomas #define CRYPTO_RETW_LOCK(w) mtx_lock(&w->crypto_ret_mtx) 19039bbca6fSFabien Thomas #define CRYPTO_RETW_UNLOCK(w) mtx_unlock(&w->crypto_ret_mtx) 19139bbca6fSFabien Thomas 19239bbca6fSFabien Thomas static int crypto_workers_num = 0; 193c0341432SJohn Baldwin SYSCTL_INT(_kern_crypto, OID_AUTO, num_workers, CTLFLAG_RDTUN, 194c0341432SJohn Baldwin &crypto_workers_num, 0, 195c0341432SJohn Baldwin "Number of crypto workers used to dispatch crypto jobs"); 196c0341432SJohn Baldwin #ifdef COMPAT_FREEBSD12 19739bbca6fSFabien Thomas SYSCTL_INT(_kern, OID_AUTO, crypto_workers_num, CTLFLAG_RDTUN, 19839bbca6fSFabien Thomas &crypto_workers_num, 0, 19939bbca6fSFabien Thomas "Number of crypto workers used to dispatch crypto jobs"); 200c0341432SJohn Baldwin #endif 201091d81d1SSam Leffler 202091d81d1SSam Leffler static uma_zone_t cryptop_zone; 203091d81d1SSam Leffler 204c0341432SJohn Baldwin int crypto_userasymcrypto = 1; 205c0341432SJohn Baldwin SYSCTL_INT(_kern_crypto, OID_AUTO, asym_enable, CTLFLAG_RW, 206c0341432SJohn Baldwin &crypto_userasymcrypto, 0, 207c0341432SJohn Baldwin "Enable user-mode access to asymmetric crypto support"); 208c0341432SJohn Baldwin #ifdef COMPAT_FREEBSD12 209091d81d1SSam Leffler SYSCTL_INT(_kern, OID_AUTO, userasymcrypto, CTLFLAG_RW, 210091d81d1SSam Leffler &crypto_userasymcrypto, 0, 211091d81d1SSam Leffler "Enable/disable user-mode access to asymmetric crypto support"); 212c0341432SJohn Baldwin #endif 213c0341432SJohn Baldwin 214c0341432SJohn Baldwin int crypto_devallowsoft = 0; 215c0341432SJohn Baldwin SYSCTL_INT(_kern_crypto, OID_AUTO, allow_soft, CTLFLAG_RW, 216c0341432SJohn Baldwin &crypto_devallowsoft, 0, 217c0341432SJohn Baldwin "Enable use of software crypto by /dev/crypto"); 218c0341432SJohn Baldwin #ifdef COMPAT_FREEBSD12 219091d81d1SSam Leffler SYSCTL_INT(_kern, OID_AUTO, cryptodevallowsoft, CTLFLAG_RW, 220091d81d1SSam Leffler &crypto_devallowsoft, 0, 2216c20d7a3SJohn-Mark Gurney "Enable/disable use of software crypto by /dev/crypto"); 222c0341432SJohn Baldwin #endif 223091d81d1SSam Leffler 224091d81d1SSam Leffler MALLOC_DEFINE(M_CRYPTO_DATA, "crypto", "crypto session records"); 225091d81d1SSam Leffler 22651e45326SSam Leffler static void crypto_proc(void); 22751e45326SSam Leffler static struct proc *cryptoproc; 22839bbca6fSFabien Thomas static void crypto_ret_proc(struct crypto_ret_worker *ret_worker); 22951e45326SSam Leffler static void crypto_destroy(void); 2304acae0acSPawel Jakub Dawidek static int crypto_invoke(struct cryptocap *cap, struct cryptop *crp, int hint); 231c0341432SJohn Baldwin static int crypto_kinvoke(struct cryptkop *krp); 23239bbca6fSFabien Thomas static void crypto_task_invoke(void *ctx, int pending); 23339bbca6fSFabien Thomas static void crypto_batch_enqueue(struct cryptop *crp); 23451e45326SSam Leffler 2357290cb47SMark Johnston static counter_u64_t cryptostats[sizeof(struct cryptostats) / sizeof(uint64_t)]; 2367290cb47SMark Johnston SYSCTL_COUNTER_U64_ARRAY(_kern_crypto, OID_AUTO, stats, CTLFLAG_RW, 2377290cb47SMark Johnston cryptostats, nitems(cryptostats), 2387290cb47SMark Johnston "Crypto system statistics"); 2397290cb47SMark Johnston 2407290cb47SMark Johnston #define CRYPTOSTAT_INC(stat) do { \ 2417290cb47SMark Johnston counter_u64_add( \ 2427290cb47SMark Johnston cryptostats[offsetof(struct cryptostats, stat) / sizeof(uint64_t)],\ 2437290cb47SMark Johnston 1); \ 2447290cb47SMark Johnston } while (0) 2457290cb47SMark Johnston 2467290cb47SMark Johnston static void 2477290cb47SMark Johnston cryptostats_init(void *arg __unused) 2487290cb47SMark Johnston { 2497290cb47SMark Johnston COUNTER_ARRAY_ALLOC(cryptostats, nitems(cryptostats), M_WAITOK); 2507290cb47SMark Johnston } 2517290cb47SMark Johnston SYSINIT(cryptostats_init, SI_SUB_COUNTER, SI_ORDER_ANY, cryptostats_init, NULL); 2527290cb47SMark Johnston 2537290cb47SMark Johnston static void 2547290cb47SMark Johnston cryptostats_fini(void *arg __unused) 2557290cb47SMark Johnston { 2567290cb47SMark Johnston COUNTER_ARRAY_FREE(cryptostats, nitems(cryptostats)); 2577290cb47SMark Johnston } 2587290cb47SMark Johnston SYSUNINIT(cryptostats_fini, SI_SUB_COUNTER, SI_ORDER_ANY, cryptostats_fini, 2597290cb47SMark Johnston NULL); 2607d1853eeSSam Leffler 261ec5c0e5bSAllan Jude /* Try to avoid directly exposing the key buffer as a symbol */ 262ec5c0e5bSAllan Jude static struct keybuf *keybuf; 263ec5c0e5bSAllan Jude 264ec5c0e5bSAllan Jude static struct keybuf empty_keybuf = { 265ec5c0e5bSAllan Jude .kb_nents = 0 266ec5c0e5bSAllan Jude }; 267ec5c0e5bSAllan Jude 268ec5c0e5bSAllan Jude /* Obtain the key buffer from boot metadata */ 269ec5c0e5bSAllan Jude static void 270ec5c0e5bSAllan Jude keybuf_init(void) 271ec5c0e5bSAllan Jude { 272ec5c0e5bSAllan Jude caddr_t kmdp; 273ec5c0e5bSAllan Jude 274ec5c0e5bSAllan Jude kmdp = preload_search_by_type("elf kernel"); 275ec5c0e5bSAllan Jude 276ec5c0e5bSAllan Jude if (kmdp == NULL) 277ec5c0e5bSAllan Jude kmdp = preload_search_by_type("elf64 kernel"); 278ec5c0e5bSAllan Jude 279ec5c0e5bSAllan Jude keybuf = (struct keybuf *)preload_search_info(kmdp, 280ec5c0e5bSAllan Jude MODINFO_METADATA | MODINFOMD_KEYBUF); 281ec5c0e5bSAllan Jude 282ec5c0e5bSAllan Jude if (keybuf == NULL) 283ec5c0e5bSAllan Jude keybuf = &empty_keybuf; 284ec5c0e5bSAllan Jude } 285ec5c0e5bSAllan Jude 286ec5c0e5bSAllan Jude /* It'd be nice if we could store these in some kind of secure memory... */ 2875973f492SJohn Baldwin struct keybuf * 2885973f492SJohn Baldwin get_keybuf(void) 2895973f492SJohn Baldwin { 290ec5c0e5bSAllan Jude 291ec5c0e5bSAllan Jude return (keybuf); 292ec5c0e5bSAllan Jude } 293ec5c0e5bSAllan Jude 294c0341432SJohn Baldwin static struct cryptocap * 295c0341432SJohn Baldwin cap_ref(struct cryptocap *cap) 296c0341432SJohn Baldwin { 297c0341432SJohn Baldwin 298c0341432SJohn Baldwin refcount_acquire(&cap->cc_refs); 299c0341432SJohn Baldwin return (cap); 300c0341432SJohn Baldwin } 301c0341432SJohn Baldwin 302c0341432SJohn Baldwin static void 303c0341432SJohn Baldwin cap_rele(struct cryptocap *cap) 304c0341432SJohn Baldwin { 305c0341432SJohn Baldwin 306c0341432SJohn Baldwin if (refcount_release(&cap->cc_refs) == 0) 307c0341432SJohn Baldwin return; 308c0341432SJohn Baldwin 309c0341432SJohn Baldwin KASSERT(cap->cc_sessions == 0, 310c0341432SJohn Baldwin ("freeing crypto driver with active sessions")); 311c0341432SJohn Baldwin KASSERT(cap->cc_koperations == 0, 312c0341432SJohn Baldwin ("freeing crypto driver with active key operations")); 313c0341432SJohn Baldwin 314c0341432SJohn Baldwin free(cap, M_CRYPTO_DATA); 315c0341432SJohn Baldwin } 316c0341432SJohn Baldwin 31751e45326SSam Leffler static int 318091d81d1SSam Leffler crypto_init(void) 319091d81d1SSam Leffler { 32039bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 32151e45326SSam Leffler int error; 322091d81d1SSam Leffler 3233569ae7fSSam Leffler mtx_init(&crypto_drivers_mtx, "crypto", "crypto driver table", 3243569ae7fSSam Leffler MTX_DEF|MTX_QUIET); 325091d81d1SSam Leffler 326091d81d1SSam Leffler TAILQ_INIT(&crp_q); 327091d81d1SSam Leffler TAILQ_INIT(&crp_kq); 3283569ae7fSSam Leffler mtx_init(&crypto_q_mtx, "crypto", "crypto op queues", MTX_DEF); 329091d81d1SSam Leffler 330e5587cbbSMark Johnston cryptop_zone = uma_zcreate("cryptop", 331e5587cbbSMark Johnston sizeof(struct cryptop), NULL, NULL, NULL, NULL, 33251e45326SSam Leffler UMA_ALIGN_PTR, UMA_ZONE_ZINIT); 3331b0909d5SConrad Meyer 334c0341432SJohn Baldwin crypto_drivers_size = CRYPTO_DRIVERS_INITIAL; 335c0341432SJohn Baldwin crypto_drivers = malloc(crypto_drivers_size * 336e5587cbbSMark Johnston sizeof(struct cryptocap), M_CRYPTO_DATA, M_WAITOK | M_ZERO); 33751e45326SSam Leffler 33839bbca6fSFabien Thomas if (crypto_workers_num < 1 || crypto_workers_num > mp_ncpus) 33939bbca6fSFabien Thomas crypto_workers_num = mp_ncpus; 34039bbca6fSFabien Thomas 34139bbca6fSFabien Thomas crypto_tq = taskqueue_create("crypto", M_WAITOK | M_ZERO, 34239bbca6fSFabien Thomas taskqueue_thread_enqueue, &crypto_tq); 34339bbca6fSFabien Thomas 34439bbca6fSFabien Thomas taskqueue_start_threads(&crypto_tq, crypto_workers_num, PRI_MIN_KERN, 34539bbca6fSFabien Thomas "crypto"); 34639bbca6fSFabien Thomas 3473745c395SJulian Elischer error = kproc_create((void (*)(void *)) crypto_proc, NULL, 34851e45326SSam Leffler &cryptoproc, 0, 0, "crypto"); 34951e45326SSam Leffler if (error) { 35051e45326SSam Leffler printf("crypto_init: cannot start crypto thread; error %d", 35151e45326SSam Leffler error); 35251e45326SSam Leffler goto bad; 35351e45326SSam Leffler } 35451e45326SSam Leffler 355e5587cbbSMark Johnston crypto_ret_workers = mallocarray(crypto_workers_num, 356e5587cbbSMark Johnston sizeof(struct crypto_ret_worker), M_CRYPTO_DATA, M_WAITOK | M_ZERO); 35739bbca6fSFabien Thomas 35839bbca6fSFabien Thomas FOREACH_CRYPTO_RETW(ret_worker) { 35939bbca6fSFabien Thomas TAILQ_INIT(&ret_worker->crp_ordered_ret_q); 36039bbca6fSFabien Thomas TAILQ_INIT(&ret_worker->crp_ret_q); 36139bbca6fSFabien Thomas TAILQ_INIT(&ret_worker->crp_ret_kq); 36239bbca6fSFabien Thomas 36339bbca6fSFabien Thomas ret_worker->reorder_ops = 0; 36439bbca6fSFabien Thomas ret_worker->reorder_cur_seq = 0; 36539bbca6fSFabien Thomas 36639bbca6fSFabien Thomas mtx_init(&ret_worker->crypto_ret_mtx, "crypto", "crypto return queues", MTX_DEF); 36739bbca6fSFabien Thomas 36839bbca6fSFabien Thomas error = kproc_create((void (*)(void *)) crypto_ret_proc, ret_worker, 36939bbca6fSFabien Thomas &ret_worker->cryptoretproc, 0, 0, "crypto returns %td", CRYPTO_RETW_ID(ret_worker)); 37051e45326SSam Leffler if (error) { 37151e45326SSam Leffler printf("crypto_init: cannot start cryptoret thread; error %d", 37251e45326SSam Leffler error); 37351e45326SSam Leffler goto bad; 37451e45326SSam Leffler } 37539bbca6fSFabien Thomas } 376ec5c0e5bSAllan Jude 377ec5c0e5bSAllan Jude keybuf_init(); 378ec5c0e5bSAllan Jude 37951e45326SSam Leffler return 0; 38051e45326SSam Leffler bad: 38151e45326SSam Leffler crypto_destroy(); 38251e45326SSam Leffler return error; 38351e45326SSam Leffler } 38451e45326SSam Leffler 38551e45326SSam Leffler /* 38651e45326SSam Leffler * Signal a crypto thread to terminate. We use the driver 38751e45326SSam Leffler * table lock to synchronize the sleep/wakeups so that we 38851e45326SSam Leffler * are sure the threads have terminated before we release 38951e45326SSam Leffler * the data structures they use. See crypto_finis below 39051e45326SSam Leffler * for the other half of this song-and-dance. 39151e45326SSam Leffler */ 39251e45326SSam Leffler static void 39351e45326SSam Leffler crypto_terminate(struct proc **pp, void *q) 39451e45326SSam Leffler { 39551e45326SSam Leffler struct proc *p; 39651e45326SSam Leffler 39751e45326SSam Leffler mtx_assert(&crypto_drivers_mtx, MA_OWNED); 39851e45326SSam Leffler p = *pp; 39951e45326SSam Leffler *pp = NULL; 40051e45326SSam Leffler if (p) { 40151e45326SSam Leffler wakeup_one(q); 40251e45326SSam Leffler PROC_LOCK(p); /* NB: insure we don't miss wakeup */ 40351e45326SSam Leffler CRYPTO_DRIVER_UNLOCK(); /* let crypto_finis progress */ 40451e45326SSam Leffler msleep(p, &p->p_mtx, PWAIT, "crypto_destroy", 0); 40551e45326SSam Leffler PROC_UNLOCK(p); 40651e45326SSam Leffler CRYPTO_DRIVER_LOCK(); 40751e45326SSam Leffler } 40851e45326SSam Leffler } 40951e45326SSam Leffler 41051e45326SSam Leffler static void 411d588dc7dSMark Johnston hmac_init_pad(const struct auth_hash *axf, const char *key, int klen, 412d588dc7dSMark Johnston void *auth_ctx, uint8_t padval) 413c0341432SJohn Baldwin { 414c0341432SJohn Baldwin uint8_t hmac_key[HMAC_MAX_BLOCK_LEN]; 415c0341432SJohn Baldwin u_int i; 416c0341432SJohn Baldwin 417c0341432SJohn Baldwin KASSERT(axf->blocksize <= sizeof(hmac_key), 418c0341432SJohn Baldwin ("Invalid HMAC block size %d", axf->blocksize)); 419c0341432SJohn Baldwin 420c0341432SJohn Baldwin /* 421c0341432SJohn Baldwin * If the key is larger than the block size, use the digest of 422c0341432SJohn Baldwin * the key as the key instead. 423c0341432SJohn Baldwin */ 424c0341432SJohn Baldwin memset(hmac_key, 0, sizeof(hmac_key)); 425c0341432SJohn Baldwin if (klen > axf->blocksize) { 426c0341432SJohn Baldwin axf->Init(auth_ctx); 427c0341432SJohn Baldwin axf->Update(auth_ctx, key, klen); 428c0341432SJohn Baldwin axf->Final(hmac_key, auth_ctx); 429c0341432SJohn Baldwin klen = axf->hashsize; 430c0341432SJohn Baldwin } else 431c0341432SJohn Baldwin memcpy(hmac_key, key, klen); 432c0341432SJohn Baldwin 433c0341432SJohn Baldwin for (i = 0; i < axf->blocksize; i++) 434c0341432SJohn Baldwin hmac_key[i] ^= padval; 435c0341432SJohn Baldwin 436c0341432SJohn Baldwin axf->Init(auth_ctx); 437c0341432SJohn Baldwin axf->Update(auth_ctx, hmac_key, axf->blocksize); 43817a831eaSJohn Baldwin explicit_bzero(hmac_key, sizeof(hmac_key)); 439c0341432SJohn Baldwin } 440c0341432SJohn Baldwin 441c0341432SJohn Baldwin void 442d588dc7dSMark Johnston hmac_init_ipad(const struct auth_hash *axf, const char *key, int klen, 443c0341432SJohn Baldwin void *auth_ctx) 444c0341432SJohn Baldwin { 445c0341432SJohn Baldwin 446c0341432SJohn Baldwin hmac_init_pad(axf, key, klen, auth_ctx, HMAC_IPAD_VAL); 447c0341432SJohn Baldwin } 448c0341432SJohn Baldwin 449c0341432SJohn Baldwin void 450d588dc7dSMark Johnston hmac_init_opad(const struct auth_hash *axf, const char *key, int klen, 451c0341432SJohn Baldwin void *auth_ctx) 452c0341432SJohn Baldwin { 453c0341432SJohn Baldwin 454c0341432SJohn Baldwin hmac_init_pad(axf, key, klen, auth_ctx, HMAC_OPAD_VAL); 455c0341432SJohn Baldwin } 456c0341432SJohn Baldwin 457c0341432SJohn Baldwin static void 45851e45326SSam Leffler crypto_destroy(void) 45951e45326SSam Leffler { 46039bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 461c0341432SJohn Baldwin int i; 46239bbca6fSFabien Thomas 46351e45326SSam Leffler /* 46451e45326SSam Leffler * Terminate any crypto threads. 46551e45326SSam Leffler */ 46639bbca6fSFabien Thomas if (crypto_tq != NULL) 46739bbca6fSFabien Thomas taskqueue_drain_all(crypto_tq); 46851e45326SSam Leffler CRYPTO_DRIVER_LOCK(); 46951e45326SSam Leffler crypto_terminate(&cryptoproc, &crp_q); 47039bbca6fSFabien Thomas FOREACH_CRYPTO_RETW(ret_worker) 47139bbca6fSFabien Thomas crypto_terminate(&ret_worker->cryptoretproc, &ret_worker->crp_ret_q); 47251e45326SSam Leffler CRYPTO_DRIVER_UNLOCK(); 47351e45326SSam Leffler 47451e45326SSam Leffler /* XXX flush queues??? */ 47551e45326SSam Leffler 47651e45326SSam Leffler /* 47751e45326SSam Leffler * Reclaim dynamically allocated resources. 47851e45326SSam Leffler */ 479c0341432SJohn Baldwin for (i = 0; i < crypto_drivers_size; i++) { 480c0341432SJohn Baldwin if (crypto_drivers[i] != NULL) 481c0341432SJohn Baldwin cap_rele(crypto_drivers[i]); 482c0341432SJohn Baldwin } 48351e45326SSam Leffler free(crypto_drivers, M_CRYPTO_DATA); 48451e45326SSam Leffler 48551e45326SSam Leffler if (cryptop_zone != NULL) 48651e45326SSam Leffler uma_zdestroy(cryptop_zone); 48751e45326SSam Leffler mtx_destroy(&crypto_q_mtx); 48839bbca6fSFabien Thomas FOREACH_CRYPTO_RETW(ret_worker) 48939bbca6fSFabien Thomas mtx_destroy(&ret_worker->crypto_ret_mtx); 49039bbca6fSFabien Thomas free(crypto_ret_workers, M_CRYPTO_DATA); 49139bbca6fSFabien Thomas if (crypto_tq != NULL) 49239bbca6fSFabien Thomas taskqueue_free(crypto_tq); 49351e45326SSam Leffler mtx_destroy(&crypto_drivers_mtx); 494091d81d1SSam Leffler } 495f544a528SMark Murray 4961b0909d5SConrad Meyer uint32_t 4971b0909d5SConrad Meyer crypto_ses2hid(crypto_session_t crypto_session) 4981b0909d5SConrad Meyer { 499c0341432SJohn Baldwin return (crypto_session->cap->cc_hid); 5001b0909d5SConrad Meyer } 5011b0909d5SConrad Meyer 5021b0909d5SConrad Meyer uint32_t 5031b0909d5SConrad Meyer crypto_ses2caps(crypto_session_t crypto_session) 5041b0909d5SConrad Meyer { 505c0341432SJohn Baldwin return (crypto_session->cap->cc_flags & 0xff000000); 5061b0909d5SConrad Meyer } 5071b0909d5SConrad Meyer 5081b0909d5SConrad Meyer void * 5091b0909d5SConrad Meyer crypto_get_driver_session(crypto_session_t crypto_session) 5101b0909d5SConrad Meyer { 511d1816248SMark Johnston return (crypto_session + 1); 5121b0909d5SConrad Meyer } 5131b0909d5SConrad Meyer 514c0341432SJohn Baldwin const struct crypto_session_params * 515c0341432SJohn Baldwin crypto_get_params(crypto_session_t crypto_session) 516c0341432SJohn Baldwin { 517c0341432SJohn Baldwin return (&crypto_session->csp); 518c0341432SJohn Baldwin } 519c0341432SJohn Baldwin 520c0341432SJohn Baldwin struct auth_hash * 521c0341432SJohn Baldwin crypto_auth_hash(const struct crypto_session_params *csp) 522c0341432SJohn Baldwin { 523c0341432SJohn Baldwin 524c0341432SJohn Baldwin switch (csp->csp_auth_alg) { 525c0341432SJohn Baldwin case CRYPTO_SHA1_HMAC: 526c0341432SJohn Baldwin return (&auth_hash_hmac_sha1); 527c0341432SJohn Baldwin case CRYPTO_SHA2_224_HMAC: 528c0341432SJohn Baldwin return (&auth_hash_hmac_sha2_224); 529c0341432SJohn Baldwin case CRYPTO_SHA2_256_HMAC: 530c0341432SJohn Baldwin return (&auth_hash_hmac_sha2_256); 531c0341432SJohn Baldwin case CRYPTO_SHA2_384_HMAC: 532c0341432SJohn Baldwin return (&auth_hash_hmac_sha2_384); 533c0341432SJohn Baldwin case CRYPTO_SHA2_512_HMAC: 534c0341432SJohn Baldwin return (&auth_hash_hmac_sha2_512); 535c0341432SJohn Baldwin case CRYPTO_NULL_HMAC: 536c0341432SJohn Baldwin return (&auth_hash_null); 537c0341432SJohn Baldwin case CRYPTO_RIPEMD160_HMAC: 538c0341432SJohn Baldwin return (&auth_hash_hmac_ripemd_160); 539c0341432SJohn Baldwin case CRYPTO_SHA1: 540c0341432SJohn Baldwin return (&auth_hash_sha1); 541c0341432SJohn Baldwin case CRYPTO_SHA2_224: 542c0341432SJohn Baldwin return (&auth_hash_sha2_224); 543c0341432SJohn Baldwin case CRYPTO_SHA2_256: 544c0341432SJohn Baldwin return (&auth_hash_sha2_256); 545c0341432SJohn Baldwin case CRYPTO_SHA2_384: 546c0341432SJohn Baldwin return (&auth_hash_sha2_384); 547c0341432SJohn Baldwin case CRYPTO_SHA2_512: 548c0341432SJohn Baldwin return (&auth_hash_sha2_512); 549c0341432SJohn Baldwin case CRYPTO_AES_NIST_GMAC: 550c0341432SJohn Baldwin switch (csp->csp_auth_klen) { 551c0341432SJohn Baldwin case 128 / 8: 552c0341432SJohn Baldwin return (&auth_hash_nist_gmac_aes_128); 553c0341432SJohn Baldwin case 192 / 8: 554c0341432SJohn Baldwin return (&auth_hash_nist_gmac_aes_192); 555c0341432SJohn Baldwin case 256 / 8: 556c0341432SJohn Baldwin return (&auth_hash_nist_gmac_aes_256); 557c0341432SJohn Baldwin default: 558c0341432SJohn Baldwin return (NULL); 559c0341432SJohn Baldwin } 560c0341432SJohn Baldwin case CRYPTO_BLAKE2B: 561c0341432SJohn Baldwin return (&auth_hash_blake2b); 562c0341432SJohn Baldwin case CRYPTO_BLAKE2S: 563c0341432SJohn Baldwin return (&auth_hash_blake2s); 564c0341432SJohn Baldwin case CRYPTO_POLY1305: 565c0341432SJohn Baldwin return (&auth_hash_poly1305); 566c0341432SJohn Baldwin case CRYPTO_AES_CCM_CBC_MAC: 567c0341432SJohn Baldwin switch (csp->csp_auth_klen) { 568c0341432SJohn Baldwin case 128 / 8: 569c0341432SJohn Baldwin return (&auth_hash_ccm_cbc_mac_128); 570c0341432SJohn Baldwin case 192 / 8: 571c0341432SJohn Baldwin return (&auth_hash_ccm_cbc_mac_192); 572c0341432SJohn Baldwin case 256 / 8: 573c0341432SJohn Baldwin return (&auth_hash_ccm_cbc_mac_256); 574c0341432SJohn Baldwin default: 575c0341432SJohn Baldwin return (NULL); 576c0341432SJohn Baldwin } 577c0341432SJohn Baldwin default: 578c0341432SJohn Baldwin return (NULL); 579c0341432SJohn Baldwin } 580c0341432SJohn Baldwin } 581c0341432SJohn Baldwin 582c0341432SJohn Baldwin struct enc_xform * 583c0341432SJohn Baldwin crypto_cipher(const struct crypto_session_params *csp) 584c0341432SJohn Baldwin { 585c0341432SJohn Baldwin 586c0341432SJohn Baldwin switch (csp->csp_cipher_alg) { 587c0341432SJohn Baldwin case CRYPTO_RIJNDAEL128_CBC: 588c0341432SJohn Baldwin return (&enc_xform_rijndael128); 589c0341432SJohn Baldwin case CRYPTO_AES_XTS: 590c0341432SJohn Baldwin return (&enc_xform_aes_xts); 591c0341432SJohn Baldwin case CRYPTO_AES_ICM: 592c0341432SJohn Baldwin return (&enc_xform_aes_icm); 593c0341432SJohn Baldwin case CRYPTO_AES_NIST_GCM_16: 594c0341432SJohn Baldwin return (&enc_xform_aes_nist_gcm); 595c0341432SJohn Baldwin case CRYPTO_CAMELLIA_CBC: 596c0341432SJohn Baldwin return (&enc_xform_camellia); 597c0341432SJohn Baldwin case CRYPTO_NULL_CBC: 598c0341432SJohn Baldwin return (&enc_xform_null); 599c0341432SJohn Baldwin case CRYPTO_CHACHA20: 600c0341432SJohn Baldwin return (&enc_xform_chacha20); 601c0341432SJohn Baldwin case CRYPTO_AES_CCM_16: 602c0341432SJohn Baldwin return (&enc_xform_ccm); 603*fc8fc743SJohn Baldwin case CRYPTO_CHACHA20_POLY1305: 604*fc8fc743SJohn Baldwin return (&enc_xform_chacha20_poly1305); 605c0341432SJohn Baldwin default: 606c0341432SJohn Baldwin return (NULL); 607c0341432SJohn Baldwin } 608c0341432SJohn Baldwin } 609c0341432SJohn Baldwin 6106810ad6fSSam Leffler static struct cryptocap * 611d3d79e96SJohn Baldwin crypto_checkdriver(uint32_t hid) 6126810ad6fSSam Leffler { 6136810ad6fSSam Leffler 614c0341432SJohn Baldwin return (hid >= crypto_drivers_size ? NULL : crypto_drivers[hid]); 615f544a528SMark Murray } 616f544a528SMark Murray 617091d81d1SSam Leffler /* 6186810ad6fSSam Leffler * Select a driver for a new session that supports the specified 6196810ad6fSSam Leffler * algorithms and, optionally, is constrained according to the flags. 620091d81d1SSam Leffler */ 6216810ad6fSSam Leffler static struct cryptocap * 622c0341432SJohn Baldwin crypto_select_driver(const struct crypto_session_params *csp, int flags) 6236810ad6fSSam Leffler { 6246810ad6fSSam Leffler struct cryptocap *cap, *best; 625c0341432SJohn Baldwin int best_match, error, hid; 6266810ad6fSSam Leffler 6276810ad6fSSam Leffler CRYPTO_DRIVER_ASSERT(); 628091d81d1SSam Leffler 6296810ad6fSSam Leffler best = NULL; 630c0341432SJohn Baldwin for (hid = 0; hid < crypto_drivers_size; hid++) { 631091d81d1SSam Leffler /* 632c0341432SJohn Baldwin * If there is no driver for this slot, or the driver 633c0341432SJohn Baldwin * is not appropriate (hardware or software based on 634c0341432SJohn Baldwin * match), then skip. 635091d81d1SSam Leffler */ 636c0341432SJohn Baldwin cap = crypto_drivers[hid]; 637c0341432SJohn Baldwin if (cap == NULL || 638c0341432SJohn Baldwin (cap->cc_flags & flags) == 0) 639091d81d1SSam Leffler continue; 640091d81d1SSam Leffler 641c0341432SJohn Baldwin error = CRYPTODEV_PROBESESSION(cap->cc_dev, csp); 642c0341432SJohn Baldwin if (error >= 0) 643c0341432SJohn Baldwin continue; 644c0341432SJohn Baldwin 645c0341432SJohn Baldwin /* 646c0341432SJohn Baldwin * Use the driver with the highest probe value. 647c0341432SJohn Baldwin * Hardware drivers use a higher probe value than 648c0341432SJohn Baldwin * software. In case of a tie, prefer the driver with 649c0341432SJohn Baldwin * the fewest active sessions. 650c0341432SJohn Baldwin */ 651c0341432SJohn Baldwin if (best == NULL || error > best_match || 652c0341432SJohn Baldwin (error == best_match && 653c0341432SJohn Baldwin cap->cc_sessions < best->cc_sessions)) { 6546810ad6fSSam Leffler best = cap; 655c0341432SJohn Baldwin best_match = error; 6566810ad6fSSam Leffler } 6576810ad6fSSam Leffler } 6586810ad6fSSam Leffler return best; 6596810ad6fSSam Leffler } 660091d81d1SSam Leffler 661ad557055SJohn Baldwin static enum alg_type { 662ad557055SJohn Baldwin ALG_NONE = 0, 663ad557055SJohn Baldwin ALG_CIPHER, 664ad557055SJohn Baldwin ALG_DIGEST, 665ad557055SJohn Baldwin ALG_KEYED_DIGEST, 666ad557055SJohn Baldwin ALG_COMPRESSION, 667ad557055SJohn Baldwin ALG_AEAD 668ad557055SJohn Baldwin } alg_types[] = { 669ad557055SJohn Baldwin [CRYPTO_SHA1_HMAC] = ALG_KEYED_DIGEST, 670ad557055SJohn Baldwin [CRYPTO_RIPEMD160_HMAC] = ALG_KEYED_DIGEST, 671ad557055SJohn Baldwin [CRYPTO_AES_CBC] = ALG_CIPHER, 672ad557055SJohn Baldwin [CRYPTO_SHA1] = ALG_DIGEST, 673ad557055SJohn Baldwin [CRYPTO_NULL_HMAC] = ALG_DIGEST, 674ad557055SJohn Baldwin [CRYPTO_NULL_CBC] = ALG_CIPHER, 675ad557055SJohn Baldwin [CRYPTO_DEFLATE_COMP] = ALG_COMPRESSION, 676ad557055SJohn Baldwin [CRYPTO_SHA2_256_HMAC] = ALG_KEYED_DIGEST, 677ad557055SJohn Baldwin [CRYPTO_SHA2_384_HMAC] = ALG_KEYED_DIGEST, 678ad557055SJohn Baldwin [CRYPTO_SHA2_512_HMAC] = ALG_KEYED_DIGEST, 679ad557055SJohn Baldwin [CRYPTO_CAMELLIA_CBC] = ALG_CIPHER, 680ad557055SJohn Baldwin [CRYPTO_AES_XTS] = ALG_CIPHER, 681ad557055SJohn Baldwin [CRYPTO_AES_ICM] = ALG_CIPHER, 682ad557055SJohn Baldwin [CRYPTO_AES_NIST_GMAC] = ALG_KEYED_DIGEST, 683ad557055SJohn Baldwin [CRYPTO_AES_NIST_GCM_16] = ALG_AEAD, 684ad557055SJohn Baldwin [CRYPTO_BLAKE2B] = ALG_KEYED_DIGEST, 685ad557055SJohn Baldwin [CRYPTO_BLAKE2S] = ALG_KEYED_DIGEST, 686ad557055SJohn Baldwin [CRYPTO_CHACHA20] = ALG_CIPHER, 687ad557055SJohn Baldwin [CRYPTO_SHA2_224_HMAC] = ALG_KEYED_DIGEST, 688ad557055SJohn Baldwin [CRYPTO_RIPEMD160] = ALG_DIGEST, 689ad557055SJohn Baldwin [CRYPTO_SHA2_224] = ALG_DIGEST, 690ad557055SJohn Baldwin [CRYPTO_SHA2_256] = ALG_DIGEST, 691ad557055SJohn Baldwin [CRYPTO_SHA2_384] = ALG_DIGEST, 692ad557055SJohn Baldwin [CRYPTO_SHA2_512] = ALG_DIGEST, 693ad557055SJohn Baldwin [CRYPTO_POLY1305] = ALG_KEYED_DIGEST, 694ad557055SJohn Baldwin [CRYPTO_AES_CCM_CBC_MAC] = ALG_KEYED_DIGEST, 695ad557055SJohn Baldwin [CRYPTO_AES_CCM_16] = ALG_AEAD, 696*fc8fc743SJohn Baldwin [CRYPTO_CHACHA20_POLY1305] = ALG_AEAD, 697ad557055SJohn Baldwin }; 698ad557055SJohn Baldwin 699ad557055SJohn Baldwin static enum alg_type 700ad557055SJohn Baldwin alg_type(int alg) 701ad557055SJohn Baldwin { 702ad557055SJohn Baldwin 703ad557055SJohn Baldwin if (alg < nitems(alg_types)) 704ad557055SJohn Baldwin return (alg_types[alg]); 705ad557055SJohn Baldwin return (ALG_NONE); 706ad557055SJohn Baldwin } 707ad557055SJohn Baldwin 708c0341432SJohn Baldwin static bool 709c0341432SJohn Baldwin alg_is_compression(int alg) 710c0341432SJohn Baldwin { 711c0341432SJohn Baldwin 712ad557055SJohn Baldwin return (alg_type(alg) == ALG_COMPRESSION); 713c0341432SJohn Baldwin } 714c0341432SJohn Baldwin 715c0341432SJohn Baldwin static bool 716c0341432SJohn Baldwin alg_is_cipher(int alg) 717c0341432SJohn Baldwin { 718c0341432SJohn Baldwin 719ad557055SJohn Baldwin return (alg_type(alg) == ALG_CIPHER); 720c0341432SJohn Baldwin } 721c0341432SJohn Baldwin 722c0341432SJohn Baldwin static bool 723c0341432SJohn Baldwin alg_is_digest(int alg) 724c0341432SJohn Baldwin { 725c0341432SJohn Baldwin 726ad557055SJohn Baldwin return (alg_type(alg) == ALG_DIGEST || 727ad557055SJohn Baldwin alg_type(alg) == ALG_KEYED_DIGEST); 728c0341432SJohn Baldwin } 729c0341432SJohn Baldwin 730c0341432SJohn Baldwin static bool 731c0341432SJohn Baldwin alg_is_keyed_digest(int alg) 732c0341432SJohn Baldwin { 733c0341432SJohn Baldwin 734ad557055SJohn Baldwin return (alg_type(alg) == ALG_KEYED_DIGEST); 735c0341432SJohn Baldwin } 736c0341432SJohn Baldwin 737c0341432SJohn Baldwin static bool 738c0341432SJohn Baldwin alg_is_aead(int alg) 739c0341432SJohn Baldwin { 740c0341432SJohn Baldwin 741ad557055SJohn Baldwin return (alg_type(alg) == ALG_AEAD); 742c0341432SJohn Baldwin } 743c0341432SJohn Baldwin 7447e89ae49SMarcin Wojtas #define SUPPORTED_SES (CSP_F_SEPARATE_OUTPUT | CSP_F_SEPARATE_AAD | CSP_F_ESN) 7457e89ae49SMarcin Wojtas 746c0341432SJohn Baldwin /* Various sanity checks on crypto session parameters. */ 747c0341432SJohn Baldwin static bool 748c0341432SJohn Baldwin check_csp(const struct crypto_session_params *csp) 749c0341432SJohn Baldwin { 750c0341432SJohn Baldwin struct auth_hash *axf; 751c0341432SJohn Baldwin 752c0341432SJohn Baldwin /* Mode-independent checks. */ 7537e89ae49SMarcin Wojtas if ((csp->csp_flags & ~(SUPPORTED_SES)) != 0) 754c0341432SJohn Baldwin return (false); 755c0341432SJohn Baldwin if (csp->csp_ivlen < 0 || csp->csp_cipher_klen < 0 || 756c0341432SJohn Baldwin csp->csp_auth_klen < 0 || csp->csp_auth_mlen < 0) 757c0341432SJohn Baldwin return (false); 758c0341432SJohn Baldwin if (csp->csp_auth_key != NULL && csp->csp_auth_klen == 0) 759c0341432SJohn Baldwin return (false); 760c0341432SJohn Baldwin if (csp->csp_cipher_key != NULL && csp->csp_cipher_klen == 0) 761c0341432SJohn Baldwin return (false); 762c0341432SJohn Baldwin 763c0341432SJohn Baldwin switch (csp->csp_mode) { 764c0341432SJohn Baldwin case CSP_MODE_COMPRESS: 765c0341432SJohn Baldwin if (!alg_is_compression(csp->csp_cipher_alg)) 766c0341432SJohn Baldwin return (false); 7679c0e3d3aSJohn Baldwin if (csp->csp_flags & CSP_F_SEPARATE_OUTPUT) 768c0341432SJohn Baldwin return (false); 7699b774dc0SJohn Baldwin if (csp->csp_flags & CSP_F_SEPARATE_AAD) 7709b774dc0SJohn Baldwin return (false); 771c0341432SJohn Baldwin if (csp->csp_cipher_klen != 0 || csp->csp_ivlen != 0 || 772c0341432SJohn Baldwin csp->csp_auth_alg != 0 || csp->csp_auth_klen != 0 || 773c0341432SJohn Baldwin csp->csp_auth_mlen != 0) 774c0341432SJohn Baldwin return (false); 775c0341432SJohn Baldwin break; 776c0341432SJohn Baldwin case CSP_MODE_CIPHER: 777c0341432SJohn Baldwin if (!alg_is_cipher(csp->csp_cipher_alg)) 778c0341432SJohn Baldwin return (false); 7799b774dc0SJohn Baldwin if (csp->csp_flags & CSP_F_SEPARATE_AAD) 7809b774dc0SJohn Baldwin return (false); 781c0341432SJohn Baldwin if (csp->csp_cipher_alg != CRYPTO_NULL_CBC) { 782c0341432SJohn Baldwin if (csp->csp_cipher_klen == 0) 783c0341432SJohn Baldwin return (false); 784c0341432SJohn Baldwin if (csp->csp_ivlen == 0) 785c0341432SJohn Baldwin return (false); 786c0341432SJohn Baldwin } 787c0341432SJohn Baldwin if (csp->csp_ivlen >= EALG_MAX_BLOCK_LEN) 788c0341432SJohn Baldwin return (false); 789c0341432SJohn Baldwin if (csp->csp_auth_alg != 0 || csp->csp_auth_klen != 0 || 790c0341432SJohn Baldwin csp->csp_auth_mlen != 0) 791c0341432SJohn Baldwin return (false); 792c0341432SJohn Baldwin break; 793c0341432SJohn Baldwin case CSP_MODE_DIGEST: 794c0341432SJohn Baldwin if (csp->csp_cipher_alg != 0 || csp->csp_cipher_klen != 0) 795c0341432SJohn Baldwin return (false); 796c0341432SJohn Baldwin 7979b774dc0SJohn Baldwin if (csp->csp_flags & CSP_F_SEPARATE_AAD) 7989b774dc0SJohn Baldwin return (false); 7999b774dc0SJohn Baldwin 800c0341432SJohn Baldwin /* IV is optional for digests (e.g. GMAC). */ 801c0341432SJohn Baldwin if (csp->csp_ivlen >= EALG_MAX_BLOCK_LEN) 802c0341432SJohn Baldwin return (false); 803c0341432SJohn Baldwin if (!alg_is_digest(csp->csp_auth_alg)) 804c0341432SJohn Baldwin return (false); 805c0341432SJohn Baldwin 806c0341432SJohn Baldwin /* Key is optional for BLAKE2 digests. */ 807c0341432SJohn Baldwin if (csp->csp_auth_alg == CRYPTO_BLAKE2B || 808c0341432SJohn Baldwin csp->csp_auth_alg == CRYPTO_BLAKE2S) 809c0341432SJohn Baldwin ; 810c0341432SJohn Baldwin else if (alg_is_keyed_digest(csp->csp_auth_alg)) { 811c0341432SJohn Baldwin if (csp->csp_auth_klen == 0) 812c0341432SJohn Baldwin return (false); 813c0341432SJohn Baldwin } else { 814c0341432SJohn Baldwin if (csp->csp_auth_klen != 0) 815c0341432SJohn Baldwin return (false); 816c0341432SJohn Baldwin } 817c0341432SJohn Baldwin if (csp->csp_auth_mlen != 0) { 818c0341432SJohn Baldwin axf = crypto_auth_hash(csp); 819c0341432SJohn Baldwin if (axf == NULL || csp->csp_auth_mlen > axf->hashsize) 820c0341432SJohn Baldwin return (false); 821c0341432SJohn Baldwin } 822c0341432SJohn Baldwin break; 823c0341432SJohn Baldwin case CSP_MODE_AEAD: 824c0341432SJohn Baldwin if (!alg_is_aead(csp->csp_cipher_alg)) 825c0341432SJohn Baldwin return (false); 826c0341432SJohn Baldwin if (csp->csp_cipher_klen == 0) 827c0341432SJohn Baldwin return (false); 828c0341432SJohn Baldwin if (csp->csp_ivlen == 0 || 829c0341432SJohn Baldwin csp->csp_ivlen >= EALG_MAX_BLOCK_LEN) 830c0341432SJohn Baldwin return (false); 831c0341432SJohn Baldwin if (csp->csp_auth_alg != 0 || csp->csp_auth_klen != 0) 832c0341432SJohn Baldwin return (false); 833c0341432SJohn Baldwin 834c0341432SJohn Baldwin /* 835c0341432SJohn Baldwin * XXX: Would be nice to have a better way to get this 836c0341432SJohn Baldwin * value. 837c0341432SJohn Baldwin */ 838c0341432SJohn Baldwin switch (csp->csp_cipher_alg) { 839c0341432SJohn Baldwin case CRYPTO_AES_NIST_GCM_16: 840c0341432SJohn Baldwin case CRYPTO_AES_CCM_16: 841*fc8fc743SJohn Baldwin case CRYPTO_CHACHA20_POLY1305: 842c0341432SJohn Baldwin if (csp->csp_auth_mlen > 16) 843c0341432SJohn Baldwin return (false); 844c0341432SJohn Baldwin break; 845c0341432SJohn Baldwin } 846c0341432SJohn Baldwin break; 847c0341432SJohn Baldwin case CSP_MODE_ETA: 848c0341432SJohn Baldwin if (!alg_is_cipher(csp->csp_cipher_alg)) 849c0341432SJohn Baldwin return (false); 850c0341432SJohn Baldwin if (csp->csp_cipher_alg != CRYPTO_NULL_CBC) { 851c0341432SJohn Baldwin if (csp->csp_cipher_klen == 0) 852c0341432SJohn Baldwin return (false); 853c0341432SJohn Baldwin if (csp->csp_ivlen == 0) 854c0341432SJohn Baldwin return (false); 855c0341432SJohn Baldwin } 856c0341432SJohn Baldwin if (csp->csp_ivlen >= EALG_MAX_BLOCK_LEN) 857c0341432SJohn Baldwin return (false); 858c0341432SJohn Baldwin if (!alg_is_digest(csp->csp_auth_alg)) 859c0341432SJohn Baldwin return (false); 860c0341432SJohn Baldwin 861c0341432SJohn Baldwin /* Key is optional for BLAKE2 digests. */ 862c0341432SJohn Baldwin if (csp->csp_auth_alg == CRYPTO_BLAKE2B || 863c0341432SJohn Baldwin csp->csp_auth_alg == CRYPTO_BLAKE2S) 864c0341432SJohn Baldwin ; 865c0341432SJohn Baldwin else if (alg_is_keyed_digest(csp->csp_auth_alg)) { 866c0341432SJohn Baldwin if (csp->csp_auth_klen == 0) 867c0341432SJohn Baldwin return (false); 868c0341432SJohn Baldwin } else { 869c0341432SJohn Baldwin if (csp->csp_auth_klen != 0) 870c0341432SJohn Baldwin return (false); 871c0341432SJohn Baldwin } 872c0341432SJohn Baldwin if (csp->csp_auth_mlen != 0) { 873c0341432SJohn Baldwin axf = crypto_auth_hash(csp); 874c0341432SJohn Baldwin if (axf == NULL || csp->csp_auth_mlen > axf->hashsize) 875c0341432SJohn Baldwin return (false); 876c0341432SJohn Baldwin } 877c0341432SJohn Baldwin break; 878c0341432SJohn Baldwin default: 879c0341432SJohn Baldwin return (false); 880c0341432SJohn Baldwin } 881c0341432SJohn Baldwin 882c0341432SJohn Baldwin return (true); 883c0341432SJohn Baldwin } 884c0341432SJohn Baldwin 885c0341432SJohn Baldwin /* 886c0341432SJohn Baldwin * Delete a session after it has been detached from its driver. 887c0341432SJohn Baldwin */ 888c0341432SJohn Baldwin static void 889c0341432SJohn Baldwin crypto_deletesession(crypto_session_t cses) 890c0341432SJohn Baldwin { 891c0341432SJohn Baldwin struct cryptocap *cap; 892c0341432SJohn Baldwin 893c0341432SJohn Baldwin cap = cses->cap; 894c0341432SJohn Baldwin 895d1816248SMark Johnston zfree(cses, M_CRYPTO_DATA); 896c0341432SJohn Baldwin 897c0341432SJohn Baldwin CRYPTO_DRIVER_LOCK(); 898c0341432SJohn Baldwin cap->cc_sessions--; 899c0341432SJohn Baldwin if (cap->cc_sessions == 0 && cap->cc_flags & CRYPTOCAP_F_CLEANUP) 900c0341432SJohn Baldwin wakeup(cap); 901c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 902c0341432SJohn Baldwin cap_rele(cap); 903c0341432SJohn Baldwin } 904c0341432SJohn Baldwin 905694e0113SPawel Jakub Dawidek /* 9066810ad6fSSam Leffler * Create a new session. The crid argument specifies a crypto 9076810ad6fSSam Leffler * driver to use or constraints on a driver to select (hardware 9086810ad6fSSam Leffler * only, software only, either). Whatever driver is selected 9096810ad6fSSam Leffler * must be capable of the requested crypto algorithms. 910694e0113SPawel Jakub Dawidek */ 9116810ad6fSSam Leffler int 912c0341432SJohn Baldwin crypto_newsession(crypto_session_t *cses, 913c0341432SJohn Baldwin const struct crypto_session_params *csp, int crid) 9146810ad6fSSam Leffler { 91598d788c8SMark Johnston static uint64_t sessid = 0; 9161b0909d5SConrad Meyer crypto_session_t res; 9176810ad6fSSam Leffler struct cryptocap *cap; 9186810ad6fSSam Leffler int err; 9196810ad6fSSam Leffler 920c0341432SJohn Baldwin if (!check_csp(csp)) 921c0341432SJohn Baldwin return (EINVAL); 922c0341432SJohn Baldwin 9231b0909d5SConrad Meyer res = NULL; 9241b0909d5SConrad Meyer 9256810ad6fSSam Leffler CRYPTO_DRIVER_LOCK(); 9266810ad6fSSam Leffler if ((crid & (CRYPTOCAP_F_HARDWARE | CRYPTOCAP_F_SOFTWARE)) == 0) { 927694e0113SPawel Jakub Dawidek /* 9286810ad6fSSam Leffler * Use specified driver; verify it is capable. 929694e0113SPawel Jakub Dawidek */ 9306810ad6fSSam Leffler cap = crypto_checkdriver(crid); 931c0341432SJohn Baldwin if (cap != NULL && CRYPTODEV_PROBESESSION(cap->cc_dev, csp) > 0) 932694e0113SPawel Jakub Dawidek cap = NULL; 9336810ad6fSSam Leffler } else { 9346810ad6fSSam Leffler /* 9356810ad6fSSam Leffler * No requested driver; select based on crid flags. 9366810ad6fSSam Leffler */ 937c0341432SJohn Baldwin cap = crypto_select_driver(csp, crid); 938694e0113SPawel Jakub Dawidek } 9391b0909d5SConrad Meyer if (cap == NULL) { 940c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 94108fca7a5SJohn-Mark Gurney CRYPTDEB("no driver"); 942c0341432SJohn Baldwin return (EOPNOTSUPP); 94308fca7a5SJohn-Mark Gurney } 944c0341432SJohn Baldwin cap_ref(cap); 9451b0909d5SConrad Meyer cap->cc_sessions++; 946091d81d1SSam Leffler CRYPTO_DRIVER_UNLOCK(); 9471b0909d5SConrad Meyer 9488adcc757SMark Johnston /* Allocate a single block for the generic session and driver softc. */ 949d1816248SMark Johnston res = malloc(sizeof(*res) + cap->cc_session_size, M_CRYPTO_DATA, 950d1816248SMark Johnston M_WAITOK | M_ZERO); 951c0341432SJohn Baldwin res->cap = cap; 952c0341432SJohn Baldwin res->csp = *csp; 95398d788c8SMark Johnston res->id = atomic_fetchadd_64(&sessid, 1); 9541b0909d5SConrad Meyer 9551b0909d5SConrad Meyer /* Call the driver initialization routine. */ 956c0341432SJohn Baldwin err = CRYPTODEV_NEWSESSION(cap->cc_dev, res, csp); 9571b0909d5SConrad Meyer if (err != 0) { 9581b0909d5SConrad Meyer CRYPTDEB("dev newsession failed: %d", err); 959c0341432SJohn Baldwin crypto_deletesession(res); 960c0341432SJohn Baldwin return (err); 9611b0909d5SConrad Meyer } 9621b0909d5SConrad Meyer 9631b0909d5SConrad Meyer *cses = res; 964c0341432SJohn Baldwin return (0); 9654acae0acSPawel Jakub Dawidek } 9664acae0acSPawel Jakub Dawidek 967091d81d1SSam Leffler /* 968091d81d1SSam Leffler * Delete an existing session (or a reserved session on an unregistered 969091d81d1SSam Leffler * driver). 970091d81d1SSam Leffler */ 9711b0909d5SConrad Meyer void 9721b0909d5SConrad Meyer crypto_freesession(crypto_session_t cses) 973091d81d1SSam Leffler { 9744acae0acSPawel Jakub Dawidek struct cryptocap *cap; 9751b0909d5SConrad Meyer 9761b0909d5SConrad Meyer if (cses == NULL) 9771b0909d5SConrad Meyer return; 978091d81d1SSam Leffler 979c0341432SJohn Baldwin cap = cses->cap; 980091d81d1SSam Leffler 981091d81d1SSam Leffler /* Call the driver cleanup routine, if available. */ 9821b0909d5SConrad Meyer CRYPTODEV_FREESESSION(cap->cc_dev, cses); 9831b0909d5SConrad Meyer 984c0341432SJohn Baldwin crypto_deletesession(cses); 985091d81d1SSam Leffler } 986091d81d1SSam Leffler 987091d81d1SSam Leffler /* 988c0341432SJohn Baldwin * Return a new driver id. Registers a driver with the system so that 989c0341432SJohn Baldwin * it can be probed by subsequent sessions. 990091d81d1SSam Leffler */ 991091d81d1SSam Leffler int32_t 9921b0909d5SConrad Meyer crypto_get_driverid(device_t dev, size_t sessionsize, int flags) 993091d81d1SSam Leffler { 994c0341432SJohn Baldwin struct cryptocap *cap, **newdrv; 995091d81d1SSam Leffler int i; 996091d81d1SSam Leffler 9976810ad6fSSam Leffler if ((flags & (CRYPTOCAP_F_HARDWARE | CRYPTOCAP_F_SOFTWARE)) == 0) { 998c0341432SJohn Baldwin device_printf(dev, 999c0341432SJohn Baldwin "no flags specified when registering driver\n"); 10006810ad6fSSam Leffler return -1; 10016810ad6fSSam Leffler } 10026810ad6fSSam Leffler 1003c0341432SJohn Baldwin cap = malloc(sizeof(*cap), M_CRYPTO_DATA, M_WAITOK | M_ZERO); 1004c0341432SJohn Baldwin cap->cc_dev = dev; 1005c0341432SJohn Baldwin cap->cc_session_size = sessionsize; 1006c0341432SJohn Baldwin cap->cc_flags = flags; 1007c0341432SJohn Baldwin refcount_init(&cap->cc_refs, 1); 1008c0341432SJohn Baldwin 1009091d81d1SSam Leffler CRYPTO_DRIVER_LOCK(); 1010c0341432SJohn Baldwin for (;;) { 1011c0341432SJohn Baldwin for (i = 0; i < crypto_drivers_size; i++) { 1012c0341432SJohn Baldwin if (crypto_drivers[i] == NULL) 1013091d81d1SSam Leffler break; 10144acae0acSPawel Jakub Dawidek } 1015c0341432SJohn Baldwin 1016c0341432SJohn Baldwin if (i < crypto_drivers_size) 1017c0341432SJohn Baldwin break; 1018091d81d1SSam Leffler 1019091d81d1SSam Leffler /* Out of entries, allocate some more. */ 1020c0341432SJohn Baldwin 1021c0341432SJohn Baldwin if (2 * crypto_drivers_size <= crypto_drivers_size) { 1022091d81d1SSam Leffler CRYPTO_DRIVER_UNLOCK(); 1023091d81d1SSam Leffler printf("crypto: driver count wraparound!\n"); 1024c0341432SJohn Baldwin cap_rele(cap); 1025c0341432SJohn Baldwin return (-1); 1026091d81d1SSam Leffler } 1027091d81d1SSam Leffler CRYPTO_DRIVER_UNLOCK(); 1028091d81d1SSam Leffler 1029c0341432SJohn Baldwin newdrv = malloc(2 * crypto_drivers_size * 1030c0341432SJohn Baldwin sizeof(*crypto_drivers), M_CRYPTO_DATA, M_WAITOK | M_ZERO); 1031091d81d1SSam Leffler 1032c0341432SJohn Baldwin CRYPTO_DRIVER_LOCK(); 1033c0341432SJohn Baldwin memcpy(newdrv, crypto_drivers, 1034c0341432SJohn Baldwin crypto_drivers_size * sizeof(*crypto_drivers)); 1035c0341432SJohn Baldwin 1036c0341432SJohn Baldwin crypto_drivers_size *= 2; 1037091d81d1SSam Leffler 1038091d81d1SSam Leffler free(crypto_drivers, M_CRYPTO_DATA); 1039091d81d1SSam Leffler crypto_drivers = newdrv; 1040091d81d1SSam Leffler } 1041091d81d1SSam Leffler 1042c0341432SJohn Baldwin cap->cc_hid = i; 1043c0341432SJohn Baldwin crypto_drivers[i] = cap; 1044c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 1045c0341432SJohn Baldwin 1046091d81d1SSam Leffler if (bootverbose) 1047d7d2f0d4SConrad Meyer printf("crypto: assign %s driver id %u, flags 0x%x\n", 10486810ad6fSSam Leffler device_get_nameunit(dev), i, flags); 1049091d81d1SSam Leffler 1050091d81d1SSam Leffler return i; 1051091d81d1SSam Leffler } 1052091d81d1SSam Leffler 10536810ad6fSSam Leffler /* 10546810ad6fSSam Leffler * Lookup a driver by name. We match against the full device 10556810ad6fSSam Leffler * name and unit, and against just the name. The latter gives 10566810ad6fSSam Leffler * us a simple widlcarding by device name. On success return the 10576810ad6fSSam Leffler * driver/hardware identifier; otherwise return -1. 10586810ad6fSSam Leffler */ 10596810ad6fSSam Leffler int 10606810ad6fSSam Leffler crypto_find_driver(const char *match) 1061091d81d1SSam Leffler { 1062c0341432SJohn Baldwin struct cryptocap *cap; 10636810ad6fSSam Leffler int i, len = strlen(match); 10646810ad6fSSam Leffler 10656810ad6fSSam Leffler CRYPTO_DRIVER_LOCK(); 1066c0341432SJohn Baldwin for (i = 0; i < crypto_drivers_size; i++) { 1067c0341432SJohn Baldwin if (crypto_drivers[i] == NULL) 10686810ad6fSSam Leffler continue; 1069c0341432SJohn Baldwin cap = crypto_drivers[i]; 1070c0341432SJohn Baldwin if (strncmp(match, device_get_nameunit(cap->cc_dev), len) == 0 || 1071c0341432SJohn Baldwin strncmp(match, device_get_name(cap->cc_dev), len) == 0) { 1072c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 1073c0341432SJohn Baldwin return (i); 1074c0341432SJohn Baldwin } 10756810ad6fSSam Leffler } 10766810ad6fSSam Leffler CRYPTO_DRIVER_UNLOCK(); 1077c0341432SJohn Baldwin return (-1); 10786810ad6fSSam Leffler } 10796810ad6fSSam Leffler 10806810ad6fSSam Leffler /* 10816810ad6fSSam Leffler * Return the device_t for the specified driver or NULL 10826810ad6fSSam Leffler * if the driver identifier is invalid. 10836810ad6fSSam Leffler */ 10846810ad6fSSam Leffler device_t 10856810ad6fSSam Leffler crypto_find_device_byhid(int hid) 10866810ad6fSSam Leffler { 1087c0341432SJohn Baldwin struct cryptocap *cap; 1088c0341432SJohn Baldwin device_t dev; 1089c0341432SJohn Baldwin 1090c0341432SJohn Baldwin dev = NULL; 1091c0341432SJohn Baldwin CRYPTO_DRIVER_LOCK(); 1092c0341432SJohn Baldwin cap = crypto_checkdriver(hid); 1093c0341432SJohn Baldwin if (cap != NULL) 1094c0341432SJohn Baldwin dev = cap->cc_dev; 1095c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 1096c0341432SJohn Baldwin return (dev); 10976810ad6fSSam Leffler } 10986810ad6fSSam Leffler 10996810ad6fSSam Leffler /* 11006810ad6fSSam Leffler * Return the device/driver capabilities. 11016810ad6fSSam Leffler */ 11026810ad6fSSam Leffler int 11036810ad6fSSam Leffler crypto_getcaps(int hid) 11046810ad6fSSam Leffler { 1105c0341432SJohn Baldwin struct cryptocap *cap; 1106c0341432SJohn Baldwin int flags; 1107c0341432SJohn Baldwin 1108c0341432SJohn Baldwin flags = 0; 1109c0341432SJohn Baldwin CRYPTO_DRIVER_LOCK(); 1110c0341432SJohn Baldwin cap = crypto_checkdriver(hid); 1111c0341432SJohn Baldwin if (cap != NULL) 1112c0341432SJohn Baldwin flags = cap->cc_flags; 1113c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 1114c0341432SJohn Baldwin return (flags); 1115091d81d1SSam Leffler } 1116091d81d1SSam Leffler 1117091d81d1SSam Leffler /* 1118091d81d1SSam Leffler * Register support for a key-related algorithm. This routine 1119091d81d1SSam Leffler * is called once for each algorithm supported a driver. 1120091d81d1SSam Leffler */ 1121091d81d1SSam Leffler int 1122d3d79e96SJohn Baldwin crypto_kregister(uint32_t driverid, int kalg, uint32_t flags) 1123091d81d1SSam Leffler { 1124091d81d1SSam Leffler struct cryptocap *cap; 1125091d81d1SSam Leffler int err; 1126091d81d1SSam Leffler 1127091d81d1SSam Leffler CRYPTO_DRIVER_LOCK(); 1128091d81d1SSam Leffler 1129091d81d1SSam Leffler cap = crypto_checkdriver(driverid); 1130091d81d1SSam Leffler if (cap != NULL && 1131091d81d1SSam Leffler (CRK_ALGORITM_MIN <= kalg && kalg <= CRK_ALGORITHM_MAX)) { 1132091d81d1SSam Leffler /* 1133091d81d1SSam Leffler * XXX Do some performance testing to determine placing. 1134091d81d1SSam Leffler * XXX We probably need an auxiliary data structure that 1135091d81d1SSam Leffler * XXX describes relative performances. 1136091d81d1SSam Leffler */ 1137091d81d1SSam Leffler 1138091d81d1SSam Leffler cap->cc_kalg[kalg] = flags | CRYPTO_ALG_FLAG_SUPPORTED; 1139091d81d1SSam Leffler if (bootverbose) 11406810ad6fSSam Leffler printf("crypto: %s registers key alg %u flags %u\n" 11416810ad6fSSam Leffler , device_get_nameunit(cap->cc_dev) 1142091d81d1SSam Leffler , kalg 1143091d81d1SSam Leffler , flags 1144091d81d1SSam Leffler ); 1145ecedef53SJohn Baldwin gone_in_dev(cap->cc_dev, 14, "asymmetric crypto"); 1146091d81d1SSam Leffler err = 0; 1147091d81d1SSam Leffler } else 1148091d81d1SSam Leffler err = EINVAL; 1149091d81d1SSam Leffler 1150091d81d1SSam Leffler CRYPTO_DRIVER_UNLOCK(); 1151091d81d1SSam Leffler return err; 1152091d81d1SSam Leffler } 1153091d81d1SSam Leffler 1154091d81d1SSam Leffler /* 1155091d81d1SSam Leffler * Unregister all algorithms associated with a crypto driver. 1156091d81d1SSam Leffler * If there are pending sessions using it, leave enough information 1157091d81d1SSam Leffler * around so that subsequent calls using those sessions will 1158091d81d1SSam Leffler * correctly detect the driver has been unregistered and reroute 1159091d81d1SSam Leffler * requests. 1160091d81d1SSam Leffler */ 1161091d81d1SSam Leffler int 1162d3d79e96SJohn Baldwin crypto_unregister_all(uint32_t driverid) 1163091d81d1SSam Leffler { 1164091d81d1SSam Leffler struct cryptocap *cap; 1165091d81d1SSam Leffler 1166091d81d1SSam Leffler CRYPTO_DRIVER_LOCK(); 1167091d81d1SSam Leffler cap = crypto_checkdriver(driverid); 1168c0341432SJohn Baldwin if (cap == NULL) { 1169091d81d1SSam Leffler CRYPTO_DRIVER_UNLOCK(); 1170c0341432SJohn Baldwin return (EINVAL); 1171c0341432SJohn Baldwin } 11726810ad6fSSam Leffler 1173c0341432SJohn Baldwin cap->cc_flags |= CRYPTOCAP_F_CLEANUP; 1174c0341432SJohn Baldwin crypto_drivers[driverid] = NULL; 1175c0341432SJohn Baldwin 1176c0341432SJohn Baldwin /* 1177c0341432SJohn Baldwin * XXX: This doesn't do anything to kick sessions that 1178c0341432SJohn Baldwin * have no pending operations. 1179c0341432SJohn Baldwin */ 1180c0341432SJohn Baldwin while (cap->cc_sessions != 0 || cap->cc_koperations != 0) 1181c0341432SJohn Baldwin mtx_sleep(cap, &crypto_drivers_mtx, 0, "cryunreg", 0); 1182c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 1183c0341432SJohn Baldwin cap_rele(cap); 1184c0341432SJohn Baldwin 1185c0341432SJohn Baldwin return (0); 1186091d81d1SSam Leffler } 1187091d81d1SSam Leffler 1188091d81d1SSam Leffler /* 1189091d81d1SSam Leffler * Clear blockage on a driver. The what parameter indicates whether 1190091d81d1SSam Leffler * the driver is now ready for cryptop's and/or cryptokop's. 1191091d81d1SSam Leffler */ 1192091d81d1SSam Leffler int 1193d3d79e96SJohn Baldwin crypto_unblock(uint32_t driverid, int what) 1194091d81d1SSam Leffler { 1195091d81d1SSam Leffler struct cryptocap *cap; 11963a865c82SPawel Jakub Dawidek int err; 1197091d81d1SSam Leffler 1198091d81d1SSam Leffler CRYPTO_Q_LOCK(); 1199091d81d1SSam Leffler cap = crypto_checkdriver(driverid); 1200091d81d1SSam Leffler if (cap != NULL) { 12013a865c82SPawel Jakub Dawidek if (what & CRYPTO_SYMQ) 1202091d81d1SSam Leffler cap->cc_qblocked = 0; 12033a865c82SPawel Jakub Dawidek if (what & CRYPTO_ASYMQ) 1204091d81d1SSam Leffler cap->cc_kqblocked = 0; 12053a865c82SPawel Jakub Dawidek if (crp_sleep) 12061a91ccccSSam Leffler wakeup_one(&crp_q); 1207091d81d1SSam Leffler err = 0; 1208091d81d1SSam Leffler } else 1209091d81d1SSam Leffler err = EINVAL; 1210091d81d1SSam Leffler CRYPTO_Q_UNLOCK(); 1211091d81d1SSam Leffler 1212091d81d1SSam Leffler return err; 1213091d81d1SSam Leffler } 1214091d81d1SSam Leffler 12159c0e3d3aSJohn Baldwin size_t 12169c0e3d3aSJohn Baldwin crypto_buffer_len(struct crypto_buffer *cb) 12179c0e3d3aSJohn Baldwin { 12189c0e3d3aSJohn Baldwin switch (cb->cb_type) { 12199c0e3d3aSJohn Baldwin case CRYPTO_BUF_CONTIG: 12209c0e3d3aSJohn Baldwin return (cb->cb_buf_len); 12219c0e3d3aSJohn Baldwin case CRYPTO_BUF_MBUF: 12229c0e3d3aSJohn Baldwin if (cb->cb_mbuf->m_flags & M_PKTHDR) 12239c0e3d3aSJohn Baldwin return (cb->cb_mbuf->m_pkthdr.len); 12249c0e3d3aSJohn Baldwin return (m_length(cb->cb_mbuf, NULL)); 1225e6f6d0c9SAlan Somers case CRYPTO_BUF_VMPAGE: 1226e6f6d0c9SAlan Somers return (cb->cb_vm_page_len); 12279c0e3d3aSJohn Baldwin case CRYPTO_BUF_UIO: 12289c0e3d3aSJohn Baldwin return (cb->cb_uio->uio_resid); 12299c0e3d3aSJohn Baldwin default: 12309c0e3d3aSJohn Baldwin return (0); 12319c0e3d3aSJohn Baldwin } 12329c0e3d3aSJohn Baldwin } 12339c0e3d3aSJohn Baldwin 1234c0341432SJohn Baldwin #ifdef INVARIANTS 1235c0341432SJohn Baldwin /* Various sanity checks on crypto requests. */ 1236c0341432SJohn Baldwin static void 12379c0e3d3aSJohn Baldwin cb_sanity(struct crypto_buffer *cb, const char *name) 12389c0e3d3aSJohn Baldwin { 12399c0e3d3aSJohn Baldwin KASSERT(cb->cb_type > CRYPTO_BUF_NONE && cb->cb_type <= CRYPTO_BUF_LAST, 12409c0e3d3aSJohn Baldwin ("incoming crp with invalid %s buffer type", name)); 1241e6f6d0c9SAlan Somers switch (cb->cb_type) { 1242e6f6d0c9SAlan Somers case CRYPTO_BUF_CONTIG: 12439c0e3d3aSJohn Baldwin KASSERT(cb->cb_buf_len >= 0, 12449c0e3d3aSJohn Baldwin ("incoming crp with -ve %s buffer length", name)); 1245e6f6d0c9SAlan Somers break; 1246e6f6d0c9SAlan Somers case CRYPTO_BUF_VMPAGE: 1247e6f6d0c9SAlan Somers KASSERT(CRYPTO_HAS_VMPAGE, 1248e6f6d0c9SAlan Somers ("incoming crp uses dmap on supported arch")); 1249e6f6d0c9SAlan Somers KASSERT(cb->cb_vm_page_len >= 0, 1250e6f6d0c9SAlan Somers ("incoming crp with -ve %s buffer length", name)); 1251e6f6d0c9SAlan Somers KASSERT(cb->cb_vm_page_offset >= 0, 1252e6f6d0c9SAlan Somers ("incoming crp with -ve %s buffer offset", name)); 1253e6f6d0c9SAlan Somers KASSERT(cb->cb_vm_page_offset < PAGE_SIZE, 1254e6f6d0c9SAlan Somers ("incoming crp with %s buffer offset greater than page size" 1255e6f6d0c9SAlan Somers , name)); 1256e6f6d0c9SAlan Somers break; 1257e6f6d0c9SAlan Somers default: 1258e6f6d0c9SAlan Somers break; 1259e6f6d0c9SAlan Somers } 12609c0e3d3aSJohn Baldwin } 12619c0e3d3aSJohn Baldwin 12629c0e3d3aSJohn Baldwin static void 1263c0341432SJohn Baldwin crp_sanity(struct cryptop *crp) 1264c0341432SJohn Baldwin { 1265c0341432SJohn Baldwin struct crypto_session_params *csp; 12669c0e3d3aSJohn Baldwin struct crypto_buffer *out; 12679c0e3d3aSJohn Baldwin size_t ilen, len, olen; 1268c0341432SJohn Baldwin 1269c0341432SJohn Baldwin KASSERT(crp->crp_session != NULL, ("incoming crp without a session")); 12709c0e3d3aSJohn Baldwin KASSERT(crp->crp_obuf.cb_type >= CRYPTO_BUF_NONE && 12719c0e3d3aSJohn Baldwin crp->crp_obuf.cb_type <= CRYPTO_BUF_LAST, 12729c0e3d3aSJohn Baldwin ("incoming crp with invalid output buffer type")); 1273c0341432SJohn Baldwin KASSERT(crp->crp_etype == 0, ("incoming crp with error")); 1274c0341432SJohn Baldwin KASSERT(!(crp->crp_flags & CRYPTO_F_DONE), 1275c0341432SJohn Baldwin ("incoming crp already done")); 1276c0341432SJohn Baldwin 1277c0341432SJohn Baldwin csp = &crp->crp_session->csp; 12789c0e3d3aSJohn Baldwin cb_sanity(&crp->crp_buf, "input"); 12799c0e3d3aSJohn Baldwin ilen = crypto_buffer_len(&crp->crp_buf); 12809c0e3d3aSJohn Baldwin olen = ilen; 12819c0e3d3aSJohn Baldwin out = NULL; 12829c0e3d3aSJohn Baldwin if (csp->csp_flags & CSP_F_SEPARATE_OUTPUT) { 12839c0e3d3aSJohn Baldwin if (crp->crp_obuf.cb_type != CRYPTO_BUF_NONE) { 12849c0e3d3aSJohn Baldwin cb_sanity(&crp->crp_obuf, "output"); 12859c0e3d3aSJohn Baldwin out = &crp->crp_obuf; 12869c0e3d3aSJohn Baldwin olen = crypto_buffer_len(out); 12879c0e3d3aSJohn Baldwin } 12889c0e3d3aSJohn Baldwin } else 12899c0e3d3aSJohn Baldwin KASSERT(crp->crp_obuf.cb_type == CRYPTO_BUF_NONE, 12909c0e3d3aSJohn Baldwin ("incoming crp with separate output buffer " 12919c0e3d3aSJohn Baldwin "but no session support")); 12929c0e3d3aSJohn Baldwin 1293c0341432SJohn Baldwin switch (csp->csp_mode) { 1294c0341432SJohn Baldwin case CSP_MODE_COMPRESS: 1295c0341432SJohn Baldwin KASSERT(crp->crp_op == CRYPTO_OP_COMPRESS || 1296c0341432SJohn Baldwin crp->crp_op == CRYPTO_OP_DECOMPRESS, 1297c0341432SJohn Baldwin ("invalid compression op %x", crp->crp_op)); 1298c0341432SJohn Baldwin break; 1299c0341432SJohn Baldwin case CSP_MODE_CIPHER: 1300c0341432SJohn Baldwin KASSERT(crp->crp_op == CRYPTO_OP_ENCRYPT || 1301c0341432SJohn Baldwin crp->crp_op == CRYPTO_OP_DECRYPT, 1302c0341432SJohn Baldwin ("invalid cipher op %x", crp->crp_op)); 1303c0341432SJohn Baldwin break; 1304c0341432SJohn Baldwin case CSP_MODE_DIGEST: 1305c0341432SJohn Baldwin KASSERT(crp->crp_op == CRYPTO_OP_COMPUTE_DIGEST || 1306c0341432SJohn Baldwin crp->crp_op == CRYPTO_OP_VERIFY_DIGEST, 1307c0341432SJohn Baldwin ("invalid digest op %x", crp->crp_op)); 1308c0341432SJohn Baldwin break; 1309c0341432SJohn Baldwin case CSP_MODE_AEAD: 1310c0341432SJohn Baldwin KASSERT(crp->crp_op == 1311c0341432SJohn Baldwin (CRYPTO_OP_ENCRYPT | CRYPTO_OP_COMPUTE_DIGEST) || 1312c0341432SJohn Baldwin crp->crp_op == 1313c0341432SJohn Baldwin (CRYPTO_OP_DECRYPT | CRYPTO_OP_VERIFY_DIGEST), 1314c0341432SJohn Baldwin ("invalid AEAD op %x", crp->crp_op)); 1315c0341432SJohn Baldwin KASSERT(crp->crp_flags & CRYPTO_F_IV_SEPARATE, 1316*fc8fc743SJohn Baldwin ("AEAD without a separate IV")); 1317c0341432SJohn Baldwin break; 1318c0341432SJohn Baldwin case CSP_MODE_ETA: 1319c0341432SJohn Baldwin KASSERT(crp->crp_op == 1320c0341432SJohn Baldwin (CRYPTO_OP_ENCRYPT | CRYPTO_OP_COMPUTE_DIGEST) || 1321c0341432SJohn Baldwin crp->crp_op == 1322c0341432SJohn Baldwin (CRYPTO_OP_DECRYPT | CRYPTO_OP_VERIFY_DIGEST), 1323c0341432SJohn Baldwin ("invalid ETA op %x", crp->crp_op)); 1324c0341432SJohn Baldwin break; 1325c0341432SJohn Baldwin } 1326c0341432SJohn Baldwin if (csp->csp_mode == CSP_MODE_AEAD || csp->csp_mode == CSP_MODE_ETA) { 13279b774dc0SJohn Baldwin if (crp->crp_aad == NULL) { 1328c0341432SJohn Baldwin KASSERT(crp->crp_aad_start == 0 || 13299c0e3d3aSJohn Baldwin crp->crp_aad_start < ilen, 1330c0341432SJohn Baldwin ("invalid AAD start")); 13319b774dc0SJohn Baldwin KASSERT(crp->crp_aad_length != 0 || 13329b774dc0SJohn Baldwin crp->crp_aad_start == 0, 1333c0341432SJohn Baldwin ("AAD with zero length and non-zero start")); 1334c0341432SJohn Baldwin KASSERT(crp->crp_aad_length == 0 || 13359c0e3d3aSJohn Baldwin crp->crp_aad_start + crp->crp_aad_length <= ilen, 1336c0341432SJohn Baldwin ("AAD outside input length")); 1337c0341432SJohn Baldwin } else { 13389b774dc0SJohn Baldwin KASSERT(csp->csp_flags & CSP_F_SEPARATE_AAD, 13399b774dc0SJohn Baldwin ("session doesn't support separate AAD buffer")); 13409b774dc0SJohn Baldwin KASSERT(crp->crp_aad_start == 0, 13419b774dc0SJohn Baldwin ("separate AAD buffer with non-zero AAD start")); 13429b774dc0SJohn Baldwin KASSERT(crp->crp_aad_length != 0, 13439b774dc0SJohn Baldwin ("separate AAD buffer with zero length")); 13449b774dc0SJohn Baldwin } 13459b774dc0SJohn Baldwin } else { 13469b774dc0SJohn Baldwin KASSERT(crp->crp_aad == NULL && crp->crp_aad_start == 0 && 13479b774dc0SJohn Baldwin crp->crp_aad_length == 0, 1348c0341432SJohn Baldwin ("AAD region in request not supporting AAD")); 1349c0341432SJohn Baldwin } 1350c0341432SJohn Baldwin if (csp->csp_ivlen == 0) { 135129fe41ddSJohn Baldwin KASSERT((crp->crp_flags & CRYPTO_F_IV_SEPARATE) == 0, 135229fe41ddSJohn Baldwin ("IV_SEPARATE set when IV isn't used")); 1353c0341432SJohn Baldwin KASSERT(crp->crp_iv_start == 0, 1354c0341432SJohn Baldwin ("crp_iv_start set when IV isn't used")); 1355c0341432SJohn Baldwin } else if (crp->crp_flags & CRYPTO_F_IV_SEPARATE) { 1356c0341432SJohn Baldwin KASSERT(crp->crp_iv_start == 0, 1357c0341432SJohn Baldwin ("IV_SEPARATE used with non-zero IV start")); 1358c0341432SJohn Baldwin } else { 13599c0e3d3aSJohn Baldwin KASSERT(crp->crp_iv_start < ilen, 1360c0341432SJohn Baldwin ("invalid IV start")); 13619c0e3d3aSJohn Baldwin KASSERT(crp->crp_iv_start + csp->csp_ivlen <= ilen, 13629c0e3d3aSJohn Baldwin ("IV outside buffer length")); 1363c0341432SJohn Baldwin } 13649c0e3d3aSJohn Baldwin /* XXX: payload_start of 0 should always be < ilen? */ 1365c0341432SJohn Baldwin KASSERT(crp->crp_payload_start == 0 || 13669c0e3d3aSJohn Baldwin crp->crp_payload_start < ilen, 1367c0341432SJohn Baldwin ("invalid payload start")); 1368c0341432SJohn Baldwin KASSERT(crp->crp_payload_start + crp->crp_payload_length <= 13699c0e3d3aSJohn Baldwin ilen, ("payload outside input buffer")); 13709c0e3d3aSJohn Baldwin if (out == NULL) { 13719c0e3d3aSJohn Baldwin KASSERT(crp->crp_payload_output_start == 0, 13729c0e3d3aSJohn Baldwin ("payload output start non-zero without output buffer")); 13739c0e3d3aSJohn Baldwin } else { 13749c0e3d3aSJohn Baldwin KASSERT(crp->crp_payload_output_start < olen, 13759c0e3d3aSJohn Baldwin ("invalid payload output start")); 13769c0e3d3aSJohn Baldwin KASSERT(crp->crp_payload_output_start + 13779c0e3d3aSJohn Baldwin crp->crp_payload_length <= olen, 13789c0e3d3aSJohn Baldwin ("payload outside output buffer")); 13799c0e3d3aSJohn Baldwin } 1380c0341432SJohn Baldwin if (csp->csp_mode == CSP_MODE_DIGEST || 1381c0341432SJohn Baldwin csp->csp_mode == CSP_MODE_AEAD || csp->csp_mode == CSP_MODE_ETA) { 13829c0e3d3aSJohn Baldwin if (crp->crp_op & CRYPTO_OP_VERIFY_DIGEST) 13839c0e3d3aSJohn Baldwin len = ilen; 13849c0e3d3aSJohn Baldwin else 13859c0e3d3aSJohn Baldwin len = olen; 1386c0341432SJohn Baldwin KASSERT(crp->crp_digest_start == 0 || 13879c0e3d3aSJohn Baldwin crp->crp_digest_start < len, 1388c0341432SJohn Baldwin ("invalid digest start")); 1389c0341432SJohn Baldwin /* XXX: For the mlen == 0 case this check isn't perfect. */ 13909c0e3d3aSJohn Baldwin KASSERT(crp->crp_digest_start + csp->csp_auth_mlen <= len, 13919c0e3d3aSJohn Baldwin ("digest outside buffer")); 1392c0341432SJohn Baldwin } else { 1393c0341432SJohn Baldwin KASSERT(crp->crp_digest_start == 0, 1394c0341432SJohn Baldwin ("non-zero digest start for request without a digest")); 1395c0341432SJohn Baldwin } 1396c0341432SJohn Baldwin if (csp->csp_cipher_klen != 0) 1397c0341432SJohn Baldwin KASSERT(csp->csp_cipher_key != NULL || 1398c0341432SJohn Baldwin crp->crp_cipher_key != NULL, 1399c0341432SJohn Baldwin ("cipher request without a key")); 1400c0341432SJohn Baldwin if (csp->csp_auth_klen != 0) 1401c0341432SJohn Baldwin KASSERT(csp->csp_auth_key != NULL || crp->crp_auth_key != NULL, 1402c0341432SJohn Baldwin ("auth request without a key")); 1403c0341432SJohn Baldwin KASSERT(crp->crp_callback != NULL, ("incoming crp without callback")); 1404c0341432SJohn Baldwin } 1405c0341432SJohn Baldwin #endif 1406c0341432SJohn Baldwin 140768f6800cSMark Johnston static int 140868f6800cSMark Johnston crypto_dispatch_one(struct cryptop *crp, int hint) 1409091d81d1SSam Leffler { 14104acae0acSPawel Jakub Dawidek struct cryptocap *cap; 14114acae0acSPawel Jakub Dawidek int result; 1412091d81d1SSam Leffler 1413c0341432SJohn Baldwin #ifdef INVARIANTS 1414c0341432SJohn Baldwin crp_sanity(crp); 1415c0341432SJohn Baldwin #endif 14167290cb47SMark Johnston CRYPTOSTAT_INC(cs_ops); 14177d1853eeSSam Leffler 141898d788c8SMark Johnston crp->crp_retw_id = crp->crp_session->id % crypto_workers_num; 1419de2b2c90SFabien Thomas 142068f6800cSMark Johnston /* 142168f6800cSMark Johnston * Caller marked the request to be processed immediately; dispatch it 142268f6800cSMark Johnston * directly to the driver unless the driver is currently blocked, in 142368f6800cSMark Johnston * which case it is queued for deferred dispatch. 142468f6800cSMark Johnston */ 142568f6800cSMark Johnston cap = crp->crp_session->cap; 142668f6800cSMark Johnston if (!atomic_load_int(&cap->cc_qblocked)) { 142768f6800cSMark Johnston result = crypto_invoke(cap, crp, hint); 142868f6800cSMark Johnston if (result != ERESTART) 142968f6800cSMark Johnston return (result); 143068f6800cSMark Johnston 143168f6800cSMark Johnston /* 143268f6800cSMark Johnston * The driver ran out of resources, put the request on the 143368f6800cSMark Johnston * queue. 143468f6800cSMark Johnston */ 143568f6800cSMark Johnston } 143668f6800cSMark Johnston crypto_batch_enqueue(crp); 143768f6800cSMark Johnston return (0); 143868f6800cSMark Johnston } 143968f6800cSMark Johnston 144068f6800cSMark Johnston int 144168f6800cSMark Johnston crypto_dispatch(struct cryptop *crp) 144268f6800cSMark Johnston { 144368f6800cSMark Johnston return (crypto_dispatch_one(crp, 0)); 144468f6800cSMark Johnston } 144568f6800cSMark Johnston 144668f6800cSMark Johnston int 144768f6800cSMark Johnston crypto_dispatch_async(struct cryptop *crp, int flags) 144868f6800cSMark Johnston { 144939bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 145039bbca6fSFabien Thomas 145168f6800cSMark Johnston if (!CRYPTO_SESS_SYNC(crp->crp_session)) { 145268f6800cSMark Johnston /* 145368f6800cSMark Johnston * The driver issues completions asynchonously, don't bother 145468f6800cSMark Johnston * deferring dispatch to a worker thread. 145568f6800cSMark Johnston */ 145668f6800cSMark Johnston return (crypto_dispatch(crp)); 145768f6800cSMark Johnston } 145839bbca6fSFabien Thomas 145968f6800cSMark Johnston #ifdef INVARIANTS 146068f6800cSMark Johnston crp_sanity(crp); 146168f6800cSMark Johnston #endif 146268f6800cSMark Johnston CRYPTOSTAT_INC(cs_ops); 146368f6800cSMark Johnston 146468f6800cSMark Johnston crp->crp_retw_id = crp->crp_session->id % crypto_workers_num; 146568f6800cSMark Johnston if ((flags & CRYPTO_ASYNC_ORDERED) != 0) { 146668f6800cSMark Johnston crp->crp_flags |= CRYPTO_F_ASYNC_ORDERED; 146768f6800cSMark Johnston ret_worker = CRYPTO_RETW(crp->crp_retw_id); 146839bbca6fSFabien Thomas CRYPTO_RETW_LOCK(ret_worker); 146939bbca6fSFabien Thomas crp->crp_seq = ret_worker->reorder_ops++; 147039bbca6fSFabien Thomas CRYPTO_RETW_UNLOCK(ret_worker); 147139bbca6fSFabien Thomas } 147239bbca6fSFabien Thomas TASK_INIT(&crp->crp_task, 0, crypto_task_invoke, crp); 147339bbca6fSFabien Thomas taskqueue_enqueue(crypto_tq, &crp->crp_task); 147439bbca6fSFabien Thomas return (0); 147539bbca6fSFabien Thomas } 14764acae0acSPawel Jakub Dawidek 147768f6800cSMark Johnston void 147868f6800cSMark Johnston crypto_dispatch_batch(struct cryptopq *crpq, int flags) 147968f6800cSMark Johnston { 148068f6800cSMark Johnston struct cryptop *crp; 148168f6800cSMark Johnston int hint; 148268f6800cSMark Johnston 148368f6800cSMark Johnston while ((crp = TAILQ_FIRST(crpq)) != NULL) { 148468f6800cSMark Johnston hint = TAILQ_NEXT(crp, crp_next) != NULL ? CRYPTO_HINT_MORE : 0; 148568f6800cSMark Johnston TAILQ_REMOVE(crpq, crp, crp_next); 148668f6800cSMark Johnston if (crypto_dispatch_one(crp, hint) != 0) 148739bbca6fSFabien Thomas crypto_batch_enqueue(crp); 148868f6800cSMark Johnston } 148939bbca6fSFabien Thomas } 149039bbca6fSFabien Thomas 149168f6800cSMark Johnston static void 149239bbca6fSFabien Thomas crypto_batch_enqueue(struct cryptop *crp) 149339bbca6fSFabien Thomas { 149439bbca6fSFabien Thomas 14954acae0acSPawel Jakub Dawidek CRYPTO_Q_LOCK(); 14964acae0acSPawel Jakub Dawidek TAILQ_INSERT_TAIL(&crp_q, crp, crp_next); 14973a865c82SPawel Jakub Dawidek if (crp_sleep) 14983a865c82SPawel Jakub Dawidek wakeup_one(&crp_q); 14993569ae7fSSam Leffler CRYPTO_Q_UNLOCK(); 1500091d81d1SSam Leffler } 1501091d81d1SSam Leffler 1502091d81d1SSam Leffler /* 1503091d81d1SSam Leffler * Add an asymetric crypto request to a queue, 1504091d81d1SSam Leffler * to be processed by the kernel thread. 1505091d81d1SSam Leffler */ 1506091d81d1SSam Leffler int 1507091d81d1SSam Leffler crypto_kdispatch(struct cryptkop *krp) 1508091d81d1SSam Leffler { 15096810ad6fSSam Leffler int error; 1510091d81d1SSam Leffler 15117290cb47SMark Johnston CRYPTOSTAT_INC(cs_kops); 15127d1853eeSSam Leffler 1513c0341432SJohn Baldwin krp->krp_cap = NULL; 1514c0341432SJohn Baldwin error = crypto_kinvoke(krp); 15156810ad6fSSam Leffler if (error == ERESTART) { 1516091d81d1SSam Leffler CRYPTO_Q_LOCK(); 15174acae0acSPawel Jakub Dawidek TAILQ_INSERT_TAIL(&crp_kq, krp, krp_next); 15183a865c82SPawel Jakub Dawidek if (crp_sleep) 15193a865c82SPawel Jakub Dawidek wakeup_one(&crp_q); 1520091d81d1SSam Leffler CRYPTO_Q_UNLOCK(); 15216810ad6fSSam Leffler error = 0; 15226810ad6fSSam Leffler } 15236810ad6fSSam Leffler return error; 1524091d81d1SSam Leffler } 1525091d81d1SSam Leffler 1526091d81d1SSam Leffler /* 15276810ad6fSSam Leffler * Verify a driver is suitable for the specified operation. 15286810ad6fSSam Leffler */ 15296810ad6fSSam Leffler static __inline int 15306810ad6fSSam Leffler kdriver_suitable(const struct cryptocap *cap, const struct cryptkop *krp) 15316810ad6fSSam Leffler { 15326810ad6fSSam Leffler return (cap->cc_kalg[krp->krp_op] & CRYPTO_ALG_FLAG_SUPPORTED) != 0; 15336810ad6fSSam Leffler } 15346810ad6fSSam Leffler 15356810ad6fSSam Leffler /* 15366810ad6fSSam Leffler * Select a driver for an asym operation. The driver must 15376810ad6fSSam Leffler * support the necessary algorithm. The caller can constrain 15386810ad6fSSam Leffler * which device is selected with the flags parameter. The 15396810ad6fSSam Leffler * algorithm we use here is pretty stupid; just use the first 15406810ad6fSSam Leffler * driver that supports the algorithms we need. If there are 15416810ad6fSSam Leffler * multiple suitable drivers we choose the driver with the 15426810ad6fSSam Leffler * fewest active operations. We prefer hardware-backed 15436810ad6fSSam Leffler * drivers to software ones when either may be used. 15446810ad6fSSam Leffler */ 15456810ad6fSSam Leffler static struct cryptocap * 15466810ad6fSSam Leffler crypto_select_kdriver(const struct cryptkop *krp, int flags) 15476810ad6fSSam Leffler { 1548151ba793SAlexander Kabaev struct cryptocap *cap, *best; 15496810ad6fSSam Leffler int match, hid; 15506810ad6fSSam Leffler 15516810ad6fSSam Leffler CRYPTO_DRIVER_ASSERT(); 15526810ad6fSSam Leffler 15536810ad6fSSam Leffler /* 15546810ad6fSSam Leffler * Look first for hardware crypto devices if permitted. 15556810ad6fSSam Leffler */ 15566810ad6fSSam Leffler if (flags & CRYPTOCAP_F_HARDWARE) 15576810ad6fSSam Leffler match = CRYPTOCAP_F_HARDWARE; 15586810ad6fSSam Leffler else 15596810ad6fSSam Leffler match = CRYPTOCAP_F_SOFTWARE; 15606810ad6fSSam Leffler best = NULL; 15616810ad6fSSam Leffler again: 1562c0341432SJohn Baldwin for (hid = 0; hid < crypto_drivers_size; hid++) { 15636810ad6fSSam Leffler /* 1564c0341432SJohn Baldwin * If there is no driver for this slot, or the driver 1565c0341432SJohn Baldwin * is not appropriate (hardware or software based on 1566c0341432SJohn Baldwin * match), then skip. 15676810ad6fSSam Leffler */ 1568c0341432SJohn Baldwin cap = crypto_drivers[hid]; 1569e7f6b6cfSJohn Baldwin if (cap == NULL || 15706810ad6fSSam Leffler (cap->cc_flags & match) == 0) 15716810ad6fSSam Leffler continue; 15726810ad6fSSam Leffler 15736810ad6fSSam Leffler /* verify all the algorithms are supported. */ 15746810ad6fSSam Leffler if (kdriver_suitable(cap, krp)) { 15756810ad6fSSam Leffler if (best == NULL || 15766810ad6fSSam Leffler cap->cc_koperations < best->cc_koperations) 15776810ad6fSSam Leffler best = cap; 15786810ad6fSSam Leffler } 15796810ad6fSSam Leffler } 15806810ad6fSSam Leffler if (best != NULL) 15816810ad6fSSam Leffler return best; 15826810ad6fSSam Leffler if (match == CRYPTOCAP_F_HARDWARE && (flags & CRYPTOCAP_F_SOFTWARE)) { 15836810ad6fSSam Leffler /* sort of an Algol 68-style for loop */ 15846810ad6fSSam Leffler match = CRYPTOCAP_F_SOFTWARE; 15856810ad6fSSam Leffler goto again; 15866810ad6fSSam Leffler } 15876810ad6fSSam Leffler return best; 15886810ad6fSSam Leffler } 15896810ad6fSSam Leffler 15906810ad6fSSam Leffler /* 1591c0341432SJohn Baldwin * Choose a driver for an asymmetric crypto request. 1592091d81d1SSam Leffler */ 1593c0341432SJohn Baldwin static struct cryptocap * 1594c0341432SJohn Baldwin crypto_lookup_kdriver(struct cryptkop *krp) 1595091d81d1SSam Leffler { 1596c0341432SJohn Baldwin struct cryptocap *cap; 1597c0341432SJohn Baldwin uint32_t crid; 1598091d81d1SSam Leffler 1599c0341432SJohn Baldwin /* If this request is requeued, it might already have a driver. */ 1600c0341432SJohn Baldwin cap = krp->krp_cap; 1601c0341432SJohn Baldwin if (cap != NULL) 1602c0341432SJohn Baldwin return (cap); 1603091d81d1SSam Leffler 1604c0341432SJohn Baldwin /* Use krp_crid to choose a driver. */ 1605c0341432SJohn Baldwin crid = krp->krp_crid; 16066810ad6fSSam Leffler if ((crid & (CRYPTOCAP_F_HARDWARE | CRYPTOCAP_F_SOFTWARE)) == 0) { 16076810ad6fSSam Leffler cap = crypto_checkdriver(crid); 16086810ad6fSSam Leffler if (cap != NULL) { 16096810ad6fSSam Leffler /* 1610c0341432SJohn Baldwin * Driver present, it must support the 1611c0341432SJohn Baldwin * necessary algorithm and, if s/w drivers are 1612c0341432SJohn Baldwin * excluded, it must be registered as 1613c0341432SJohn Baldwin * hardware-backed. 16146810ad6fSSam Leffler */ 16156810ad6fSSam Leffler if (!kdriver_suitable(cap, krp) || 16166810ad6fSSam Leffler (!crypto_devallowsoft && 16176810ad6fSSam Leffler (cap->cc_flags & CRYPTOCAP_F_HARDWARE) == 0)) 16186810ad6fSSam Leffler cap = NULL; 16194acae0acSPawel Jakub Dawidek } 16206810ad6fSSam Leffler } else { 16216810ad6fSSam Leffler /* 16226810ad6fSSam Leffler * No requested driver; select based on crid flags. 16236810ad6fSSam Leffler */ 16246810ad6fSSam Leffler if (!crypto_devallowsoft) /* NB: disallow s/w drivers */ 16256810ad6fSSam Leffler crid &= ~CRYPTOCAP_F_SOFTWARE; 16266810ad6fSSam Leffler cap = crypto_select_kdriver(krp, crid); 16274acae0acSPawel Jakub Dawidek } 1628c0341432SJohn Baldwin 1629c0341432SJohn Baldwin if (cap != NULL) { 1630c0341432SJohn Baldwin krp->krp_cap = cap_ref(cap); 1631c0341432SJohn Baldwin krp->krp_hid = cap->cc_hid; 1632c0341432SJohn Baldwin } 1633c0341432SJohn Baldwin return (cap); 1634c0341432SJohn Baldwin } 1635c0341432SJohn Baldwin 1636c0341432SJohn Baldwin /* 1637c0341432SJohn Baldwin * Dispatch an asymmetric crypto request. 1638c0341432SJohn Baldwin */ 1639c0341432SJohn Baldwin static int 1640c0341432SJohn Baldwin crypto_kinvoke(struct cryptkop *krp) 1641c0341432SJohn Baldwin { 1642c0341432SJohn Baldwin struct cryptocap *cap = NULL; 1643c0341432SJohn Baldwin int error; 1644c0341432SJohn Baldwin 1645c0341432SJohn Baldwin KASSERT(krp != NULL, ("%s: krp == NULL", __func__)); 1646c0341432SJohn Baldwin KASSERT(krp->krp_callback != NULL, 1647c0341432SJohn Baldwin ("%s: krp->crp_callback == NULL", __func__)); 1648c0341432SJohn Baldwin 1649c0341432SJohn Baldwin CRYPTO_DRIVER_LOCK(); 1650c0341432SJohn Baldwin cap = crypto_lookup_kdriver(krp); 1651c0341432SJohn Baldwin if (cap == NULL) { 1652c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 1653c0341432SJohn Baldwin krp->krp_status = ENODEV; 1654c0341432SJohn Baldwin crypto_kdone(krp); 1655c0341432SJohn Baldwin return (0); 1656c0341432SJohn Baldwin } 1657c0341432SJohn Baldwin 1658c0341432SJohn Baldwin /* 1659c0341432SJohn Baldwin * If the device is blocked, return ERESTART to requeue it. 1660c0341432SJohn Baldwin */ 1661c0341432SJohn Baldwin if (cap->cc_kqblocked) { 1662c0341432SJohn Baldwin /* 1663c0341432SJohn Baldwin * XXX: Previously this set krp_status to ERESTART and 1664c0341432SJohn Baldwin * invoked crypto_kdone but the caller would still 1665c0341432SJohn Baldwin * requeue it. 1666c0341432SJohn Baldwin */ 1667c0341432SJohn Baldwin CRYPTO_DRIVER_UNLOCK(); 1668c0341432SJohn Baldwin return (ERESTART); 1669c0341432SJohn Baldwin } 1670c0341432SJohn Baldwin 16714acae0acSPawel Jakub Dawidek cap->cc_koperations++; 16724acae0acSPawel Jakub Dawidek CRYPTO_DRIVER_UNLOCK(); 16736810ad6fSSam Leffler error = CRYPTODEV_KPROCESS(cap->cc_dev, krp, 0); 16744acae0acSPawel Jakub Dawidek if (error == ERESTART) { 1675c0341432SJohn Baldwin CRYPTO_DRIVER_LOCK(); 16764acae0acSPawel Jakub Dawidek cap->cc_koperations--; 16774acae0acSPawel Jakub Dawidek CRYPTO_DRIVER_UNLOCK(); 16784acae0acSPawel Jakub Dawidek return (error); 16794acae0acSPawel Jakub Dawidek } 1680091d81d1SSam Leffler 1681c0341432SJohn Baldwin KASSERT(error == 0, ("error %d returned from crypto_kprocess", error)); 1682c0341432SJohn Baldwin return (0); 1683091d81d1SSam Leffler } 1684091d81d1SSam Leffler 168539bbca6fSFabien Thomas static void 168639bbca6fSFabien Thomas crypto_task_invoke(void *ctx, int pending) 168739bbca6fSFabien Thomas { 168839bbca6fSFabien Thomas struct cryptocap *cap; 168939bbca6fSFabien Thomas struct cryptop *crp; 1690c0341432SJohn Baldwin int result; 169139bbca6fSFabien Thomas 169239bbca6fSFabien Thomas crp = (struct cryptop *)ctx; 1693c0341432SJohn Baldwin cap = crp->crp_session->cap; 169439bbca6fSFabien Thomas result = crypto_invoke(cap, crp, 0); 169539bbca6fSFabien Thomas if (result == ERESTART) 169639bbca6fSFabien Thomas crypto_batch_enqueue(crp); 169739bbca6fSFabien Thomas } 169839bbca6fSFabien Thomas 1699091d81d1SSam Leffler /* 1700091d81d1SSam Leffler * Dispatch a crypto request to the appropriate crypto devices. 1701091d81d1SSam Leffler */ 1702091d81d1SSam Leffler static int 17034acae0acSPawel Jakub Dawidek crypto_invoke(struct cryptocap *cap, struct cryptop *crp, int hint) 1704091d81d1SSam Leffler { 17054acae0acSPawel Jakub Dawidek 17064acae0acSPawel Jakub Dawidek KASSERT(crp != NULL, ("%s: crp == NULL", __func__)); 17074acae0acSPawel Jakub Dawidek KASSERT(crp->crp_callback != NULL, 17084acae0acSPawel Jakub Dawidek ("%s: crp->crp_callback == NULL", __func__)); 1709c0341432SJohn Baldwin KASSERT(crp->crp_session != NULL, 1710c0341432SJohn Baldwin ("%s: crp->crp_session == NULL", __func__)); 1711091d81d1SSam Leffler 17124acae0acSPawel Jakub Dawidek if (cap->cc_flags & CRYPTOCAP_F_CLEANUP) { 1713c0341432SJohn Baldwin struct crypto_session_params csp; 17141b0909d5SConrad Meyer crypto_session_t nses; 1715091d81d1SSam Leffler 1716091d81d1SSam Leffler /* 1717091d81d1SSam Leffler * Driver has unregistered; migrate the session and return 1718091d81d1SSam Leffler * an error to the caller so they'll resubmit the op. 17194acae0acSPawel Jakub Dawidek * 17204acae0acSPawel Jakub Dawidek * XXX: What if there are more already queued requests for this 17214acae0acSPawel Jakub Dawidek * session? 1722c0341432SJohn Baldwin * 1723c0341432SJohn Baldwin * XXX: Real solution is to make sessions refcounted 1724c0341432SJohn Baldwin * and force callers to hold a reference when 1725c0341432SJohn Baldwin * assigning to crp_session. Could maybe change 1726c0341432SJohn Baldwin * crypto_getreq to accept a session pointer to make 1727c0341432SJohn Baldwin * that work. Alternatively, we could abandon the 1728c0341432SJohn Baldwin * notion of rewriting crp_session in requests forcing 1729c0341432SJohn Baldwin * the caller to deal with allocating a new session. 1730c0341432SJohn Baldwin * Perhaps provide a method to allow a crp's session to 1731c0341432SJohn Baldwin * be swapped that callers could use. 1732091d81d1SSam Leffler */ 1733c0341432SJohn Baldwin csp = crp->crp_session->csp; 17341b0909d5SConrad Meyer crypto_freesession(crp->crp_session); 17354acae0acSPawel Jakub Dawidek 1736c0341432SJohn Baldwin /* 1737c0341432SJohn Baldwin * XXX: Key pointers may no longer be valid. If we 1738c0341432SJohn Baldwin * really want to support this we need to define the 1739c0341432SJohn Baldwin * KPI such that 'csp' is required to be valid for the 1740c0341432SJohn Baldwin * duration of a session by the caller perhaps. 1741c0341432SJohn Baldwin * 1742c0341432SJohn Baldwin * XXX: If the keys have been changed this will reuse 1743c0341432SJohn Baldwin * the old keys. This probably suggests making 1744c0341432SJohn Baldwin * rekeying more explicit and updating the key 1745c0341432SJohn Baldwin * pointers in 'csp' when the keys change. 1746c0341432SJohn Baldwin */ 1747c0341432SJohn Baldwin if (crypto_newsession(&nses, &csp, 17486810ad6fSSam Leffler CRYPTOCAP_F_HARDWARE | CRYPTOCAP_F_SOFTWARE) == 0) 17491b0909d5SConrad Meyer crp->crp_session = nses; 1750091d81d1SSam Leffler 1751091d81d1SSam Leffler crp->crp_etype = EAGAIN; 17521a91ccccSSam Leffler crypto_done(crp); 1753091d81d1SSam Leffler return 0; 1754091d81d1SSam Leffler } else { 1755091d81d1SSam Leffler /* 1756091d81d1SSam Leffler * Invoke the driver to process the request. 1757091d81d1SSam Leffler */ 17586810ad6fSSam Leffler return CRYPTODEV_PROCESS(cap->cc_dev, crp, hint); 1759091d81d1SSam Leffler } 1760091d81d1SSam Leffler } 1761091d81d1SSam Leffler 1762091d81d1SSam Leffler void 1763946b8f6fSJohn Baldwin crypto_destroyreq(struct cryptop *crp) 1764091d81d1SSam Leffler { 17650d5c337bSPawel Jakub Dawidek #ifdef DIAGNOSTIC 17660d5c337bSPawel Jakub Dawidek { 17670d5c337bSPawel Jakub Dawidek struct cryptop *crp2; 176839bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 17690d5c337bSPawel Jakub Dawidek 17700d5c337bSPawel Jakub Dawidek CRYPTO_Q_LOCK(); 17710d5c337bSPawel Jakub Dawidek TAILQ_FOREACH(crp2, &crp_q, crp_next) { 17720d5c337bSPawel Jakub Dawidek KASSERT(crp2 != crp, 17730d5c337bSPawel Jakub Dawidek ("Freeing cryptop from the crypto queue (%p).", 17740d5c337bSPawel Jakub Dawidek crp)); 17750d5c337bSPawel Jakub Dawidek } 17760d5c337bSPawel Jakub Dawidek CRYPTO_Q_UNLOCK(); 177739bbca6fSFabien Thomas 177839bbca6fSFabien Thomas FOREACH_CRYPTO_RETW(ret_worker) { 177939bbca6fSFabien Thomas CRYPTO_RETW_LOCK(ret_worker); 178039bbca6fSFabien Thomas TAILQ_FOREACH(crp2, &ret_worker->crp_ret_q, crp_next) { 17810d5c337bSPawel Jakub Dawidek KASSERT(crp2 != crp, 17820d5c337bSPawel Jakub Dawidek ("Freeing cryptop from the return queue (%p).", 17830d5c337bSPawel Jakub Dawidek crp)); 17840d5c337bSPawel Jakub Dawidek } 178539bbca6fSFabien Thomas CRYPTO_RETW_UNLOCK(ret_worker); 178639bbca6fSFabien Thomas } 17870d5c337bSPawel Jakub Dawidek } 17880d5c337bSPawel Jakub Dawidek #endif 1789946b8f6fSJohn Baldwin } 17900d5c337bSPawel Jakub Dawidek 1791946b8f6fSJohn Baldwin void 1792946b8f6fSJohn Baldwin crypto_freereq(struct cryptop *crp) 1793946b8f6fSJohn Baldwin { 1794946b8f6fSJohn Baldwin if (crp == NULL) 1795946b8f6fSJohn Baldwin return; 1796946b8f6fSJohn Baldwin 1797946b8f6fSJohn Baldwin crypto_destroyreq(crp); 1798091d81d1SSam Leffler uma_zfree(cryptop_zone, crp); 1799091d81d1SSam Leffler } 1800091d81d1SSam Leffler 1801946b8f6fSJohn Baldwin static void 1802946b8f6fSJohn Baldwin _crypto_initreq(struct cryptop *crp, crypto_session_t cses) 1803946b8f6fSJohn Baldwin { 1804946b8f6fSJohn Baldwin crp->crp_session = cses; 1805946b8f6fSJohn Baldwin } 1806946b8f6fSJohn Baldwin 1807946b8f6fSJohn Baldwin void 1808946b8f6fSJohn Baldwin crypto_initreq(struct cryptop *crp, crypto_session_t cses) 1809946b8f6fSJohn Baldwin { 1810946b8f6fSJohn Baldwin memset(crp, 0, sizeof(*crp)); 1811946b8f6fSJohn Baldwin _crypto_initreq(crp, cses); 1812946b8f6fSJohn Baldwin } 1813946b8f6fSJohn Baldwin 1814091d81d1SSam Leffler struct cryptop * 1815c0341432SJohn Baldwin crypto_getreq(crypto_session_t cses, int how) 1816091d81d1SSam Leffler { 1817091d81d1SSam Leffler struct cryptop *crp; 1818091d81d1SSam Leffler 1819c0341432SJohn Baldwin MPASS(how == M_WAITOK || how == M_NOWAIT); 1820c0341432SJohn Baldwin crp = uma_zalloc(cryptop_zone, how | M_ZERO); 1821946b8f6fSJohn Baldwin if (crp != NULL) 1822946b8f6fSJohn Baldwin _crypto_initreq(crp, cses); 1823c0341432SJohn Baldwin return (crp); 1824091d81d1SSam Leffler } 1825091d81d1SSam Leffler 1826091d81d1SSam Leffler /* 1827091d81d1SSam Leffler * Invoke the callback on behalf of the driver. 1828091d81d1SSam Leffler */ 1829091d81d1SSam Leffler void 1830091d81d1SSam Leffler crypto_done(struct cryptop *crp) 1831091d81d1SSam Leffler { 18323569ae7fSSam Leffler KASSERT((crp->crp_flags & CRYPTO_F_DONE) == 0, 18333569ae7fSSam Leffler ("crypto_done: op already done, flags 0x%x", crp->crp_flags)); 18343569ae7fSSam Leffler crp->crp_flags |= CRYPTO_F_DONE; 18357d1853eeSSam Leffler if (crp->crp_etype != 0) 18367290cb47SMark Johnston CRYPTOSTAT_INC(cs_errs); 1837a5c053f5SMark Johnston 1838d8409aafSSam Leffler /* 1839d8409aafSSam Leffler * CBIMM means unconditionally do the callback immediately; 1840d8409aafSSam Leffler * CBIFSYNC means do the callback immediately only if the 1841d8409aafSSam Leffler * operation was done synchronously. Both are used to avoid 1842d8409aafSSam Leffler * doing extraneous context switches; the latter is mostly 1843d8409aafSSam Leffler * used with the software crypto driver. 1844d8409aafSSam Leffler */ 184568f6800cSMark Johnston if ((crp->crp_flags & CRYPTO_F_ASYNC_ORDERED) == 0 && 184668f6800cSMark Johnston ((crp->crp_flags & CRYPTO_F_CBIMM) != 0 || 184768f6800cSMark Johnston ((crp->crp_flags & CRYPTO_F_CBIFSYNC) != 0 && 184868f6800cSMark Johnston CRYPTO_SESS_SYNC(crp->crp_session)))) { 1849eb73a605SSam Leffler /* 1850eb73a605SSam Leffler * Do the callback directly. This is ok when the 1851eb73a605SSam Leffler * callback routine does very little (e.g. the 1852eb73a605SSam Leffler * /dev/crypto callback method just does a wakeup). 1853eb73a605SSam Leffler */ 1854eb73a605SSam Leffler crp->crp_callback(crp); 1855eb73a605SSam Leffler } else { 185639bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 185739bbca6fSFabien Thomas bool wake; 185839bbca6fSFabien Thomas 185939bbca6fSFabien Thomas ret_worker = CRYPTO_RETW(crp->crp_retw_id); 186039bbca6fSFabien Thomas 1861eb73a605SSam Leffler /* 1862eb73a605SSam Leffler * Normal case; queue the callback for the thread. 1863eb73a605SSam Leffler */ 186439bbca6fSFabien Thomas CRYPTO_RETW_LOCK(ret_worker); 186568f6800cSMark Johnston if ((crp->crp_flags & CRYPTO_F_ASYNC_ORDERED) != 0) { 186639bbca6fSFabien Thomas struct cryptop *tmp; 186739bbca6fSFabien Thomas 186868f6800cSMark Johnston TAILQ_FOREACH_REVERSE(tmp, 186968f6800cSMark Johnston &ret_worker->crp_ordered_ret_q, cryptop_q, 187068f6800cSMark Johnston crp_next) { 187139bbca6fSFabien Thomas if (CRYPTO_SEQ_GT(crp->crp_seq, tmp->crp_seq)) { 187268f6800cSMark Johnston TAILQ_INSERT_AFTER( 187368f6800cSMark Johnston &ret_worker->crp_ordered_ret_q, tmp, 187468f6800cSMark Johnston crp, crp_next); 187539bbca6fSFabien Thomas break; 187639bbca6fSFabien Thomas } 187739bbca6fSFabien Thomas } 187839bbca6fSFabien Thomas if (tmp == NULL) { 187968f6800cSMark Johnston TAILQ_INSERT_HEAD( 188068f6800cSMark Johnston &ret_worker->crp_ordered_ret_q, crp, 188168f6800cSMark Johnston crp_next); 188239bbca6fSFabien Thomas } 188339bbca6fSFabien Thomas 188468f6800cSMark Johnston wake = crp->crp_seq == ret_worker->reorder_cur_seq; 188568f6800cSMark Johnston } else { 188668f6800cSMark Johnston wake = TAILQ_EMPTY(&ret_worker->crp_ret_q); 188768f6800cSMark Johnston TAILQ_INSERT_TAIL(&ret_worker->crp_ret_q, crp, 188868f6800cSMark Johnston crp_next); 188939bbca6fSFabien Thomas } 189039bbca6fSFabien Thomas 189139bbca6fSFabien Thomas if (wake) 189239bbca6fSFabien Thomas wakeup_one(&ret_worker->crp_ret_q); /* shared wait channel */ 189339bbca6fSFabien Thomas CRYPTO_RETW_UNLOCK(ret_worker); 1894091d81d1SSam Leffler } 1895eb73a605SSam Leffler } 1896091d81d1SSam Leffler 1897091d81d1SSam Leffler /* 1898091d81d1SSam Leffler * Invoke the callback on behalf of the driver. 1899091d81d1SSam Leffler */ 1900091d81d1SSam Leffler void 1901091d81d1SSam Leffler crypto_kdone(struct cryptkop *krp) 1902091d81d1SSam Leffler { 190339bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 19044acae0acSPawel Jakub Dawidek struct cryptocap *cap; 1905091d81d1SSam Leffler 19067d1853eeSSam Leffler if (krp->krp_status != 0) 19077290cb47SMark Johnston CRYPTOSTAT_INC(cs_kerrs); 1908c0341432SJohn Baldwin cap = krp->krp_cap; 1909e7f6b6cfSJohn Baldwin if (cap != NULL) { 1910e7f6b6cfSJohn Baldwin CRYPTO_DRIVER_LOCK(); 1911fb17b4c5SJohn-Mark Gurney KASSERT(cap->cc_koperations > 0, ("cc_koperations == 0")); 19124acae0acSPawel Jakub Dawidek cap->cc_koperations--; 1913e7f6b6cfSJohn Baldwin if (cap->cc_koperations == 0 && 1914e7f6b6cfSJohn Baldwin cap->cc_flags & CRYPTOCAP_F_CLEANUP) 1915c0341432SJohn Baldwin wakeup(cap); 19164acae0acSPawel Jakub Dawidek CRYPTO_DRIVER_UNLOCK(); 1917c0341432SJohn Baldwin krp->krp_cap = NULL; 1918c0341432SJohn Baldwin cap_rele(cap); 1919e7f6b6cfSJohn Baldwin } 192039bbca6fSFabien Thomas 192139bbca6fSFabien Thomas ret_worker = CRYPTO_RETW(0); 192239bbca6fSFabien Thomas 192339bbca6fSFabien Thomas CRYPTO_RETW_LOCK(ret_worker); 192468f6800cSMark Johnston if (TAILQ_EMPTY(&ret_worker->crp_ret_kq)) 192539bbca6fSFabien Thomas wakeup_one(&ret_worker->crp_ret_q); /* shared wait channel */ 192639bbca6fSFabien Thomas TAILQ_INSERT_TAIL(&ret_worker->crp_ret_kq, krp, krp_next); 192739bbca6fSFabien Thomas CRYPTO_RETW_UNLOCK(ret_worker); 1928091d81d1SSam Leffler } 1929091d81d1SSam Leffler 1930091d81d1SSam Leffler int 1931091d81d1SSam Leffler crypto_getfeat(int *featp) 1932091d81d1SSam Leffler { 1933091d81d1SSam Leffler int hid, kalg, feat = 0; 1934091d81d1SSam Leffler 1935091d81d1SSam Leffler CRYPTO_DRIVER_LOCK(); 1936c0341432SJohn Baldwin for (hid = 0; hid < crypto_drivers_size; hid++) { 1937c0341432SJohn Baldwin const struct cryptocap *cap = crypto_drivers[hid]; 19386810ad6fSSam Leffler 1939c0341432SJohn Baldwin if (cap == NULL || 1940c0341432SJohn Baldwin ((cap->cc_flags & CRYPTOCAP_F_SOFTWARE) && 1941c0341432SJohn Baldwin !crypto_devallowsoft)) { 1942091d81d1SSam Leffler continue; 1943091d81d1SSam Leffler } 1944091d81d1SSam Leffler for (kalg = 0; kalg < CRK_ALGORITHM_MAX; kalg++) 19456810ad6fSSam Leffler if (cap->cc_kalg[kalg] & CRYPTO_ALG_FLAG_SUPPORTED) 1946091d81d1SSam Leffler feat |= 1 << kalg; 1947091d81d1SSam Leffler } 1948091d81d1SSam Leffler CRYPTO_DRIVER_UNLOCK(); 1949091d81d1SSam Leffler *featp = feat; 1950091d81d1SSam Leffler return (0); 1951091d81d1SSam Leffler } 1952091d81d1SSam Leffler 195351e45326SSam Leffler /* 195451e45326SSam Leffler * Terminate a thread at module unload. The process that 195551e45326SSam Leffler * initiated this is waiting for us to signal that we're gone; 195651e45326SSam Leffler * wake it up and exit. We use the driver table lock to insure 195751e45326SSam Leffler * we don't do the wakeup before they're waiting. There is no 195851e45326SSam Leffler * race here because the waiter sleeps on the proc lock for the 195951e45326SSam Leffler * thread so it gets notified at the right time because of an 196051e45326SSam Leffler * extra wakeup that's done in exit1(). 196151e45326SSam Leffler */ 1962091d81d1SSam Leffler static void 196351e45326SSam Leffler crypto_finis(void *chan) 1964091d81d1SSam Leffler { 196551e45326SSam Leffler CRYPTO_DRIVER_LOCK(); 196651e45326SSam Leffler wakeup_one(chan); 196751e45326SSam Leffler CRYPTO_DRIVER_UNLOCK(); 19683745c395SJulian Elischer kproc_exit(0); 1969091d81d1SSam Leffler } 1970091d81d1SSam Leffler 1971091d81d1SSam Leffler /* 19721a91ccccSSam Leffler * Crypto thread, dispatches crypto requests. 1973091d81d1SSam Leffler */ 1974091d81d1SSam Leffler static void 1975091d81d1SSam Leffler crypto_proc(void) 1976091d81d1SSam Leffler { 19771a91ccccSSam Leffler struct cryptop *crp, *submit; 19781a91ccccSSam Leffler struct cryptkop *krp; 1979091d81d1SSam Leffler struct cryptocap *cap; 1980091d81d1SSam Leffler int result, hint; 1981091d81d1SSam Leffler 19826ed982a2SAndrew Turner #if defined(__i386__) || defined(__amd64__) || defined(__aarch64__) 198304c49e68SKonstantin Belousov fpu_kern_thread(FPU_KERN_NORMAL); 198404c49e68SKonstantin Belousov #endif 198504c49e68SKonstantin Belousov 19861a91ccccSSam Leffler CRYPTO_Q_LOCK(); 1987091d81d1SSam Leffler for (;;) { 1988091d81d1SSam Leffler /* 1989091d81d1SSam Leffler * Find the first element in the queue that can be 1990091d81d1SSam Leffler * processed and look-ahead to see if multiple ops 1991091d81d1SSam Leffler * are ready for the same driver. 1992091d81d1SSam Leffler */ 1993091d81d1SSam Leffler submit = NULL; 1994091d81d1SSam Leffler hint = 0; 1995091d81d1SSam Leffler TAILQ_FOREACH(crp, &crp_q, crp_next) { 1996c0341432SJohn Baldwin cap = crp->crp_session->cap; 19974acae0acSPawel Jakub Dawidek /* 19984acae0acSPawel Jakub Dawidek * Driver cannot disappeared when there is an active 19994acae0acSPawel Jakub Dawidek * session. 20004acae0acSPawel Jakub Dawidek */ 2001c3c82036SPawel Jakub Dawidek KASSERT(cap != NULL, ("%s:%u Driver disappeared.", 2002c3c82036SPawel Jakub Dawidek __func__, __LINE__)); 2003c0341432SJohn Baldwin if (cap->cc_flags & CRYPTOCAP_F_CLEANUP) { 2004091d81d1SSam Leffler /* Op needs to be migrated, process it. */ 2005091d81d1SSam Leffler if (submit == NULL) 2006091d81d1SSam Leffler submit = crp; 2007091d81d1SSam Leffler break; 2008091d81d1SSam Leffler } 2009091d81d1SSam Leffler if (!cap->cc_qblocked) { 2010091d81d1SSam Leffler if (submit != NULL) { 2011091d81d1SSam Leffler /* 2012091d81d1SSam Leffler * We stop on finding another op, 2013091d81d1SSam Leffler * regardless whether its for the same 2014091d81d1SSam Leffler * driver or not. We could keep 2015091d81d1SSam Leffler * searching the queue but it might be 2016091d81d1SSam Leffler * better to just use a per-driver 2017091d81d1SSam Leffler * queue instead. 2018091d81d1SSam Leffler */ 2019c0341432SJohn Baldwin if (submit->crp_session->cap == cap) 2020091d81d1SSam Leffler hint = CRYPTO_HINT_MORE; 2021091d81d1SSam Leffler } else { 2022091d81d1SSam Leffler submit = crp; 2023091d81d1SSam Leffler } 202468f6800cSMark Johnston break; 2025091d81d1SSam Leffler } 2026091d81d1SSam Leffler } 2027091d81d1SSam Leffler if (submit != NULL) { 2028091d81d1SSam Leffler TAILQ_REMOVE(&crp_q, submit, crp_next); 2029c0341432SJohn Baldwin cap = submit->crp_session->cap; 2030c3c82036SPawel Jakub Dawidek KASSERT(cap != NULL, ("%s:%u Driver disappeared.", 2031c3c82036SPawel Jakub Dawidek __func__, __LINE__)); 2032c0341432SJohn Baldwin CRYPTO_Q_UNLOCK(); 20334acae0acSPawel Jakub Dawidek result = crypto_invoke(cap, submit, hint); 2034c0341432SJohn Baldwin CRYPTO_Q_LOCK(); 2035091d81d1SSam Leffler if (result == ERESTART) { 2036091d81d1SSam Leffler /* 2037091d81d1SSam Leffler * The driver ran out of resources, mark the 2038091d81d1SSam Leffler * driver ``blocked'' for cryptop's and put 2039091d81d1SSam Leffler * the request back in the queue. It would 2040091d81d1SSam Leffler * best to put the request back where we got 2041091d81d1SSam Leffler * it but that's hard so for now we put it 2042091d81d1SSam Leffler * at the front. This should be ok; putting 2043091d81d1SSam Leffler * it at the end does not work. 2044091d81d1SSam Leffler */ 2045c0341432SJohn Baldwin cap->cc_qblocked = 1; 2046091d81d1SSam Leffler TAILQ_INSERT_HEAD(&crp_q, submit, crp_next); 20477290cb47SMark Johnston CRYPTOSTAT_INC(cs_blocks); 2048091d81d1SSam Leffler } 2049091d81d1SSam Leffler } 2050091d81d1SSam Leffler 2051091d81d1SSam Leffler /* As above, but for key ops */ 2052091d81d1SSam Leffler TAILQ_FOREACH(krp, &crp_kq, krp_next) { 2053c0341432SJohn Baldwin cap = krp->krp_cap; 2054c0341432SJohn Baldwin if (cap->cc_flags & CRYPTOCAP_F_CLEANUP) { 20556810ad6fSSam Leffler /* 2056c0341432SJohn Baldwin * Operation needs to be migrated, 2057c0341432SJohn Baldwin * clear krp_cap so a new driver is 2058c0341432SJohn Baldwin * selected. 20596810ad6fSSam Leffler */ 2060c0341432SJohn Baldwin krp->krp_cap = NULL; 2061c0341432SJohn Baldwin cap_rele(cap); 2062091d81d1SSam Leffler break; 2063091d81d1SSam Leffler } 2064091d81d1SSam Leffler if (!cap->cc_kqblocked) 2065091d81d1SSam Leffler break; 2066091d81d1SSam Leffler } 2067091d81d1SSam Leffler if (krp != NULL) { 2068091d81d1SSam Leffler TAILQ_REMOVE(&crp_kq, krp, krp_next); 2069c0341432SJohn Baldwin CRYPTO_Q_UNLOCK(); 2070c0341432SJohn Baldwin result = crypto_kinvoke(krp); 2071c0341432SJohn Baldwin CRYPTO_Q_LOCK(); 2072091d81d1SSam Leffler if (result == ERESTART) { 2073091d81d1SSam Leffler /* 2074091d81d1SSam Leffler * The driver ran out of resources, mark the 2075091d81d1SSam Leffler * driver ``blocked'' for cryptkop's and put 2076091d81d1SSam Leffler * the request back in the queue. It would 2077091d81d1SSam Leffler * best to put the request back where we got 2078091d81d1SSam Leffler * it but that's hard so for now we put it 2079091d81d1SSam Leffler * at the front. This should be ok; putting 2080091d81d1SSam Leffler * it at the end does not work. 2081091d81d1SSam Leffler */ 2082c0341432SJohn Baldwin krp->krp_cap->cc_kqblocked = 1; 2083091d81d1SSam Leffler TAILQ_INSERT_HEAD(&crp_kq, krp, krp_next); 20847290cb47SMark Johnston CRYPTOSTAT_INC(cs_kblocks); 2085091d81d1SSam Leffler } 2086091d81d1SSam Leffler } 2087091d81d1SSam Leffler 20881a91ccccSSam Leffler if (submit == NULL && krp == NULL) { 2089091d81d1SSam Leffler /* 2090091d81d1SSam Leffler * Nothing more to be processed. Sleep until we're 2091091d81d1SSam Leffler * woken because there are more ops to process. 2092091d81d1SSam Leffler * This happens either by submission or by a driver 2093091d81d1SSam Leffler * becoming unblocked and notifying us through 2094091d81d1SSam Leffler * crypto_unblock. Note that when we wakeup we 2095091d81d1SSam Leffler * start processing each queue again from the 2096091d81d1SSam Leffler * front. It's not clear that it's important to 2097091d81d1SSam Leffler * preserve this ordering since ops may finish 2098091d81d1SSam Leffler * out of order if dispatched to different devices 2099091d81d1SSam Leffler * and some become blocked while others do not. 2100091d81d1SSam Leffler */ 21013a865c82SPawel Jakub Dawidek crp_sleep = 1; 21021a91ccccSSam Leffler msleep(&crp_q, &crypto_q_mtx, PWAIT, "crypto_wait", 0); 21033a865c82SPawel Jakub Dawidek crp_sleep = 0; 210451e45326SSam Leffler if (cryptoproc == NULL) 210551e45326SSam Leffler break; 21067290cb47SMark Johnston CRYPTOSTAT_INC(cs_intrs); 2107091d81d1SSam Leffler } 2108091d81d1SSam Leffler } 210951e45326SSam Leffler CRYPTO_Q_UNLOCK(); 21101a91ccccSSam Leffler 211151e45326SSam Leffler crypto_finis(&crp_q); 21121a91ccccSSam Leffler } 21131a91ccccSSam Leffler 21141a91ccccSSam Leffler /* 21151a91ccccSSam Leffler * Crypto returns thread, does callbacks for processed crypto requests. 21161a91ccccSSam Leffler * Callbacks are done here, rather than in the crypto drivers, because 21171a91ccccSSam Leffler * callbacks typically are expensive and would slow interrupt handling. 21181a91ccccSSam Leffler */ 21191a91ccccSSam Leffler static void 212039bbca6fSFabien Thomas crypto_ret_proc(struct crypto_ret_worker *ret_worker) 21211a91ccccSSam Leffler { 21221a91ccccSSam Leffler struct cryptop *crpt; 21231a91ccccSSam Leffler struct cryptkop *krpt; 21241a91ccccSSam Leffler 212539bbca6fSFabien Thomas CRYPTO_RETW_LOCK(ret_worker); 21261a91ccccSSam Leffler for (;;) { 21271a91ccccSSam Leffler /* Harvest return q's for completed ops */ 212839bbca6fSFabien Thomas crpt = TAILQ_FIRST(&ret_worker->crp_ordered_ret_q); 212939bbca6fSFabien Thomas if (crpt != NULL) { 213039bbca6fSFabien Thomas if (crpt->crp_seq == ret_worker->reorder_cur_seq) { 213139bbca6fSFabien Thomas TAILQ_REMOVE(&ret_worker->crp_ordered_ret_q, crpt, crp_next); 213239bbca6fSFabien Thomas ret_worker->reorder_cur_seq++; 213339bbca6fSFabien Thomas } else { 213439bbca6fSFabien Thomas crpt = NULL; 213539bbca6fSFabien Thomas } 213639bbca6fSFabien Thomas } 21371a91ccccSSam Leffler 213839bbca6fSFabien Thomas if (crpt == NULL) { 213939bbca6fSFabien Thomas crpt = TAILQ_FIRST(&ret_worker->crp_ret_q); 214039bbca6fSFabien Thomas if (crpt != NULL) 214139bbca6fSFabien Thomas TAILQ_REMOVE(&ret_worker->crp_ret_q, crpt, crp_next); 214239bbca6fSFabien Thomas } 214339bbca6fSFabien Thomas 214439bbca6fSFabien Thomas krpt = TAILQ_FIRST(&ret_worker->crp_ret_kq); 21451a91ccccSSam Leffler if (krpt != NULL) 214639bbca6fSFabien Thomas TAILQ_REMOVE(&ret_worker->crp_ret_kq, krpt, krp_next); 21471a91ccccSSam Leffler 21481a91ccccSSam Leffler if (crpt != NULL || krpt != NULL) { 214939bbca6fSFabien Thomas CRYPTO_RETW_UNLOCK(ret_worker); 21501a91ccccSSam Leffler /* 21511a91ccccSSam Leffler * Run callbacks unlocked. 21521a91ccccSSam Leffler */ 2153a5c053f5SMark Johnston if (crpt != NULL) 21541a91ccccSSam Leffler crpt->crp_callback(crpt); 21551a91ccccSSam Leffler if (krpt != NULL) 21561a91ccccSSam Leffler krpt->krp_callback(krpt); 215739bbca6fSFabien Thomas CRYPTO_RETW_LOCK(ret_worker); 21581a91ccccSSam Leffler } else { 21591a91ccccSSam Leffler /* 21601a91ccccSSam Leffler * Nothing more to be processed. Sleep until we're 21611a91ccccSSam Leffler * woken because there are more returns to process. 21621a91ccccSSam Leffler */ 216339bbca6fSFabien Thomas msleep(&ret_worker->crp_ret_q, &ret_worker->crypto_ret_mtx, PWAIT, 21641a91ccccSSam Leffler "crypto_ret_wait", 0); 216539bbca6fSFabien Thomas if (ret_worker->cryptoretproc == NULL) 216651e45326SSam Leffler break; 21677290cb47SMark Johnston CRYPTOSTAT_INC(cs_rets); 21681a91ccccSSam Leffler } 21691a91ccccSSam Leffler } 217039bbca6fSFabien Thomas CRYPTO_RETW_UNLOCK(ret_worker); 217151e45326SSam Leffler 217239bbca6fSFabien Thomas crypto_finis(&ret_worker->crp_ret_q); 21731a91ccccSSam Leffler } 21746810ad6fSSam Leffler 21756810ad6fSSam Leffler #ifdef DDB 21766810ad6fSSam Leffler static void 21776810ad6fSSam Leffler db_show_drivers(void) 21786810ad6fSSam Leffler { 21796810ad6fSSam Leffler int hid; 21806810ad6fSSam Leffler 21816810ad6fSSam Leffler db_printf("%12s %4s %4s %8s %2s %2s\n" 21826810ad6fSSam Leffler , "Device" 21836810ad6fSSam Leffler , "Ses" 21846810ad6fSSam Leffler , "Kops" 21856810ad6fSSam Leffler , "Flags" 21866810ad6fSSam Leffler , "QB" 21876810ad6fSSam Leffler , "KB" 21886810ad6fSSam Leffler ); 2189c0341432SJohn Baldwin for (hid = 0; hid < crypto_drivers_size; hid++) { 2190c0341432SJohn Baldwin const struct cryptocap *cap = crypto_drivers[hid]; 2191c0341432SJohn Baldwin if (cap == NULL) 21926810ad6fSSam Leffler continue; 21936810ad6fSSam Leffler db_printf("%-12s %4u %4u %08x %2u %2u\n" 21946810ad6fSSam Leffler , device_get_nameunit(cap->cc_dev) 21956810ad6fSSam Leffler , cap->cc_sessions 21966810ad6fSSam Leffler , cap->cc_koperations 21976810ad6fSSam Leffler , cap->cc_flags 21986810ad6fSSam Leffler , cap->cc_qblocked 21996810ad6fSSam Leffler , cap->cc_kqblocked 22006810ad6fSSam Leffler ); 22016810ad6fSSam Leffler } 22026810ad6fSSam Leffler } 22036810ad6fSSam Leffler 22046810ad6fSSam Leffler DB_SHOW_COMMAND(crypto, db_show_crypto) 22056810ad6fSSam Leffler { 22066810ad6fSSam Leffler struct cryptop *crp; 220739bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 22086810ad6fSSam Leffler 22096810ad6fSSam Leffler db_show_drivers(); 22106810ad6fSSam Leffler db_printf("\n"); 22116810ad6fSSam Leffler 22126810ad6fSSam Leffler db_printf("%4s %8s %4s %4s %4s %4s %8s %8s\n", 22136810ad6fSSam Leffler "HID", "Caps", "Ilen", "Olen", "Etype", "Flags", 2214c0341432SJohn Baldwin "Device", "Callback"); 22156810ad6fSSam Leffler TAILQ_FOREACH(crp, &crp_q, crp_next) { 22169c0e3d3aSJohn Baldwin db_printf("%4u %08x %4u %4u %04x %8p %8p\n" 2217c0341432SJohn Baldwin , crp->crp_session->cap->cc_hid 22181b0909d5SConrad Meyer , (int) crypto_ses2caps(crp->crp_session) 22199c0e3d3aSJohn Baldwin , crp->crp_olen 22206810ad6fSSam Leffler , crp->crp_etype 22216810ad6fSSam Leffler , crp->crp_flags 2222c0341432SJohn Baldwin , device_get_nameunit(crp->crp_session->cap->cc_dev) 22236810ad6fSSam Leffler , crp->crp_callback 22246810ad6fSSam Leffler ); 22256810ad6fSSam Leffler } 222639bbca6fSFabien Thomas FOREACH_CRYPTO_RETW(ret_worker) { 222739bbca6fSFabien Thomas db_printf("\n%8s %4s %4s %4s %8s\n", 222839bbca6fSFabien Thomas "ret_worker", "HID", "Etype", "Flags", "Callback"); 222939bbca6fSFabien Thomas if (!TAILQ_EMPTY(&ret_worker->crp_ret_q)) { 223039bbca6fSFabien Thomas TAILQ_FOREACH(crp, &ret_worker->crp_ret_q, crp_next) { 223139bbca6fSFabien Thomas db_printf("%8td %4u %4u %04x %8p\n" 223239bbca6fSFabien Thomas , CRYPTO_RETW_ID(ret_worker) 2233c0341432SJohn Baldwin , crp->crp_session->cap->cc_hid 22346810ad6fSSam Leffler , crp->crp_etype 22356810ad6fSSam Leffler , crp->crp_flags 22366810ad6fSSam Leffler , crp->crp_callback 22376810ad6fSSam Leffler ); 22386810ad6fSSam Leffler } 22396810ad6fSSam Leffler } 22406810ad6fSSam Leffler } 224139bbca6fSFabien Thomas } 22426810ad6fSSam Leffler 22436810ad6fSSam Leffler DB_SHOW_COMMAND(kcrypto, db_show_kcrypto) 22446810ad6fSSam Leffler { 22456810ad6fSSam Leffler struct cryptkop *krp; 224639bbca6fSFabien Thomas struct crypto_ret_worker *ret_worker; 22476810ad6fSSam Leffler 22486810ad6fSSam Leffler db_show_drivers(); 22496810ad6fSSam Leffler db_printf("\n"); 22506810ad6fSSam Leffler 22516810ad6fSSam Leffler db_printf("%4s %5s %4s %4s %8s %4s %8s\n", 22526810ad6fSSam Leffler "Op", "Status", "#IP", "#OP", "CRID", "HID", "Callback"); 22536810ad6fSSam Leffler TAILQ_FOREACH(krp, &crp_kq, krp_next) { 22546810ad6fSSam Leffler db_printf("%4u %5u %4u %4u %08x %4u %8p\n" 22556810ad6fSSam Leffler , krp->krp_op 22566810ad6fSSam Leffler , krp->krp_status 22576810ad6fSSam Leffler , krp->krp_iparams, krp->krp_oparams 22586810ad6fSSam Leffler , krp->krp_crid, krp->krp_hid 22596810ad6fSSam Leffler , krp->krp_callback 22606810ad6fSSam Leffler ); 22616810ad6fSSam Leffler } 226239bbca6fSFabien Thomas 226339bbca6fSFabien Thomas ret_worker = CRYPTO_RETW(0); 226439bbca6fSFabien Thomas if (!TAILQ_EMPTY(&ret_worker->crp_ret_q)) { 22656810ad6fSSam Leffler db_printf("%4s %5s %8s %4s %8s\n", 22666810ad6fSSam Leffler "Op", "Status", "CRID", "HID", "Callback"); 226739bbca6fSFabien Thomas TAILQ_FOREACH(krp, &ret_worker->crp_ret_kq, krp_next) { 22686810ad6fSSam Leffler db_printf("%4u %5u %08x %4u %8p\n" 22696810ad6fSSam Leffler , krp->krp_op 22706810ad6fSSam Leffler , krp->krp_status 22716810ad6fSSam Leffler , krp->krp_crid, krp->krp_hid 22726810ad6fSSam Leffler , krp->krp_callback 22736810ad6fSSam Leffler ); 22746810ad6fSSam Leffler } 22756810ad6fSSam Leffler } 22766810ad6fSSam Leffler } 22776810ad6fSSam Leffler #endif 22786810ad6fSSam Leffler 22796810ad6fSSam Leffler int crypto_modevent(module_t mod, int type, void *unused); 22806810ad6fSSam Leffler 22816810ad6fSSam Leffler /* 22826810ad6fSSam Leffler * Initialization code, both for static and dynamic loading. 22836810ad6fSSam Leffler * Note this is not invoked with the usual MODULE_DECLARE 22846810ad6fSSam Leffler * mechanism but instead is listed as a dependency by the 22856810ad6fSSam Leffler * cryptosoft driver. This guarantees proper ordering of 22866810ad6fSSam Leffler * calls on module load/unload. 22876810ad6fSSam Leffler */ 22886810ad6fSSam Leffler int 22896810ad6fSSam Leffler crypto_modevent(module_t mod, int type, void *unused) 22906810ad6fSSam Leffler { 22916810ad6fSSam Leffler int error = EINVAL; 22926810ad6fSSam Leffler 22936810ad6fSSam Leffler switch (type) { 22946810ad6fSSam Leffler case MOD_LOAD: 22956810ad6fSSam Leffler error = crypto_init(); 22966810ad6fSSam Leffler if (error == 0 && bootverbose) 22976810ad6fSSam Leffler printf("crypto: <crypto core>\n"); 22986810ad6fSSam Leffler break; 22996810ad6fSSam Leffler case MOD_UNLOAD: 23006810ad6fSSam Leffler /*XXX disallow if active sessions */ 23016810ad6fSSam Leffler error = 0; 23026810ad6fSSam Leffler crypto_destroy(); 23036810ad6fSSam Leffler return 0; 23046810ad6fSSam Leffler } 23056810ad6fSSam Leffler return error; 23066810ad6fSSam Leffler } 23076810ad6fSSam Leffler MODULE_VERSION(crypto, 1); 23086810ad6fSSam Leffler MODULE_DEPEND(crypto, zlib, 1, 1, 1); 2309