xref: /titanic_52/usr/src/uts/common/crypto/io/dca.c (revision c7158ae983f5a04c4a998f468ecefba6d23ba721)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 /*
30  * Deimos - cryptographic acceleration based upon Broadcom 582x.
31  */
32 
33 #include <sys/types.h>
34 #include <sys/modctl.h>
35 #include <sys/conf.h>
36 #include <sys/devops.h>
37 #include <sys/ddi.h>
38 #include <sys/sunddi.h>
39 #include <sys/cmn_err.h>
40 #include <sys/varargs.h>
41 #include <sys/file.h>
42 #include <sys/stat.h>
43 #include <sys/kmem.h>
44 #include <sys/ioccom.h>
45 #include <sys/open.h>
46 #include <sys/cred.h>
47 #include <sys/kstat.h>
48 #include <sys/strsun.h>
49 #include <sys/note.h>
50 #include <sys/crypto/common.h>
51 #include <sys/crypto/spi.h>
52 #include <sys/ddifm.h>
53 #include <sys/fm/protocol.h>
54 #include <sys/fm/util.h>
55 #include <sys/fm/io/ddi.h>
56 #include <sys/crypto/dca.h>
57 
58 /*
59  * Core Deimos driver.
60  */
61 
62 static void		dca_enlist2(dca_listnode_t *, dca_listnode_t *,
63     kmutex_t *);
64 static void		dca_rmlist2(dca_listnode_t *node, kmutex_t *);
65 static dca_listnode_t	*dca_delist2(dca_listnode_t *q, kmutex_t *);
66 static void		dca_free_context_list(dca_t *dca);
67 static int		dca_free_context_low(crypto_ctx_t *ctx);
68 static int		dca_attach(dev_info_t *, ddi_attach_cmd_t);
69 static int		dca_detach(dev_info_t *, ddi_detach_cmd_t);
70 static int		dca_suspend(dca_t *);
71 static int		dca_resume(dca_t *);
72 static int		dca_init(dca_t *);
73 static int		dca_reset(dca_t *, int);
74 static int		dca_initworklist(dca_t *, dca_worklist_t *);
75 static void		dca_uninit(dca_t *);
76 static void		dca_initq(dca_listnode_t *);
77 static void		dca_enqueue(dca_listnode_t *, dca_listnode_t *);
78 static dca_listnode_t	*dca_dequeue(dca_listnode_t *);
79 static dca_listnode_t	*dca_unqueue(dca_listnode_t *);
80 static dca_request_t	*dca_newreq(dca_t *);
81 static dca_work_t	*dca_getwork(dca_t *, int);
82 static void		dca_freework(dca_work_t *);
83 static dca_work_t	*dca_newwork(dca_t *);
84 static void		dca_destroywork(dca_work_t *);
85 static void		dca_schedule(dca_t *, int);
86 static void		dca_reclaim(dca_t *, int);
87 static uint_t		dca_intr(char *);
88 static void		dca_failure(dca_t *, ddi_fault_location_t,
89 			    dca_fma_eclass_t index, uint64_t, int, char *, ...);
90 static void		dca_jobtimeout(void *);
91 static int		dca_drain(dca_t *);
92 static void		dca_undrain(dca_t *);
93 static void		dca_rejectjobs(dca_t *);
94 
95 #ifdef	SCHEDDELAY
96 static void		dca_schedtimeout(void *);
97 #endif
98 
99 /*
100  * We want these inlined for performance.
101  */
102 #ifndef	DEBUG
103 #pragma inline(dca_freereq, dca_getreq, dca_freework, dca_getwork)
104 #pragma inline(dca_enqueue, dca_dequeue, dca_rmqueue, dca_done)
105 #pragma inline(dca_reverse, dca_length)
106 #endif
107 
108 /*
109  * Device operations.
110  */
111 static struct dev_ops devops = {
112 	DEVO_REV,		/* devo_rev */
113 	0,			/* devo_refcnt */
114 	nodev,			/* devo_getinfo */
115 	nulldev,		/* devo_identify */
116 	nulldev,		/* devo_probe */
117 	dca_attach,		/* devo_attach */
118 	dca_detach,		/* devo_detach */
119 	nodev,			/* devo_reset */
120 	NULL,			/* devo_cb_ops */
121 	NULL,			/* devo_bus_ops */
122 	ddi_power		/* devo_power */
123 };
124 
125 #define	IDENT		"PCI Crypto Accelerator 2.0"
126 #define	IDENT_SYM	"Crypto Accel Sym 2.0"
127 #define	IDENT_ASYM	"Crypto Accel Asym 2.0"
128 
129 /* Space-padded, will be filled in dynamically during registration */
130 #define	IDENT3	"PCI Crypto Accelerator Mod 2.0"
131 
132 #define	VENDOR	"Sun Microsystems, Inc."
133 
134 #define	STALETIME	(30 * SECOND)
135 
136 #define	crypto_prov_notify	crypto_provider_notification
137 		/* A 28 char function name doesn't leave much line space */
138 
139 /*
140  * Module linkage.
141  */
142 static struct modldrv modldrv = {
143 	&mod_driverops,		/* drv_modops */
144 	IDENT,			/* drv_linkinfo */
145 	&devops,		/* drv_dev_ops */
146 };
147 
148 extern struct mod_ops mod_cryptoops;
149 
150 static struct modlcrypto modlcrypto = {
151 	&mod_cryptoops,
152 	IDENT3
153 };
154 
155 static struct modlinkage modlinkage = {
156 	MODREV_1,		/* ml_rev */
157 	&modldrv,		/* ml_linkage */
158 	&modlcrypto,
159 	NULL
160 };
161 
162 /*
163  * CSPI information (entry points, provider info, etc.)
164  */
165 
166 /* Mechanisms for the symmetric cipher provider */
167 static crypto_mech_info_t dca_mech_info_tab1[] = {
168 	/* DES-CBC */
169 	{SUN_CKM_DES_CBC, DES_CBC_MECH_INFO_TYPE,
170 	    CRYPTO_FG_ENCRYPT | CRYPTO_FG_DECRYPT |
171 	    CRYPTO_FG_ENCRYPT_ATOMIC | CRYPTO_FG_DECRYPT_ATOMIC,
172 	    DES_KEY_LEN, DES_KEY_LEN, CRYPTO_KEYSIZE_UNIT_IN_BYTES},
173 	/* 3DES-CBC */
174 	{SUN_CKM_DES3_CBC, DES3_CBC_MECH_INFO_TYPE,
175 	    CRYPTO_FG_ENCRYPT | CRYPTO_FG_DECRYPT |
176 	    CRYPTO_FG_ENCRYPT_ATOMIC | CRYPTO_FG_DECRYPT_ATOMIC,
177 	    DES3_KEY_LEN, DES3_KEY_LEN, CRYPTO_KEYSIZE_UNIT_IN_BYTES}
178 };
179 
180 /* Mechanisms for the asymmetric cipher provider */
181 static crypto_mech_info_t dca_mech_info_tab2[] = {
182 	/* DSA */
183 	{SUN_CKM_DSA, DSA_MECH_INFO_TYPE,
184 	    CRYPTO_FG_SIGN | CRYPTO_FG_VERIFY |
185 	    CRYPTO_FG_SIGN_ATOMIC | CRYPTO_FG_VERIFY_ATOMIC,
186 	    DSA_MIN_KEY_LEN * 8, DSA_MAX_KEY_LEN * 8,
187 	    CRYPTO_KEYSIZE_UNIT_IN_BITS},
188 
189 	/* RSA */
190 	{SUN_CKM_RSA_X_509, RSA_X_509_MECH_INFO_TYPE,
191 	    CRYPTO_FG_ENCRYPT | CRYPTO_FG_DECRYPT | CRYPTO_FG_SIGN |
192 	    CRYPTO_FG_SIGN_RECOVER | CRYPTO_FG_VERIFY |
193 	    CRYPTO_FG_VERIFY_RECOVER |
194 	    CRYPTO_FG_ENCRYPT_ATOMIC | CRYPTO_FG_DECRYPT_ATOMIC |
195 	    CRYPTO_FG_SIGN_ATOMIC | CRYPTO_FG_SIGN_RECOVER_ATOMIC |
196 	    CRYPTO_FG_VERIFY_ATOMIC | CRYPTO_FG_VERIFY_RECOVER_ATOMIC,
197 	    RSA_MIN_KEY_LEN * 8, RSA_MAX_KEY_LEN * 8,
198 	    CRYPTO_KEYSIZE_UNIT_IN_BITS},
199 	{SUN_CKM_RSA_PKCS, RSA_PKCS_MECH_INFO_TYPE,
200 	    CRYPTO_FG_ENCRYPT | CRYPTO_FG_DECRYPT | CRYPTO_FG_SIGN |
201 	    CRYPTO_FG_SIGN_RECOVER | CRYPTO_FG_VERIFY |
202 	    CRYPTO_FG_VERIFY_RECOVER |
203 	    CRYPTO_FG_ENCRYPT_ATOMIC | CRYPTO_FG_DECRYPT_ATOMIC |
204 	    CRYPTO_FG_SIGN_ATOMIC | CRYPTO_FG_SIGN_RECOVER_ATOMIC |
205 	    CRYPTO_FG_VERIFY_ATOMIC | CRYPTO_FG_VERIFY_RECOVER_ATOMIC,
206 	    RSA_MIN_KEY_LEN * 8, RSA_MAX_KEY_LEN * 8,
207 	    CRYPTO_KEYSIZE_UNIT_IN_BITS}
208 };
209 
210 static void dca_provider_status(crypto_provider_handle_t, uint_t *);
211 
212 static crypto_control_ops_t dca_control_ops = {
213 	dca_provider_status
214 };
215 
216 static int dca_encrypt_init(crypto_ctx_t *, crypto_mechanism_t *,
217     crypto_key_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
218 static int dca_encrypt(crypto_ctx_t *, crypto_data_t *, crypto_data_t *,
219     crypto_req_handle_t);
220 static int dca_encrypt_update(crypto_ctx_t *, crypto_data_t *,
221     crypto_data_t *, crypto_req_handle_t);
222 static int dca_encrypt_final(crypto_ctx_t *, crypto_data_t *,
223     crypto_req_handle_t);
224 static int dca_encrypt_atomic(crypto_provider_handle_t, crypto_session_id_t,
225     crypto_mechanism_t *, crypto_key_t *, crypto_data_t *,
226     crypto_data_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
227 
228 static int dca_decrypt_init(crypto_ctx_t *, crypto_mechanism_t *,
229     crypto_key_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
230 static int dca_decrypt(crypto_ctx_t *, crypto_data_t *, crypto_data_t *,
231     crypto_req_handle_t);
232 static int dca_decrypt_update(crypto_ctx_t *, crypto_data_t *,
233     crypto_data_t *, crypto_req_handle_t);
234 static int dca_decrypt_final(crypto_ctx_t *, crypto_data_t *,
235     crypto_req_handle_t);
236 static int dca_decrypt_atomic(crypto_provider_handle_t, crypto_session_id_t,
237     crypto_mechanism_t *, crypto_key_t *, crypto_data_t *,
238     crypto_data_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
239 
240 static crypto_cipher_ops_t dca_cipher_ops = {
241 	dca_encrypt_init,
242 	dca_encrypt,
243 	dca_encrypt_update,
244 	dca_encrypt_final,
245 	dca_encrypt_atomic,
246 	dca_decrypt_init,
247 	dca_decrypt,
248 	dca_decrypt_update,
249 	dca_decrypt_final,
250 	dca_decrypt_atomic
251 };
252 
253 static int dca_sign_init(crypto_ctx_t *, crypto_mechanism_t *, crypto_key_t *,
254     crypto_spi_ctx_template_t, crypto_req_handle_t);
255 static int dca_sign(crypto_ctx_t *, crypto_data_t *, crypto_data_t *,
256     crypto_req_handle_t);
257 static int dca_sign_update(crypto_ctx_t *, crypto_data_t *,
258     crypto_req_handle_t);
259 static int dca_sign_final(crypto_ctx_t *, crypto_data_t *,
260     crypto_req_handle_t);
261 static int dca_sign_atomic(crypto_provider_handle_t, crypto_session_id_t,
262     crypto_mechanism_t *, crypto_key_t *, crypto_data_t *, crypto_data_t *,
263     crypto_spi_ctx_template_t, crypto_req_handle_t);
264 static int dca_sign_recover_init(crypto_ctx_t *, crypto_mechanism_t *,
265     crypto_key_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
266 static int dca_sign_recover(crypto_ctx_t *, crypto_data_t *, crypto_data_t *,
267     crypto_req_handle_t);
268 static int dca_sign_recover_atomic(crypto_provider_handle_t,
269     crypto_session_id_t, crypto_mechanism_t *, crypto_key_t *, crypto_data_t *,
270     crypto_data_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
271 
272 static crypto_sign_ops_t dca_sign_ops = {
273 	dca_sign_init,
274 	dca_sign,
275 	dca_sign_update,
276 	dca_sign_final,
277 	dca_sign_atomic,
278 	dca_sign_recover_init,
279 	dca_sign_recover,
280 	dca_sign_recover_atomic
281 };
282 
283 static int dca_verify_init(crypto_ctx_t *, crypto_mechanism_t *,
284     crypto_key_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
285 static int dca_verify(crypto_ctx_t *, crypto_data_t *, crypto_data_t *,
286     crypto_req_handle_t);
287 static int dca_verify_update(crypto_ctx_t *, crypto_data_t *,
288     crypto_req_handle_t);
289 static int dca_verify_final(crypto_ctx_t *, crypto_data_t *,
290     crypto_req_handle_t);
291 static int dca_verify_atomic(crypto_provider_handle_t, crypto_session_id_t,
292     crypto_mechanism_t *, crypto_key_t *, crypto_data_t *,
293     crypto_data_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
294 static int dca_verify_recover_init(crypto_ctx_t *, crypto_mechanism_t *,
295     crypto_key_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
296 static int dca_verify_recover(crypto_ctx_t *, crypto_data_t *,
297     crypto_data_t *, crypto_req_handle_t);
298 static int dca_verify_recover_atomic(crypto_provider_handle_t,
299     crypto_session_id_t, crypto_mechanism_t *, crypto_key_t *, crypto_data_t *,
300     crypto_data_t *, crypto_spi_ctx_template_t, crypto_req_handle_t);
301 
302 static crypto_verify_ops_t dca_verify_ops = {
303 	dca_verify_init,
304 	dca_verify,
305 	dca_verify_update,
306 	dca_verify_final,
307 	dca_verify_atomic,
308 	dca_verify_recover_init,
309 	dca_verify_recover,
310 	dca_verify_recover_atomic
311 };
312 
313 static int dca_generate_random(crypto_provider_handle_t, crypto_session_id_t,
314     uchar_t *, size_t, crypto_req_handle_t);
315 
316 static crypto_random_number_ops_t dca_random_number_ops = {
317 	NULL,
318 	dca_generate_random
319 };
320 
321 static int ext_info_sym(crypto_provider_handle_t prov,
322     crypto_provider_ext_info_t *ext_info, crypto_req_handle_t cfreq);
323 static int ext_info_asym(crypto_provider_handle_t prov,
324     crypto_provider_ext_info_t *ext_info, crypto_req_handle_t cfreq);
325 static int ext_info_base(crypto_provider_handle_t prov,
326     crypto_provider_ext_info_t *ext_info, crypto_req_handle_t cfreq, char *id);
327 
328 static crypto_provider_management_ops_t dca_provmanage_ops_1 = {
329 	ext_info_sym,		/* ext_info */
330 	NULL,			/* init_token */
331 	NULL,			/* init_pin */
332 	NULL			/* set_pin */
333 };
334 
335 static crypto_provider_management_ops_t dca_provmanage_ops_2 = {
336 	ext_info_asym,		/* ext_info */
337 	NULL,			/* init_token */
338 	NULL,			/* init_pin */
339 	NULL			/* set_pin */
340 };
341 
342 int dca_free_context(crypto_ctx_t *);
343 
344 static crypto_ctx_ops_t dca_ctx_ops = {
345 	NULL,
346 	dca_free_context
347 };
348 
349 /* Operations for the symmetric cipher provider */
350 static crypto_ops_t dca_crypto_ops1 = {
351 	&dca_control_ops,
352 	NULL,				/* digest_ops */
353 	&dca_cipher_ops,
354 	NULL,				/* mac_ops */
355 	NULL,				/* sign_ops */
356 	NULL,				/* verify_ops */
357 	NULL,				/* dual_ops */
358 	NULL,				/* cipher_mac_ops */
359 	NULL,				/* random_number_ops */
360 	NULL,				/* session_ops */
361 	NULL,				/* object_ops */
362 	NULL,				/* key_ops */
363 	&dca_provmanage_ops_1,		/* management_ops */
364 	&dca_ctx_ops
365 };
366 
367 /* Operations for the asymmetric cipher provider */
368 static crypto_ops_t dca_crypto_ops2 = {
369 	&dca_control_ops,
370 	NULL,				/* digest_ops */
371 	&dca_cipher_ops,
372 	NULL,				/* mac_ops */
373 	&dca_sign_ops,
374 	&dca_verify_ops,
375 	NULL,				/* dual_ops */
376 	NULL,				/* cipher_mac_ops */
377 	&dca_random_number_ops,
378 	NULL,				/* session_ops */
379 	NULL,				/* object_ops */
380 	NULL,				/* key_ops */
381 	&dca_provmanage_ops_2,		/* management_ops */
382 	&dca_ctx_ops
383 };
384 
385 /* Provider information for the symmetric cipher provider */
386 static crypto_provider_info_t dca_prov_info1 = {
387 	CRYPTO_SPI_VERSION_1,
388 	NULL,				/* pi_provider_description */
389 	CRYPTO_HW_PROVIDER,
390 	NULL,				/* pi_provider_dev */
391 	NULL,				/* pi_provider_handle */
392 	&dca_crypto_ops1,
393 	sizeof (dca_mech_info_tab1)/sizeof (crypto_mech_info_t),
394 	dca_mech_info_tab1,
395 	0,				/* pi_logical_provider_count */
396 	NULL				/* pi_logical_providers */
397 };
398 
399 /* Provider information for the asymmetric cipher provider */
400 static crypto_provider_info_t dca_prov_info2 = {
401 	CRYPTO_SPI_VERSION_1,
402 	NULL,				/* pi_provider_description */
403 	CRYPTO_HW_PROVIDER,
404 	NULL,				/* pi_provider_dev */
405 	NULL,				/* pi_provider_handle */
406 	&dca_crypto_ops2,
407 	sizeof (dca_mech_info_tab2)/sizeof (crypto_mech_info_t),
408 	dca_mech_info_tab2,
409 	0,				/* pi_logical_provider_count */
410 	NULL				/* pi_logical_providers */
411 };
412 
413 /* Convenience macros */
414 /* Retrieve the softc and instance number from a SPI crypto context */
415 #define	DCA_SOFTC_FROM_CTX(ctx, softc, instance) {		\
416 	(softc) = (dca_t *)(ctx)->cc_provider;			\
417 	(instance) = ddi_get_instance((softc)->dca_dip);	\
418 }
419 
420 #define	DCA_MECH_FROM_CTX(ctx) \
421 	(((dca_request_t *)(ctx)->cc_provider_private)->dr_ctx.ctx_cm_type)
422 
423 static int dca_bindchains_one(dca_request_t *reqp, size_t cnt, int dr_offset,
424     caddr_t kaddr, ddi_dma_handle_t handle, uint_t flags,
425     dca_chain_t *head, int *n_chain);
426 static uint64_t dca_ena(uint64_t ena);
427 static caddr_t dca_bufdaddr_out(crypto_data_t *data);
428 static char *dca_fma_eclass_string(char *model, dca_fma_eclass_t index);
429 static int dca_check_acc_handle(dca_t *dca, ddi_acc_handle_t handle,
430     dca_fma_eclass_t eclass_index);
431 
432 static void dca_fma_init(dca_t *dca);
433 static void dca_fma_fini(dca_t *dca);
434 static int dca_fm_error_cb(dev_info_t *dip, ddi_fm_error_t *err,
435     const void *impl_data);
436 
437 
438 static dca_device_t dca_devices[] = {
439 	/* Broadcom vanilla variants */
440 	{	0x14e4, 0x5820, "Broadcom 5820" },
441 	{	0x14e4, 0x5821, "Broadcom 5821" },
442 	{	0x14e4, 0x5822, "Broadcom 5822" },
443 	{	0x14e4, 0x5825, "Broadcom 5825" },
444 	/* Sun specific OEMd variants */
445 	{	0x108e, 0x5454, "SCA" },
446 	{	0x108e, 0x5455, "SCA 1000" },
447 	{	0x108e, 0x5457, "SCA 500" },
448 	/* subsysid should be 0x5457, but got 0x1 from HW. Assume both here. */
449 	{	0x108e, 0x1, "SCA 500" },
450 };
451 
452 /*
453  * Device attributes.
454  */
455 static struct ddi_device_acc_attr dca_regsattr = {
456 	DDI_DEVICE_ATTR_V0,
457 	DDI_STRUCTURE_LE_ACC,
458 	DDI_STRICTORDER_ACC,
459 	DDI_FLAGERR_ACC
460 };
461 
462 static struct ddi_device_acc_attr dca_devattr = {
463 	DDI_DEVICE_ATTR_V0,
464 	DDI_STRUCTURE_LE_ACC,
465 	DDI_STRICTORDER_ACC,
466 	DDI_FLAGERR_ACC
467 };
468 
469 #if !defined(i386) && !defined(__i386)
470 static struct ddi_device_acc_attr dca_bufattr = {
471 	DDI_DEVICE_ATTR_V0,
472 	DDI_NEVERSWAP_ACC,
473 	DDI_STRICTORDER_ACC,
474 	DDI_FLAGERR_ACC
475 };
476 #endif
477 
478 static struct ddi_dma_attr dca_dmaattr = {
479 	DMA_ATTR_V0,		/* dma_attr_version */
480 	0x0,			/* dma_attr_addr_lo */
481 	0xffffffffUL,		/* dma_attr_addr_hi */
482 	0x00ffffffUL,		/* dma_attr_count_max */
483 	0x40,			/* dma_attr_align */
484 	0x40,			/* dma_attr_burstsizes */
485 	0x1,			/* dma_attr_minxfer */
486 	0x00ffffffUL,		/* dma_attr_maxxfer */
487 	0xffffffffUL,		/* dma_attr_seg */
488 #if defined(i386) || defined(__i386) || defined(__amd64)
489 	512,			/* dma_attr_sgllen */
490 #else
491 	1,			/* dma_attr_sgllen */
492 #endif
493 	1,			/* dma_attr_granular */
494 	DDI_DMA_FLAGERR		/* dma_attr_flags */
495 };
496 
497 static void	*dca_state = NULL;
498 int	dca_mindma = 2500;
499 
500 /*
501  * FMA eclass string definitions. Note that these string arrays must be
502  * consistent with the dca_fma_eclass_t enum.
503  */
504 static char *dca_fma_eclass_sca1000[] = {
505 	"sca1000.hw.device",
506 	"sca1000.hw.timeout",
507 	"sca1000.none"
508 };
509 
510 static char *dca_fma_eclass_sca500[] = {
511 	"sca500.hw.device",
512 	"sca500.hw.timeout",
513 	"sca500.none"
514 };
515 
516 /*
517  * DDI entry points.
518  */
519 int
520 _init(void)
521 {
522 	int rv;
523 
524 	DBG(NULL, DMOD, "dca: in _init");
525 
526 	if ((rv = ddi_soft_state_init(&dca_state, sizeof (dca_t), 1)) != 0) {
527 		/* this should *never* happen! */
528 		return (rv);
529 	}
530 
531 	if ((rv = mod_install(&modlinkage)) != 0) {
532 		/* cleanup here */
533 		ddi_soft_state_fini(&dca_state);
534 		return (rv);
535 	}
536 
537 	return (0);
538 }
539 
540 int
541 _fini(void)
542 {
543 	int rv;
544 
545 	DBG(NULL, DMOD, "dca: in _fini");
546 
547 	if ((rv = mod_remove(&modlinkage)) == 0) {
548 		/* cleanup here */
549 		ddi_soft_state_fini(&dca_state);
550 	}
551 	return (rv);
552 }
553 
554 int
555 _info(struct modinfo *modinfop)
556 {
557 	DBG(NULL, DMOD, "dca: in _info");
558 
559 	return (mod_info(&modlinkage, modinfop));
560 }
561 
562 int
563 dca_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
564 {
565 	ddi_acc_handle_t	pci;
566 	int			instance;
567 	ddi_iblock_cookie_t	ibc;
568 	int			intr_added = 0;
569 	dca_t			*dca;
570 	ushort_t		venid;
571 	ushort_t		devid;
572 	ushort_t		revid;
573 	ushort_t		subsysid;
574 	ushort_t		subvenid;
575 	int			i;
576 	int			ret;
577 	char			ID[64];
578 	static char		*unknowndev = "Unknown device";
579 
580 #if DEBUG
581 	/* these are only used for debugging */
582 	ushort_t		pcicomm;
583 	ushort_t		pcistat;
584 	uchar_t			cachelinesz;
585 	uchar_t			mingnt;
586 	uchar_t			maxlat;
587 	uchar_t			lattmr;
588 #endif
589 
590 	instance = ddi_get_instance(dip);
591 
592 	DBG(NULL, DMOD, "dca: in dca_attach() for %d", instance);
593 
594 	switch (cmd) {
595 	case DDI_RESUME:
596 		if ((dca = (dca_t *)ddi_get_driver_private(dip)) == NULL) {
597 			dca_diperror(dip, "no soft state in detach");
598 			return (DDI_FAILURE);
599 		}
600 		/* assumption: we won't be DDI_DETACHed until we return */
601 		return (dca_resume(dca));
602 	case DDI_ATTACH:
603 		break;
604 	default:
605 		return (DDI_FAILURE);
606 	}
607 
608 	if (ddi_slaveonly(dip) == DDI_SUCCESS) {
609 		dca_diperror(dip, "slot does not support PCI bus-master");
610 		return (DDI_FAILURE);
611 	}
612 
613 	if (ddi_intr_hilevel(dip, 0) != 0) {
614 		dca_diperror(dip, "hilevel interrupts not supported");
615 		return (DDI_FAILURE);
616 	}
617 
618 	if (pci_config_setup(dip, &pci) != DDI_SUCCESS) {
619 		dca_diperror(dip, "unable to setup PCI config handle");
620 		return (DDI_FAILURE);
621 	}
622 
623 	/* common PCI attributes */
624 	venid = pci_config_get16(pci, PCI_VENID);
625 	devid = pci_config_get16(pci, PCI_DEVID);
626 	revid = pci_config_get8(pci, PCI_REVID);
627 	subvenid = pci_config_get16(pci, PCI_SUBVENID);
628 	subsysid = pci_config_get16(pci, PCI_SUBSYSID);
629 
630 	/*
631 	 * Broadcom-specific timings.
632 	 * We disable these timers/counters since they can cause
633 	 * incorrect false failures when the bus is just a little
634 	 * bit slow, or busy.
635 	 */
636 	pci_config_put8(pci, PCI_TRDYTO, 0);
637 	pci_config_put8(pci, PCI_RETRIES, 0);
638 
639 	/* initialize PCI access settings */
640 	pci_config_put16(pci, PCI_COMM, PCICOMM_SEE |
641 	    PCICOMM_PEE | PCICOMM_BME | PCICOMM_MAE);
642 
643 	/* set up our PCI latency timer */
644 	pci_config_put8(pci, PCI_LATTMR, 0x40);
645 
646 #if DEBUG
647 	/* read registers (for debugging) */
648 	pcicomm = pci_config_get16(pci, PCI_COMM);
649 	pcistat = pci_config_get16(pci, PCI_STATUS);
650 	cachelinesz = pci_config_get8(pci, PCI_CACHELINESZ);
651 	mingnt = pci_config_get8(pci, PCI_MINGNT);
652 	maxlat = pci_config_get8(pci, PCI_MAXLAT);
653 	lattmr = pci_config_get8(pci, PCI_LATTMR);
654 #endif
655 
656 	pci_config_teardown(&pci);
657 
658 	if (ddi_get_iblock_cookie(dip, 0, &ibc) != DDI_SUCCESS) {
659 		dca_diperror(dip, "unable to get iblock cookie");
660 		return (DDI_FAILURE);
661 	}
662 
663 	if (ddi_soft_state_zalloc(dca_state, instance) != DDI_SUCCESS) {
664 		dca_diperror(dip, "unable to allocate soft state");
665 		return (DDI_FAILURE);
666 	}
667 
668 	dca = ddi_get_soft_state(dca_state, instance);
669 	ASSERT(dca != NULL);
670 	dca->dca_dip = dip;
671 	WORKLIST(dca, MCR1)->dwl_prov = NULL;
672 	WORKLIST(dca, MCR2)->dwl_prov = NULL;
673 	/* figure pagesize */
674 	dca->dca_pagesize = ddi_ptob(dip, 1);
675 
676 	/*
677 	 * Search for the device in our supported devices table.  This
678 	 * is here for two reasons.  First, we want to ensure that
679 	 * only Sun-qualified (and presumably Sun-labeled) devices can
680 	 * be used with this driver.  Second, some devices have
681 	 * specific differences.  E.g. the 5821 has support for a
682 	 * special mode of RC4, deeper queues, power management, and
683 	 * other changes.  Also, the export versions of some of these
684 	 * chips don't support RC4 or 3DES, so we catch that here.
685 	 *
686 	 * Note that we only look at the upper nibble of the device
687 	 * id, which is used to distinguish export vs. domestic
688 	 * versions of the chip.  (The lower nibble is used for
689 	 * stepping information.)
690 	 */
691 	for (i = 0; i < (sizeof (dca_devices) / sizeof (dca_device_t)); i++) {
692 		/*
693 		 * Try to match the subsystem information first.
694 		 */
695 		if (subvenid && (subvenid == dca_devices[i].dd_vendor_id) &&
696 		    subsysid && (subsysid == dca_devices[i].dd_device_id)) {
697 			dca->dca_model = dca_devices[i].dd_model;
698 			dca->dca_devid = dca_devices[i].dd_device_id;
699 			break;
700 		}
701 		/*
702 		 * Failing that, try the generic vendor and device id.
703 		 * Even if we find a match, we keep searching anyway,
704 		 * since we would prefer to find a match based on the
705 		 * subsystem ids.
706 		 */
707 		if ((venid == dca_devices[i].dd_vendor_id) &&
708 		    (devid == dca_devices[i].dd_device_id)) {
709 			dca->dca_model = dca_devices[i].dd_model;
710 			dca->dca_devid = dca_devices[i].dd_device_id;
711 		}
712 	}
713 	/* try and handle an unrecognized device */
714 	if (dca->dca_model == NULL) {
715 		dca->dca_model = unknowndev;
716 		dca_error(dca, "device not recognized, not supported");
717 		DBG(dca, DPCI, "i=%d venid=%x devid=%x rev=%d",
718 		    i, venid, devid, revid);
719 	}
720 
721 	if (ddi_prop_update_string(DDI_DEV_T_NONE, dip, "description",
722 	    dca->dca_model) != DDI_SUCCESS) {
723 		dca_error(dca, "unable to create description property");
724 		return (DDI_FAILURE);
725 	}
726 
727 	DBG(dca, DPCI, "PCI command=0x%x status=%x cachelinesz=%x",
728 	    pcicomm, pcistat, cachelinesz);
729 	DBG(dca, DPCI, "mingnt=0x%x maxlat=0x%x lattmr=0x%x",
730 	    mingnt, maxlat, lattmr);
731 
732 	/*
733 	 * initialize locks, etc.
734 	 */
735 	(void) mutex_init(&dca->dca_intrlock, NULL, MUTEX_DRIVER, ibc);
736 
737 	/* use RNGSHA1 by default */
738 	if (ddi_getprop(DDI_DEV_T_ANY, dip,
739 	    DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS, "rngdirect", 0) == 0) {
740 		dca->dca_flags |= DCA_RNGSHA1;
741 	}
742 
743 	/* initialize FMA */
744 	dca_fma_init(dca);
745 
746 	/* initialize some key data structures */
747 	if (dca_init(dca) != DDI_SUCCESS) {
748 		goto failed;
749 	}
750 
751 	/* initialize kstats */
752 	dca_ksinit(dca);
753 
754 	/* setup access to registers */
755 	if (ddi_regs_map_setup(dip, 1, (caddr_t *)&dca->dca_regs,
756 	    0, 0, &dca_regsattr, &dca->dca_regs_handle) != DDI_SUCCESS) {
757 		dca_error(dca, "unable to map registers");
758 		goto failed;
759 	}
760 
761 	DBG(dca, DCHATTY, "MCR1 = %x", GETCSR(dca, CSR_MCR1));
762 	DBG(dca, DCHATTY, "CONTROL = %x", GETCSR(dca, CSR_DMACTL));
763 	DBG(dca, DCHATTY, "STATUS = %x", GETCSR(dca, CSR_DMASTAT));
764 	DBG(dca, DCHATTY, "DMAEA = %x", GETCSR(dca, CSR_DMAEA));
765 	DBG(dca, DCHATTY, "MCR2 = %x", GETCSR(dca, CSR_MCR2));
766 
767 	/* reset the chip */
768 	if (dca_reset(dca, 0) < 0) {
769 		goto failed;
770 	}
771 
772 	/* initialize the chip */
773 	PUTCSR(dca, CSR_DMACTL, DMACTL_BE32 | DMACTL_BE64);
774 	if (dca_check_acc_handle(dca, dca->dca_regs_handle,
775 	    DCA_FM_ECLASS_NONE) != DDI_SUCCESS) {
776 		goto failed;
777 	}
778 
779 	/* add the interrupt */
780 	if (ddi_add_intr(dip, 0, &dca->dca_icookie, NULL, dca_intr,
781 	    (void *)dca) != DDI_SUCCESS) {
782 		DBG(dca, DWARN, "ddi_add_intr failed");
783 		goto failed;
784 	} else {
785 		intr_added = 1;
786 	}
787 
788 	/* enable interrupts on the device */
789 	/*
790 	 * XXX: Note, 5820A1 errata indicates that this may clobber
791 	 * bits 24 and 23, which affect the speed of the RNG.  Since
792 	 * we always want to run in full-speed mode, this should be
793 	 * harmless.
794 	 */
795 	if (dca->dca_devid == 0x5825) {
796 		/* for 5825 - increase the DMA read size */
797 		SETBIT(dca, CSR_DMACTL,
798 		    DMACTL_MCR1IE | DMACTL_MCR2IE | DMACTL_EIE | DMACTL_RD256);
799 	} else {
800 		SETBIT(dca, CSR_DMACTL,
801 		    DMACTL_MCR1IE | DMACTL_MCR2IE | DMACTL_EIE);
802 	}
803 	if (dca_check_acc_handle(dca, dca->dca_regs_handle,
804 	    DCA_FM_ECLASS_NONE) != DDI_SUCCESS) {
805 		goto failed;
806 	}
807 
808 	/* register MCR1 with the crypto framework */
809 	/* Be careful not to exceed 32 chars */
810 	(void) sprintf(ID, "%s/%d %s",
811 	    ddi_driver_name(dip), ddi_get_instance(dip), IDENT_SYM);
812 	dca_prov_info1.pi_provider_description = ID;
813 	dca_prov_info1.pi_provider_dev.pd_hw = dip;
814 	dca_prov_info1.pi_provider_handle = dca;
815 	if ((ret = crypto_register_provider(&dca_prov_info1,
816 	    &WORKLIST(dca, MCR1)->dwl_prov)) != CRYPTO_SUCCESS) {
817 		cmn_err(CE_WARN,
818 		    "crypto_register_provider() failed (%d) for MCR1", ret);
819 		goto failed;
820 	}
821 
822 	/* register MCR2 with the crypto framework */
823 	/* Be careful not to exceed 32 chars */
824 	(void) sprintf(ID, "%s/%d %s",
825 	    ddi_driver_name(dip), ddi_get_instance(dip), IDENT_ASYM);
826 	dca_prov_info2.pi_provider_description = ID;
827 	dca_prov_info2.pi_provider_dev.pd_hw = dip;
828 	dca_prov_info2.pi_provider_handle = dca;
829 	if ((ret = crypto_register_provider(&dca_prov_info2,
830 	    &WORKLIST(dca, MCR2)->dwl_prov)) != CRYPTO_SUCCESS) {
831 		cmn_err(CE_WARN,
832 		    "crypto_register_provider() failed (%d) for MCR2", ret);
833 		goto failed;
834 	}
835 
836 	crypto_prov_notify(WORKLIST(dca, MCR1)->dwl_prov,
837 	    CRYPTO_PROVIDER_READY);
838 	crypto_prov_notify(WORKLIST(dca, MCR2)->dwl_prov,
839 	    CRYPTO_PROVIDER_READY);
840 
841 	/* Initialize the local random number pool for this instance */
842 	if ((ret = dca_random_init(dca)) != CRYPTO_SUCCESS) {
843 		goto failed;
844 	}
845 
846 	mutex_enter(&dca->dca_intrlock);
847 	dca->dca_jobtid = timeout(dca_jobtimeout, (void *)dca,
848 	    drv_usectohz(SECOND));
849 	mutex_exit(&dca->dca_intrlock);
850 
851 	ddi_set_driver_private(dip, (caddr_t)dca);
852 
853 	ddi_report_dev(dip);
854 
855 	if (ddi_get_devstate(dca->dca_dip) != DDI_DEVSTATE_UP) {
856 		ddi_fm_service_impact(dca->dca_dip, DDI_SERVICE_RESTORED);
857 	}
858 
859 	return (DDI_SUCCESS);
860 
861 failed:
862 	/* unregister from the crypto framework */
863 	if (WORKLIST(dca, MCR1)->dwl_prov != NULL) {
864 		(void) crypto_unregister_provider(
865 		    WORKLIST(dca, MCR1)->dwl_prov);
866 	}
867 	if (WORKLIST(dca, MCR2)->dwl_prov != NULL) {
868 		(void) crypto_unregister_provider(
869 		    WORKLIST(dca, MCR2)->dwl_prov);
870 	}
871 	if (intr_added) {
872 		CLRBIT(dca, CSR_DMACTL,
873 		    DMACTL_MCR1IE | DMACTL_MCR2IE | DMACTL_EIE);
874 		/* unregister intr handler */
875 		ddi_remove_intr(dip, 0, dca->dca_icookie);
876 	}
877 	if (dca->dca_regs_handle) {
878 		ddi_regs_map_free(&dca->dca_regs_handle);
879 	}
880 	if (dca->dca_intrstats) {
881 		kstat_delete(dca->dca_intrstats);
882 	}
883 	if (dca->dca_ksp) {
884 		kstat_delete(dca->dca_ksp);
885 	}
886 	dca_uninit(dca);
887 
888 	/* finalize FMA */
889 	dca_fma_fini(dca);
890 
891 	mutex_destroy(&dca->dca_intrlock);
892 	ddi_soft_state_free(dca_state, instance);
893 	return (DDI_FAILURE);
894 
895 }
896 
897 int
898 dca_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
899 {
900 	int		instance;
901 	dca_t		*dca;
902 	timeout_id_t	tid;
903 
904 	instance = ddi_get_instance(dip);
905 
906 	DBG(NULL, DMOD, "dca: in dca_detach() for %d", instance);
907 
908 	switch (cmd) {
909 	case DDI_SUSPEND:
910 		if ((dca = (dca_t *)ddi_get_driver_private(dip)) == NULL) {
911 			dca_diperror(dip, "no soft state in detach");
912 			return (DDI_FAILURE);
913 		}
914 		/* assumption: we won't be DDI_DETACHed until we return */
915 		return (dca_suspend(dca));
916 
917 	case DDI_DETACH:
918 		break;
919 	default:
920 		return (DDI_FAILURE);
921 	}
922 
923 	if ((dca = (dca_t *)ddi_get_driver_private(dip)) == NULL) {
924 		dca_diperror(dip, "no soft state in detach");
925 		return (DDI_FAILURE);
926 	}
927 
928 	/*
929 	 * Unregister from kCF.
930 	 * This needs to be done at the beginning of detach.
931 	 */
932 	if (WORKLIST(dca, MCR1)->dwl_prov != NULL) {
933 		if (crypto_unregister_provider(
934 		    WORKLIST(dca, MCR1)->dwl_prov) != CRYPTO_SUCCESS) {
935 			dca_error(dca, "unable to unregister MCR1 from kcf");
936 			return (DDI_FAILURE);
937 		}
938 	}
939 
940 	if (WORKLIST(dca, MCR2)->dwl_prov != NULL) {
941 		if (crypto_unregister_provider(
942 		    WORKLIST(dca, MCR2)->dwl_prov) != CRYPTO_SUCCESS) {
943 			dca_error(dca, "unable to unregister MCR2 from kcf");
944 			return (DDI_FAILURE);
945 		}
946 	}
947 
948 	/*
949 	 * Cleanup the private context list. Once the
950 	 * crypto_unregister_provider returns, it is safe to do so.
951 	 */
952 	dca_free_context_list(dca);
953 
954 	/* Cleanup the local random number pool */
955 	dca_random_fini(dca);
956 
957 	/* send any jobs in the waitq back to kCF */
958 	dca_rejectjobs(dca);
959 
960 	/* untimeout the timeouts */
961 	mutex_enter(&dca->dca_intrlock);
962 	tid = dca->dca_jobtid;
963 	dca->dca_jobtid = 0;
964 	mutex_exit(&dca->dca_intrlock);
965 	if (tid) {
966 		(void) untimeout(tid);
967 	}
968 
969 	/* disable device interrupts */
970 	CLRBIT(dca, CSR_DMACTL, DMACTL_MCR1IE | DMACTL_MCR2IE | DMACTL_EIE);
971 
972 	/* unregister interrupt handlers */
973 	ddi_remove_intr(dip, 0, dca->dca_icookie);
974 
975 	/* release our regs handle */
976 	ddi_regs_map_free(&dca->dca_regs_handle);
977 
978 	/* toss out kstats */
979 	if (dca->dca_intrstats) {
980 		kstat_delete(dca->dca_intrstats);
981 	}
982 	if (dca->dca_ksp) {
983 		kstat_delete(dca->dca_ksp);
984 	}
985 
986 	mutex_destroy(&dca->dca_intrlock);
987 	dca_uninit(dca);
988 
989 	/* finalize FMA */
990 	dca_fma_fini(dca);
991 
992 	ddi_soft_state_free(dca_state, instance);
993 
994 	return (DDI_SUCCESS);
995 }
996 
997 int
998 dca_resume(dca_t *dca)
999 {
1000 	ddi_acc_handle_t	pci;
1001 
1002 	if (pci_config_setup(dca->dca_dip, &pci) != DDI_SUCCESS) {
1003 		dca_error(dca, "unable to setup PCI config handle");
1004 		return (DDI_FAILURE);
1005 	}
1006 
1007 	/*
1008 	 * Reprogram registers in PCI configuration space.
1009 	 */
1010 
1011 	/* Broadcom-specific timers -- we disable them. */
1012 	pci_config_put8(pci, PCI_TRDYTO, 0);
1013 	pci_config_put8(pci, PCI_RETRIES, 0);
1014 
1015 	/* initialize PCI access settings */
1016 	pci_config_put16(pci, PCI_COMM, PCICOMM_SEE |
1017 	    PCICOMM_PEE | PCICOMM_BME | PCICOMM_MAE);
1018 
1019 	/* set up our PCI latency timer */
1020 	pci_config_put8(pci, PCI_LATTMR, 0x40);
1021 
1022 	pci_config_teardown(&pci);
1023 
1024 	if (dca_reset(dca, 0) < 0) {
1025 		dca_error(dca, "unable to reset device during resume");
1026 		return (DDI_FAILURE);
1027 	}
1028 
1029 	/*
1030 	 * Now restore the card-specific CSRs.
1031 	 */
1032 
1033 	/* restore endianness settings */
1034 	PUTCSR(dca, CSR_DMACTL, DMACTL_BE32 | DMACTL_BE64);
1035 	if (dca_check_acc_handle(dca, dca->dca_regs_handle,
1036 	    DCA_FM_ECLASS_NONE) != DDI_SUCCESS)
1037 		return (DDI_FAILURE);
1038 
1039 	/* restore interrupt enables */
1040 	if (dca->dca_devid == 0x5825) {
1041 		/* for 5825 set 256 byte read size to improve performance */
1042 		SETBIT(dca, CSR_DMACTL,
1043 		    DMACTL_MCR1IE | DMACTL_MCR2IE | DMACTL_EIE | DMACTL_RD256);
1044 	} else {
1045 		SETBIT(dca, CSR_DMACTL,
1046 		    DMACTL_MCR1IE | DMACTL_MCR2IE | DMACTL_EIE);
1047 	}
1048 	if (dca_check_acc_handle(dca, dca->dca_regs_handle,
1049 	    DCA_FM_ECLASS_NONE) != DDI_SUCCESS)
1050 		return (DDI_FAILURE);
1051 
1052 	/* resume scheduling jobs on the device */
1053 	dca_undrain(dca);
1054 
1055 	return (DDI_SUCCESS);
1056 }
1057 
1058 int
1059 dca_suspend(dca_t *dca)
1060 {
1061 	if ((dca_drain(dca)) != 0) {
1062 		return (DDI_FAILURE);
1063 	}
1064 	if (dca_reset(dca, 0) < 0) {
1065 		dca_error(dca, "unable to reset device during suspend");
1066 		return (DDI_FAILURE);
1067 	}
1068 	return (DDI_SUCCESS);
1069 }
1070 
1071 /*
1072  * Hardware access stuff.
1073  */
1074 int
1075 dca_reset(dca_t *dca, int failreset)
1076 {
1077 	int i;
1078 
1079 	if (dca->dca_regs_handle == NULL) {
1080 		return (-1);
1081 	}
1082 
1083 	PUTCSR(dca, CSR_DMACTL, DMACTL_RESET);
1084 	if (!failreset) {
1085 		if (dca_check_acc_handle(dca, dca->dca_regs_handle,
1086 		    DCA_FM_ECLASS_NONE) != DDI_SUCCESS)
1087 			return (-1);
1088 	}
1089 
1090 	/* now wait for a reset */
1091 	for (i = 1; i < 100; i++) {
1092 		uint32_t	dmactl;
1093 		drv_usecwait(100);
1094 		dmactl = GETCSR(dca, CSR_DMACTL);
1095 		if (!failreset) {
1096 			if (dca_check_acc_handle(dca, dca->dca_regs_handle,
1097 			    DCA_FM_ECLASS_NONE) != DDI_SUCCESS)
1098 				return (-1);
1099 		}
1100 		if ((dmactl & DMACTL_RESET) == 0) {
1101 			DBG(dca, DCHATTY, "reset in %d usec", i * 100);
1102 			return (0);
1103 		}
1104 	}
1105 	if (!failreset) {
1106 		dca_failure(dca, DDI_DEVICE_FAULT,
1107 		    DCA_FM_ECLASS_NONE, dca_ena(0), CRYPTO_DEVICE_ERROR,
1108 		    "timeout waiting for reset after %d usec", i * 100);
1109 	}
1110 	return (-1);
1111 }
1112 
1113 int
1114 dca_initworklist(dca_t *dca, dca_worklist_t *wlp)
1115 {
1116 	int	i;
1117 	int	reqprealloc = wlp->dwl_hiwater + (MAXWORK * MAXREQSPERMCR);
1118 
1119 	/*
1120 	 * Set up work queue.
1121 	 */
1122 	mutex_init(&wlp->dwl_lock, NULL, MUTEX_DRIVER, dca->dca_icookie);
1123 	mutex_init(&wlp->dwl_freereqslock, NULL, MUTEX_DRIVER,
1124 	    dca->dca_icookie);
1125 	mutex_init(&wlp->dwl_freelock, NULL, MUTEX_DRIVER, dca->dca_icookie);
1126 	cv_init(&wlp->dwl_cv, NULL, CV_DRIVER, NULL);
1127 
1128 	mutex_enter(&wlp->dwl_lock);
1129 
1130 	dca_initq(&wlp->dwl_freereqs);
1131 	dca_initq(&wlp->dwl_waitq);
1132 	dca_initq(&wlp->dwl_freework);
1133 	dca_initq(&wlp->dwl_runq);
1134 
1135 	for (i = 0; i < MAXWORK; i++) {
1136 		dca_work_t		*workp;
1137 
1138 		if ((workp = dca_newwork(dca)) == NULL) {
1139 			dca_error(dca, "unable to allocate work");
1140 			mutex_exit(&wlp->dwl_lock);
1141 			return (DDI_FAILURE);
1142 		}
1143 		workp->dw_wlp = wlp;
1144 		dca_freework(workp);
1145 	}
1146 	mutex_exit(&wlp->dwl_lock);
1147 
1148 	for (i = 0; i < reqprealloc; i++) {
1149 		dca_request_t *reqp;
1150 
1151 		if ((reqp = dca_newreq(dca)) == NULL) {
1152 			dca_error(dca, "unable to allocate request");
1153 			return (DDI_FAILURE);
1154 		}
1155 		reqp->dr_dca = dca;
1156 		reqp->dr_wlp = wlp;
1157 		dca_freereq(reqp);
1158 	}
1159 	return (DDI_SUCCESS);
1160 }
1161 
1162 int
1163 dca_init(dca_t *dca)
1164 {
1165 	dca_worklist_t		*wlp;
1166 
1167 	/* Initialize the private context list and the corresponding lock. */
1168 	mutex_init(&dca->dca_ctx_list_lock, NULL, MUTEX_DRIVER, NULL);
1169 	dca_initq(&dca->dca_ctx_list);
1170 
1171 	/*
1172 	 * MCR1 algorithms.
1173 	 */
1174 	wlp = WORKLIST(dca, MCR1);
1175 	(void) sprintf(wlp->dwl_name, "dca%d:mcr1",
1176 	    ddi_get_instance(dca->dca_dip));
1177 	wlp->dwl_lowater = ddi_getprop(DDI_DEV_T_ANY,
1178 	    dca->dca_dip, DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS,
1179 	    "mcr1_lowater", MCR1LOWATER);
1180 	wlp->dwl_hiwater = ddi_getprop(DDI_DEV_T_ANY,
1181 	    dca->dca_dip, DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS,
1182 	    "mcr1_hiwater", MCR1HIWATER);
1183 	wlp->dwl_reqspermcr = min(ddi_getprop(DDI_DEV_T_ANY,
1184 	    dca->dca_dip, DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS,
1185 	    "mcr1_maxreqs", MCR1MAXREQS), MAXREQSPERMCR);
1186 	wlp->dwl_dca = dca;
1187 	wlp->dwl_mcr = MCR1;
1188 	if (dca_initworklist(dca, wlp) != DDI_SUCCESS) {
1189 		return (DDI_FAILURE);
1190 	}
1191 
1192 	/*
1193 	 * MCR2 algorithms.
1194 	 */
1195 	wlp = WORKLIST(dca, MCR2);
1196 	(void) sprintf(wlp->dwl_name, "dca%d:mcr2",
1197 	    ddi_get_instance(dca->dca_dip));
1198 	wlp->dwl_lowater = ddi_getprop(DDI_DEV_T_ANY,
1199 	    dca->dca_dip, DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS,
1200 	    "mcr2_lowater", MCR2LOWATER);
1201 	wlp->dwl_hiwater = ddi_getprop(DDI_DEV_T_ANY,
1202 	    dca->dca_dip, DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS,
1203 	    "mcr2_hiwater", MCR2HIWATER);
1204 	wlp->dwl_reqspermcr = min(ddi_getprop(DDI_DEV_T_ANY,
1205 	    dca->dca_dip, DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS,
1206 	    "mcr2_maxreqs", MCR2MAXREQS), MAXREQSPERMCR);
1207 	wlp->dwl_dca = dca;
1208 	wlp->dwl_mcr = MCR2;
1209 	if (dca_initworklist(dca, wlp) != DDI_SUCCESS) {
1210 		return (DDI_FAILURE);
1211 	}
1212 	return (DDI_SUCCESS);
1213 }
1214 
1215 /*
1216  * Uninitialize worklists.  This routine should only be called when no
1217  * active jobs (hence DMA mappings) exist.  One way to ensure this is
1218  * to unregister from kCF before calling this routine.  (This is done
1219  * e.g. in detach(9e).)
1220  */
1221 void
1222 dca_uninit(dca_t *dca)
1223 {
1224 	int	mcr;
1225 
1226 	mutex_destroy(&dca->dca_ctx_list_lock);
1227 
1228 	for (mcr = MCR1; mcr <= MCR2; mcr++) {
1229 		dca_worklist_t	*wlp = WORKLIST(dca, mcr);
1230 		dca_work_t	*workp;
1231 		dca_request_t	*reqp;
1232 
1233 		if (dca->dca_regs_handle == NULL) {
1234 			continue;
1235 		}
1236 
1237 		mutex_enter(&wlp->dwl_lock);
1238 		while ((workp = dca_getwork(dca, mcr)) != NULL) {
1239 			dca_destroywork(workp);
1240 		}
1241 		mutex_exit(&wlp->dwl_lock);
1242 		while ((reqp = dca_getreq(dca, mcr, 0)) != NULL) {
1243 			dca_destroyreq(reqp);
1244 		}
1245 
1246 		mutex_destroy(&wlp->dwl_lock);
1247 		mutex_destroy(&wlp->dwl_freereqslock);
1248 		mutex_destroy(&wlp->dwl_freelock);
1249 		cv_destroy(&wlp->dwl_cv);
1250 		wlp->dwl_prov = NULL;
1251 	}
1252 }
1253 
1254 static void
1255 dca_enlist2(dca_listnode_t *q, dca_listnode_t *node, kmutex_t *lock)
1256 {
1257 	if (!q || !node)
1258 		return;
1259 
1260 	mutex_enter(lock);
1261 	node->dl_next2 = q;
1262 	node->dl_prev2 = q->dl_prev2;
1263 	node->dl_next2->dl_prev2 = node;
1264 	node->dl_prev2->dl_next2 = node;
1265 	mutex_exit(lock);
1266 }
1267 
1268 static void
1269 dca_rmlist2(dca_listnode_t *node, kmutex_t *lock)
1270 {
1271 	if (!node)
1272 		return;
1273 
1274 	mutex_enter(lock);
1275 	node->dl_next2->dl_prev2 = node->dl_prev2;
1276 	node->dl_prev2->dl_next2 = node->dl_next2;
1277 	node->dl_next2 = NULL;
1278 	node->dl_prev2 = NULL;
1279 	mutex_exit(lock);
1280 }
1281 
1282 static dca_listnode_t *
1283 dca_delist2(dca_listnode_t *q, kmutex_t *lock)
1284 {
1285 	dca_listnode_t *node;
1286 
1287 	mutex_enter(lock);
1288 	if ((node = q->dl_next2) == q) {
1289 		mutex_exit(lock);
1290 		return (NULL);
1291 	}
1292 
1293 	node->dl_next2->dl_prev2 = node->dl_prev2;
1294 	node->dl_prev2->dl_next2 = node->dl_next2;
1295 	node->dl_next2 = NULL;
1296 	node->dl_prev2 = NULL;
1297 	mutex_exit(lock);
1298 
1299 	return (node);
1300 }
1301 
1302 void
1303 dca_initq(dca_listnode_t *q)
1304 {
1305 	q->dl_next = q;
1306 	q->dl_prev = q;
1307 	q->dl_next2 = q;
1308 	q->dl_prev2 = q;
1309 }
1310 
1311 void
1312 dca_enqueue(dca_listnode_t *q, dca_listnode_t *node)
1313 {
1314 	/*
1315 	 * Enqueue submits at the "tail" of the list, i.e. just
1316 	 * behind the sentinel.
1317 	 */
1318 	node->dl_next = q;
1319 	node->dl_prev = q->dl_prev;
1320 	node->dl_next->dl_prev = node;
1321 	node->dl_prev->dl_next = node;
1322 }
1323 
1324 void
1325 dca_rmqueue(dca_listnode_t *node)
1326 {
1327 	node->dl_next->dl_prev = node->dl_prev;
1328 	node->dl_prev->dl_next = node->dl_next;
1329 	node->dl_next = NULL;
1330 	node->dl_prev = NULL;
1331 }
1332 
1333 dca_listnode_t *
1334 dca_dequeue(dca_listnode_t *q)
1335 {
1336 	dca_listnode_t *node;
1337 	/*
1338 	 * Dequeue takes from the "head" of the list, i.e. just after
1339 	 * the sentinel.
1340 	 */
1341 	if ((node = q->dl_next) == q) {
1342 		/* queue is empty */
1343 		return (NULL);
1344 	}
1345 	dca_rmqueue(node);
1346 	return (node);
1347 }
1348 
1349 /* this is the opposite of dequeue, it takes things off in LIFO order */
1350 dca_listnode_t *
1351 dca_unqueue(dca_listnode_t *q)
1352 {
1353 	dca_listnode_t *node;
1354 	/*
1355 	 * unqueue takes from the "tail" of the list, i.e. just before
1356 	 * the sentinel.
1357 	 */
1358 	if ((node = q->dl_prev) == q) {
1359 		/* queue is empty */
1360 		return (NULL);
1361 	}
1362 	dca_rmqueue(node);
1363 	return (node);
1364 }
1365 
1366 dca_listnode_t *
1367 dca_peekqueue(dca_listnode_t *q)
1368 {
1369 	dca_listnode_t *node;
1370 
1371 	if ((node = q->dl_next) == q) {
1372 		return (NULL);
1373 	} else {
1374 		return (node);
1375 	}
1376 }
1377 
1378 /*
1379  * Interrupt service routine.
1380  */
1381 uint_t
1382 dca_intr(char *arg)
1383 {
1384 	dca_t		*dca = (dca_t *)arg;
1385 	uint32_t	status;
1386 
1387 	mutex_enter(&dca->dca_intrlock);
1388 	status = GETCSR(dca, CSR_DMASTAT);
1389 	PUTCSR(dca, CSR_DMASTAT, status & DMASTAT_INTERRUPTS);
1390 	if (dca_check_acc_handle(dca, dca->dca_regs_handle,
1391 	    DCA_FM_ECLASS_NONE) != DDI_SUCCESS) {
1392 		mutex_exit(&dca->dca_intrlock);
1393 		return ((uint_t)DDI_FAILURE);
1394 	}
1395 
1396 	DBG(dca, DINTR, "interrupted, status = 0x%x!", status);
1397 
1398 	if ((status & DMASTAT_INTERRUPTS) == 0) {
1399 		/* increment spurious interrupt kstat */
1400 		if (dca->dca_intrstats) {
1401 			KIOIP(dca)->intrs[KSTAT_INTR_SPURIOUS]++;
1402 		}
1403 		mutex_exit(&dca->dca_intrlock);
1404 		return (DDI_INTR_UNCLAIMED);
1405 	}
1406 
1407 	if (dca->dca_intrstats) {
1408 		KIOIP(dca)->intrs[KSTAT_INTR_HARD]++;
1409 	}
1410 	if (status & DMASTAT_MCR1INT) {
1411 		DBG(dca, DINTR, "MCR1 interrupted");
1412 		mutex_enter(&(WORKLIST(dca, MCR1)->dwl_lock));
1413 		dca_schedule(dca, MCR1);
1414 		dca_reclaim(dca, MCR1);
1415 		mutex_exit(&(WORKLIST(dca, MCR1)->dwl_lock));
1416 	}
1417 
1418 	if (status & DMASTAT_MCR2INT) {
1419 		DBG(dca, DINTR, "MCR2 interrupted");
1420 		mutex_enter(&(WORKLIST(dca, MCR2)->dwl_lock));
1421 		dca_schedule(dca, MCR2);
1422 		dca_reclaim(dca, MCR2);
1423 		mutex_exit(&(WORKLIST(dca, MCR2)->dwl_lock));
1424 	}
1425 
1426 	if (status & DMASTAT_ERRINT) {
1427 		uint32_t	erraddr;
1428 		erraddr = GETCSR(dca, CSR_DMAEA);
1429 		mutex_exit(&dca->dca_intrlock);
1430 
1431 		/*
1432 		 * bit 1 of the error address indicates failure during
1433 		 * read if set, during write otherwise.
1434 		 */
1435 		dca_failure(dca, DDI_DEVICE_FAULT,
1436 		    DCA_FM_ECLASS_HW_DEVICE, dca_ena(0), CRYPTO_DEVICE_ERROR,
1437 		    "DMA master access error %s address 0x%x",
1438 		    erraddr & 0x1 ? "reading" : "writing", erraddr & ~1);
1439 		return (DDI_INTR_CLAIMED);
1440 	}
1441 
1442 	mutex_exit(&dca->dca_intrlock);
1443 
1444 	return (DDI_INTR_CLAIMED);
1445 }
1446 
1447 /*
1448  * Reverse a string of bytes from s1 into s2.  The reversal happens
1449  * from the tail of s1.  If len1 < len2, then null bytes will be
1450  * padded to the end of s2.  If len2 < len1, then (presumably null)
1451  * bytes will be dropped from the start of s1.
1452  *
1453  * The rationale here is that when s1 (source) is shorter, then we
1454  * are reversing from big-endian ordering, into device ordering, and
1455  * want to add some extra nulls to the tail (MSB) side of the device.
1456  *
1457  * Similarly, when s2 (dest) is shorter, then we are truncating what
1458  * are presumably null MSB bits from the device.
1459  *
1460  * There is an expectation when reversing from the device back into
1461  * big-endian, that the number of bytes to reverse and the target size
1462  * will match, and no truncation or padding occurs.
1463  */
1464 void
1465 dca_reverse(void *s1, void *s2, int len1, int len2)
1466 {
1467 	caddr_t	src, dst;
1468 
1469 	if (len1 == 0) {
1470 		if (len2) {
1471 			bzero(s2, len2);
1472 		}
1473 		return;
1474 	}
1475 	src = (caddr_t)s1 + len1 - 1;
1476 	dst = s2;
1477 	while ((src >= (caddr_t)s1) && (len2)) {
1478 		*dst++ = *src--;
1479 		len2--;
1480 	}
1481 	while (len2 > 0) {
1482 		*dst++ = 0;
1483 		len2--;
1484 	}
1485 }
1486 
1487 uint16_t
1488 dca_padfull(int num)
1489 {
1490 	if (num <= 512) {
1491 		return (BITS2BYTES(512));
1492 	}
1493 	if (num <= 768) {
1494 		return (BITS2BYTES(768));
1495 	}
1496 	if (num <= 1024) {
1497 		return (BITS2BYTES(1024));
1498 	}
1499 	if (num <= 1536) {
1500 		return (BITS2BYTES(1536));
1501 	}
1502 	if (num <= 2048) {
1503 		return (BITS2BYTES(2048));
1504 	}
1505 	return (0);
1506 }
1507 
1508 uint16_t
1509 dca_padhalf(int num)
1510 {
1511 	if (num <= 256) {
1512 		return (BITS2BYTES(256));
1513 	}
1514 	if (num <= 384) {
1515 		return (BITS2BYTES(384));
1516 	}
1517 	if (num <= 512) {
1518 		return (BITS2BYTES(512));
1519 	}
1520 	if (num <= 768) {
1521 		return (BITS2BYTES(768));
1522 	}
1523 	if (num <= 1024) {
1524 		return (BITS2BYTES(1024));
1525 	}
1526 	return (0);
1527 }
1528 
1529 dca_work_t *
1530 dca_newwork(dca_t *dca)
1531 {
1532 	dca_work_t		*workp;
1533 	size_t			size;
1534 	ddi_dma_cookie_t	c;
1535 	unsigned		nc;
1536 	int			rv;
1537 
1538 	workp = kmem_zalloc(sizeof (dca_work_t), KM_SLEEP);
1539 
1540 	rv = ddi_dma_alloc_handle(dca->dca_dip, &dca_dmaattr,
1541 	    DDI_DMA_SLEEP, NULL, &workp->dw_mcr_dmah);
1542 	if (rv != 0) {
1543 		dca_error(dca, "unable to alloc MCR DMA handle");
1544 		dca_destroywork(workp);
1545 		return (NULL);
1546 	}
1547 
1548 	rv = ddi_dma_mem_alloc(workp->dw_mcr_dmah,
1549 	    ROUNDUP(MCR_SIZE, dca->dca_pagesize),
1550 	    &dca_devattr, DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL,
1551 	    &workp->dw_mcr_kaddr, &size, &workp->dw_mcr_acch);
1552 	if (rv != 0) {
1553 		dca_error(dca, "unable to alloc MCR DMA memory");
1554 		dca_destroywork(workp);
1555 		return (NULL);
1556 	}
1557 
1558 	rv = ddi_dma_addr_bind_handle(workp->dw_mcr_dmah, NULL,
1559 	    workp->dw_mcr_kaddr, size, DDI_DMA_CONSISTENT | DDI_DMA_RDWR,
1560 	    DDI_DMA_SLEEP, NULL, &c, &nc);
1561 	if (rv != DDI_DMA_MAPPED) {
1562 		dca_error(dca, "unable to map MCR DMA memory");
1563 		dca_destroywork(workp);
1564 		return (NULL);
1565 	}
1566 
1567 	workp->dw_mcr_paddr = c.dmac_address;
1568 	return (workp);
1569 }
1570 
1571 void
1572 dca_destroywork(dca_work_t *workp)
1573 {
1574 	if (workp->dw_mcr_paddr) {
1575 		(void) ddi_dma_unbind_handle(workp->dw_mcr_dmah);
1576 	}
1577 	if (workp->dw_mcr_acch) {
1578 		ddi_dma_mem_free(&workp->dw_mcr_acch);
1579 	}
1580 	if (workp->dw_mcr_dmah) {
1581 		ddi_dma_free_handle(&workp->dw_mcr_dmah);
1582 	}
1583 	kmem_free(workp, sizeof (dca_work_t));
1584 }
1585 
1586 dca_request_t *
1587 dca_newreq(dca_t *dca)
1588 {
1589 	dca_request_t		*reqp;
1590 	size_t			size;
1591 	ddi_dma_cookie_t	c;
1592 	unsigned		nc;
1593 	int			rv;
1594 	int			n_chain = 0;
1595 
1596 	size = (DESC_SIZE * MAXFRAGS) + CTX_MAXLENGTH;
1597 
1598 	reqp = kmem_zalloc(sizeof (dca_request_t), KM_SLEEP);
1599 
1600 	reqp->dr_dca = dca;
1601 
1602 	/*
1603 	 * Setup the DMA region for the context and descriptors.
1604 	 */
1605 	rv = ddi_dma_alloc_handle(dca->dca_dip, &dca_dmaattr, DDI_DMA_SLEEP,
1606 	    NULL, &reqp->dr_ctx_dmah);
1607 	if (rv != DDI_SUCCESS) {
1608 		dca_error(dca, "failure allocating request DMA handle");
1609 		dca_destroyreq(reqp);
1610 		return (NULL);
1611 	}
1612 
1613 	/* for driver hardening, allocate in whole pages */
1614 	rv = ddi_dma_mem_alloc(reqp->dr_ctx_dmah,
1615 	    ROUNDUP(size, dca->dca_pagesize), &dca_devattr, DDI_DMA_CONSISTENT,
1616 	    DDI_DMA_SLEEP, NULL, &reqp->dr_ctx_kaddr, &size,
1617 	    &reqp->dr_ctx_acch);
1618 	if (rv != DDI_SUCCESS) {
1619 		dca_error(dca, "unable to alloc request DMA memory");
1620 		dca_destroyreq(reqp);
1621 		return (NULL);
1622 	}
1623 
1624 	rv = ddi_dma_addr_bind_handle(reqp->dr_ctx_dmah, NULL,
1625 	    reqp->dr_ctx_kaddr, size, DDI_DMA_CONSISTENT | DDI_DMA_WRITE,
1626 	    DDI_DMA_SLEEP, 0, &c, &nc);
1627 	if (rv != DDI_DMA_MAPPED) {
1628 		dca_error(dca, "failed binding request DMA handle");
1629 		dca_destroyreq(reqp);
1630 		return (NULL);
1631 	}
1632 	reqp->dr_ctx_paddr = c.dmac_address;
1633 
1634 	reqp->dr_dma_size = size;
1635 
1636 	/*
1637 	 * Set up the dma for our scratch/shared buffers.
1638 	 */
1639 	rv = ddi_dma_alloc_handle(dca->dca_dip, &dca_dmaattr,
1640 	    DDI_DMA_SLEEP, NULL, &reqp->dr_ibuf_dmah);
1641 	if (rv != DDI_SUCCESS) {
1642 		dca_error(dca, "failure allocating ibuf DMA handle");
1643 		dca_destroyreq(reqp);
1644 		return (NULL);
1645 	}
1646 	rv = ddi_dma_alloc_handle(dca->dca_dip, &dca_dmaattr,
1647 	    DDI_DMA_SLEEP, NULL, &reqp->dr_obuf_dmah);
1648 	if (rv != DDI_SUCCESS) {
1649 		dca_error(dca, "failure allocating obuf DMA handle");
1650 		dca_destroyreq(reqp);
1651 		return (NULL);
1652 	}
1653 
1654 	rv = ddi_dma_alloc_handle(dca->dca_dip, &dca_dmaattr,
1655 	    DDI_DMA_SLEEP, NULL, &reqp->dr_chain_in_dmah);
1656 	if (rv != DDI_SUCCESS) {
1657 		dca_error(dca, "failure allocating chain_in DMA handle");
1658 		dca_destroyreq(reqp);
1659 		return (NULL);
1660 	}
1661 
1662 	rv = ddi_dma_alloc_handle(dca->dca_dip, &dca_dmaattr,
1663 	    DDI_DMA_SLEEP, NULL, &reqp->dr_chain_out_dmah);
1664 	if (rv != DDI_SUCCESS) {
1665 		dca_error(dca, "failure allocating chain_out DMA handle");
1666 		dca_destroyreq(reqp);
1667 		return (NULL);
1668 	}
1669 
1670 	/*
1671 	 * for driver hardening, allocate in whole pages.
1672 	 */
1673 	size = ROUNDUP(MAXPACKET, dca->dca_pagesize);
1674 #if defined(i386) || defined(__i386)
1675 	/*
1676 	 * Use kmem_alloc instead of ddi_dma_mem_alloc here since the latter
1677 	 * may fail on x86 platform if a physically contigous memory chunk
1678 	 * cannot be found. From initial testing, we did not see performance
1679 	 * degration as seen on Sparc.
1680 	 */
1681 	if ((reqp->dr_ibuf_kaddr = kmem_alloc(size, KM_SLEEP)) == NULL) {
1682 		dca_error(dca, "unable to alloc request ibuf memory");
1683 		dca_destroyreq(reqp);
1684 		return (NULL);
1685 	}
1686 	if ((reqp->dr_obuf_kaddr = kmem_alloc(size, KM_SLEEP)) == NULL) {
1687 		dca_error(dca, "unable to alloc request obuf memory");
1688 		dca_destroyreq(reqp);
1689 		return (NULL);
1690 	}
1691 #else
1692 	/*
1693 	 * We could kmem_alloc for sparc too. However, it gives worse
1694 	 * performance when transfering more than one page data. For example,
1695 	 * using 4 threads and 12032 byte data and 3DES on 900MHZ sparc system,
1696 	 * kmem_alloc uses 80% CPU and ddi_dma_mem_alloc uses 50% CPU for
1697 	 * the same throughput.
1698 	 */
1699 	rv = ddi_dma_mem_alloc(reqp->dr_ibuf_dmah,
1700 	    size, &dca_bufattr,
1701 	    DDI_DMA_STREAMING, DDI_DMA_SLEEP, NULL, &reqp->dr_ibuf_kaddr,
1702 	    &size, &reqp->dr_ibuf_acch);
1703 	if (rv != DDI_SUCCESS) {
1704 		dca_error(dca, "unable to alloc request DMA memory");
1705 		dca_destroyreq(reqp);
1706 		return (NULL);
1707 	}
1708 
1709 	rv = ddi_dma_mem_alloc(reqp->dr_obuf_dmah,
1710 	    size, &dca_bufattr,
1711 	    DDI_DMA_STREAMING, DDI_DMA_SLEEP, NULL, &reqp->dr_obuf_kaddr,
1712 	    &size, &reqp->dr_obuf_acch);
1713 	if (rv != DDI_SUCCESS) {
1714 		dca_error(dca, "unable to alloc request DMA memory");
1715 		dca_destroyreq(reqp);
1716 		return (NULL);
1717 	}
1718 #endif
1719 
1720 	/* Skip the used portion in the context page */
1721 	reqp->dr_offset = CTX_MAXLENGTH;
1722 	if ((rv = dca_bindchains_one(reqp, size, reqp->dr_offset,
1723 	    reqp->dr_ibuf_kaddr, reqp->dr_ibuf_dmah,
1724 	    DDI_DMA_WRITE | DDI_DMA_STREAMING,
1725 	    &reqp->dr_ibuf_head, &n_chain)) != DDI_SUCCESS) {
1726 		(void) dca_destroyreq(reqp);
1727 		return (NULL);
1728 	}
1729 	reqp->dr_ibuf_paddr = reqp->dr_ibuf_head.dc_buffer_paddr;
1730 	/* Skip the space used by the input buffer */
1731 	reqp->dr_offset += DESC_SIZE * n_chain;
1732 
1733 	if ((rv = dca_bindchains_one(reqp, size, reqp->dr_offset,
1734 	    reqp->dr_obuf_kaddr, reqp->dr_obuf_dmah,
1735 	    DDI_DMA_READ | DDI_DMA_STREAMING,
1736 	    &reqp->dr_obuf_head, &n_chain)) != DDI_SUCCESS) {
1737 		(void) dca_destroyreq(reqp);
1738 		return (NULL);
1739 	}
1740 	reqp->dr_obuf_paddr = reqp->dr_obuf_head.dc_buffer_paddr;
1741 	/* Skip the space used by the output buffer */
1742 	reqp->dr_offset += DESC_SIZE * n_chain;
1743 
1744 	DBG(dca, DCHATTY, "CTX is 0x%p, phys 0x%x, len %d",
1745 	    reqp->dr_ctx_kaddr, reqp->dr_ctx_paddr, CTX_MAXLENGTH);
1746 	return (reqp);
1747 }
1748 
1749 void
1750 dca_destroyreq(dca_request_t *reqp)
1751 {
1752 #if defined(i386) || defined(__i386)
1753 	dca_t		*dca = reqp->dr_dca;
1754 	size_t		size = ROUNDUP(MAXPACKET, dca->dca_pagesize);
1755 #endif
1756 
1757 	/*
1758 	 * Clean up DMA for the context structure.
1759 	 */
1760 	if (reqp->dr_ctx_paddr) {
1761 		(void) ddi_dma_unbind_handle(reqp->dr_ctx_dmah);
1762 	}
1763 
1764 	if (reqp->dr_ctx_acch) {
1765 		ddi_dma_mem_free(&reqp->dr_ctx_acch);
1766 	}
1767 
1768 	if (reqp->dr_ctx_dmah) {
1769 		ddi_dma_free_handle(&reqp->dr_ctx_dmah);
1770 	}
1771 
1772 	/*
1773 	 * Clean up DMA for the scratch buffer.
1774 	 */
1775 #if defined(i386) || defined(__i386)
1776 	if (reqp->dr_ibuf_dmah) {
1777 		(void) ddi_dma_unbind_handle(reqp->dr_ibuf_dmah);
1778 		ddi_dma_free_handle(&reqp->dr_ibuf_dmah);
1779 	}
1780 	if (reqp->dr_obuf_dmah) {
1781 		(void) ddi_dma_unbind_handle(reqp->dr_obuf_dmah);
1782 		ddi_dma_free_handle(&reqp->dr_obuf_dmah);
1783 	}
1784 
1785 	kmem_free(reqp->dr_ibuf_kaddr, size);
1786 	kmem_free(reqp->dr_obuf_kaddr, size);
1787 #else
1788 	if (reqp->dr_ibuf_paddr) {
1789 		(void) ddi_dma_unbind_handle(reqp->dr_ibuf_dmah);
1790 	}
1791 	if (reqp->dr_obuf_paddr) {
1792 		(void) ddi_dma_unbind_handle(reqp->dr_obuf_dmah);
1793 	}
1794 
1795 	if (reqp->dr_ibuf_acch) {
1796 		ddi_dma_mem_free(&reqp->dr_ibuf_acch);
1797 	}
1798 	if (reqp->dr_obuf_acch) {
1799 		ddi_dma_mem_free(&reqp->dr_obuf_acch);
1800 	}
1801 
1802 	if (reqp->dr_ibuf_dmah) {
1803 		ddi_dma_free_handle(&reqp->dr_ibuf_dmah);
1804 	}
1805 	if (reqp->dr_obuf_dmah) {
1806 		ddi_dma_free_handle(&reqp->dr_obuf_dmah);
1807 	}
1808 #endif
1809 	/*
1810 	 * These two DMA handles should have been unbinded in
1811 	 * dca_unbindchains() function
1812 	 */
1813 	if (reqp->dr_chain_in_dmah) {
1814 		ddi_dma_free_handle(&reqp->dr_chain_in_dmah);
1815 	}
1816 	if (reqp->dr_chain_out_dmah) {
1817 		ddi_dma_free_handle(&reqp->dr_chain_out_dmah);
1818 	}
1819 
1820 	kmem_free(reqp, sizeof (dca_request_t));
1821 }
1822 
1823 dca_work_t *
1824 dca_getwork(dca_t *dca, int mcr)
1825 {
1826 	dca_worklist_t	*wlp = WORKLIST(dca, mcr);
1827 	dca_work_t	*workp;
1828 
1829 	mutex_enter(&wlp->dwl_freelock);
1830 	workp = (dca_work_t *)dca_dequeue(&wlp->dwl_freework);
1831 	mutex_exit(&wlp->dwl_freelock);
1832 	if (workp) {
1833 		int	nreqs;
1834 		bzero(workp->dw_mcr_kaddr, 8);
1835 
1836 		/* clear out old requests */
1837 		for (nreqs = 0; nreqs < MAXREQSPERMCR; nreqs++) {
1838 			workp->dw_reqs[nreqs] = NULL;
1839 		}
1840 	}
1841 	return (workp);
1842 }
1843 
1844 void
1845 dca_freework(dca_work_t *workp)
1846 {
1847 	mutex_enter(&workp->dw_wlp->dwl_freelock);
1848 	dca_enqueue(&workp->dw_wlp->dwl_freework, (dca_listnode_t *)workp);
1849 	mutex_exit(&workp->dw_wlp->dwl_freelock);
1850 }
1851 
1852 dca_request_t *
1853 dca_getreq(dca_t *dca, int mcr, int tryhard)
1854 {
1855 	dca_worklist_t	*wlp = WORKLIST(dca, mcr);
1856 	dca_request_t	*reqp;
1857 
1858 	mutex_enter(&wlp->dwl_freereqslock);
1859 	reqp = (dca_request_t *)dca_dequeue(&wlp->dwl_freereqs);
1860 	mutex_exit(&wlp->dwl_freereqslock);
1861 	if (reqp) {
1862 		reqp->dr_flags = 0;
1863 		reqp->dr_callback = NULL;
1864 	} else if (tryhard) {
1865 		/*
1866 		 * failed to get a free one, try an allocation, the hard way.
1867 		 * XXX: Kstat desired here.
1868 		 */
1869 		if ((reqp = dca_newreq(dca)) != NULL) {
1870 			reqp->dr_wlp = wlp;
1871 			reqp->dr_dca = dca;
1872 			reqp->dr_flags = 0;
1873 			reqp->dr_callback = NULL;
1874 		}
1875 	}
1876 	return (reqp);
1877 }
1878 
1879 void
1880 dca_freereq(dca_request_t *reqp)
1881 {
1882 	reqp->dr_kcf_req = NULL;
1883 	if (!(reqp->dr_flags & DR_NOCACHE)) {
1884 		mutex_enter(&reqp->dr_wlp->dwl_freereqslock);
1885 		dca_enqueue(&reqp->dr_wlp->dwl_freereqs,
1886 		    (dca_listnode_t *)reqp);
1887 		mutex_exit(&reqp->dr_wlp->dwl_freereqslock);
1888 	}
1889 }
1890 
1891 /*
1892  * Binds user buffers to DMA handles dynamically. On Sparc, a user buffer
1893  * is mapped to a single physicall address. On x86, a user buffer is mapped
1894  * to multiple physically addresses. These phsyical addresses are chained
1895  * using the method specified in Broadcom BCM5820 specification
1896  */
1897 int
1898 dca_bindchains(dca_request_t *reqp, size_t incnt, size_t outcnt)
1899 {
1900 	int			rv;
1901 	caddr_t			kaddr;
1902 	uint_t			flags;
1903 	int			n_chain = 0;
1904 
1905 	if (reqp->dr_flags & DR_INPLACE) {
1906 		flags = DDI_DMA_RDWR | DDI_DMA_CONSISTENT;
1907 	} else {
1908 		flags = DDI_DMA_WRITE | DDI_DMA_STREAMING;
1909 	}
1910 
1911 	/* first the input */
1912 	if (incnt) {
1913 		if ((kaddr = dca_bufdaddr(reqp->dr_in)) == NULL) {
1914 			DBG(NULL, DWARN, "unrecognised crypto data format");
1915 			return (DDI_FAILURE);
1916 		}
1917 		if ((rv = dca_bindchains_one(reqp, incnt, reqp->dr_offset,
1918 		    kaddr, reqp->dr_chain_in_dmah, flags,
1919 		    &reqp->dr_chain_in_head, &n_chain)) != DDI_SUCCESS) {
1920 			(void) dca_unbindchains(reqp);
1921 			return (rv);
1922 		}
1923 
1924 		/*
1925 		 * The offset and length are altered by the calling routine
1926 		 * reqp->dr_in->cd_offset += incnt;
1927 		 * reqp->dr_in->cd_length -= incnt;
1928 		 */
1929 		/* Save the first one in the chain for MCR */
1930 		reqp->dr_in_paddr = reqp->dr_chain_in_head.dc_buffer_paddr;
1931 		reqp->dr_in_next = reqp->dr_chain_in_head.dc_next_paddr;
1932 		reqp->dr_in_len = reqp->dr_chain_in_head.dc_buffer_length;
1933 	} else {
1934 		reqp->dr_in_paddr = NULL;
1935 		reqp->dr_in_next = 0;
1936 		reqp->dr_in_len = 0;
1937 	}
1938 
1939 	if (reqp->dr_flags & DR_INPLACE) {
1940 		reqp->dr_out_paddr = reqp->dr_in_paddr;
1941 		reqp->dr_out_len = reqp->dr_in_len;
1942 		reqp->dr_out_next = reqp->dr_in_next;
1943 		return (DDI_SUCCESS);
1944 	}
1945 
1946 	/* then the output */
1947 	if (outcnt) {
1948 		flags = DDI_DMA_READ | DDI_DMA_STREAMING;
1949 		if ((kaddr = dca_bufdaddr_out(reqp->dr_out)) == NULL) {
1950 			DBG(NULL, DWARN, "unrecognised crypto data format");
1951 			(void) dca_unbindchains(reqp);
1952 			return (DDI_FAILURE);
1953 		}
1954 		rv = dca_bindchains_one(reqp, outcnt, reqp->dr_offset +
1955 		    n_chain * DESC_SIZE, kaddr, reqp->dr_chain_out_dmah,
1956 		    flags, &reqp->dr_chain_out_head, &n_chain);
1957 		if (rv != DDI_SUCCESS) {
1958 			(void) dca_unbindchains(reqp);
1959 			return (DDI_FAILURE);
1960 		}
1961 
1962 		/* Save the first one in the chain for MCR */
1963 		reqp->dr_out_paddr = reqp->dr_chain_out_head.dc_buffer_paddr;
1964 		reqp->dr_out_next = reqp->dr_chain_out_head.dc_next_paddr;
1965 		reqp->dr_out_len = reqp->dr_chain_out_head.dc_buffer_length;
1966 	} else {
1967 		reqp->dr_out_paddr = NULL;
1968 		reqp->dr_out_next = 0;
1969 		reqp->dr_out_len = 0;
1970 	}
1971 
1972 	return (DDI_SUCCESS);
1973 }
1974 
1975 /*
1976  * Unbind the user buffers from the DMA handles.
1977  */
1978 int
1979 dca_unbindchains(dca_request_t *reqp)
1980 {
1981 	int rv = DDI_SUCCESS;
1982 	int rv1 = DDI_SUCCESS;
1983 
1984 	/* Clear the input chain */
1985 	if (reqp->dr_chain_in_head.dc_buffer_paddr != NULL) {
1986 		(void) ddi_dma_unbind_handle(reqp->dr_chain_in_dmah);
1987 		reqp->dr_chain_in_head.dc_buffer_paddr = 0;
1988 	}
1989 
1990 	if (reqp->dr_flags & DR_INPLACE) {
1991 		return (rv);
1992 	}
1993 
1994 	/* Clear the output chain */
1995 	if (reqp->dr_chain_out_head.dc_buffer_paddr != NULL) {
1996 		(void) ddi_dma_unbind_handle(reqp->dr_chain_out_dmah);
1997 		reqp->dr_chain_out_head.dc_buffer_paddr = 0;
1998 	}
1999 
2000 	return ((rv != DDI_SUCCESS)? rv : rv1);
2001 }
2002 
2003 /*
2004  * Build either input chain or output chain. It is single-item chain for Sparc,
2005  * and possible mutiple-item chain for x86.
2006  */
2007 static int
2008 dca_bindchains_one(dca_request_t *reqp, size_t cnt, int dr_offset,
2009     caddr_t kaddr, ddi_dma_handle_t handle, uint_t flags,
2010     dca_chain_t *head, int *n_chain)
2011 {
2012 	ddi_dma_cookie_t	c;
2013 	uint_t			nc;
2014 	int			rv;
2015 	caddr_t			chain_kaddr_pre;
2016 	caddr_t			chain_kaddr;
2017 	uint32_t		chain_paddr;
2018 	int 			i;
2019 
2020 	/* Advance past the context structure to the starting address */
2021 	chain_paddr = reqp->dr_ctx_paddr + dr_offset;
2022 	chain_kaddr = reqp->dr_ctx_kaddr + dr_offset;
2023 
2024 	/*
2025 	 * Bind the kernel address to the DMA handle. On x86, the actual
2026 	 * buffer is mapped into multiple physical addresses. On Sparc,
2027 	 * the actual buffer is mapped into a single address.
2028 	 */
2029 	rv = ddi_dma_addr_bind_handle(handle,
2030 	    NULL, kaddr, cnt, flags, DDI_DMA_DONTWAIT, NULL, &c, &nc);
2031 	if (rv != DDI_DMA_MAPPED) {
2032 		return (DDI_FAILURE);
2033 	}
2034 
2035 	(void) ddi_dma_sync(handle, 0, cnt, DDI_DMA_SYNC_FORDEV);
2036 	if ((rv = dca_check_dma_handle(reqp->dr_dca, handle,
2037 	    DCA_FM_ECLASS_NONE)) != DDI_SUCCESS) {
2038 		reqp->destroy = TRUE;
2039 		return (rv);
2040 	}
2041 
2042 	*n_chain = nc;
2043 
2044 	/* Setup the data buffer chain for DMA transfer */
2045 	chain_kaddr_pre = NULL;
2046 	head->dc_buffer_paddr = 0;
2047 	head->dc_next_paddr = 0;
2048 	head->dc_buffer_length = 0;
2049 	for (i = 0; i < nc; i++) {
2050 		/* PIO */
2051 		PUTDESC32(reqp, chain_kaddr, DESC_BUFADDR, c.dmac_address);
2052 		PUTDESC16(reqp, chain_kaddr, DESC_RSVD, 0);
2053 		PUTDESC16(reqp, chain_kaddr, DESC_LENGTH, c.dmac_size);
2054 
2055 		/* Remember the head of the chain */
2056 		if (head->dc_buffer_paddr == 0) {
2057 			head->dc_buffer_paddr = c.dmac_address;
2058 			head->dc_buffer_length = c.dmac_size;
2059 		}
2060 
2061 		/* Link to the previous one if one exists */
2062 		if (chain_kaddr_pre) {
2063 			PUTDESC32(reqp, chain_kaddr_pre, DESC_NEXT,
2064 			    chain_paddr);
2065 			if (head->dc_next_paddr == 0)
2066 				head->dc_next_paddr = chain_paddr;
2067 		}
2068 		chain_kaddr_pre = chain_kaddr;
2069 
2070 		/* Maintain pointers */
2071 		chain_paddr += DESC_SIZE;
2072 		chain_kaddr += DESC_SIZE;
2073 
2074 		/* Retrieve the next cookie if there is one */
2075 		if (i < nc-1)
2076 			ddi_dma_nextcookie(handle, &c);
2077 	}
2078 
2079 	/* Set the next pointer in the last entry to NULL */
2080 	PUTDESC32(reqp, chain_kaddr_pre, DESC_NEXT, 0);
2081 
2082 	return (DDI_SUCCESS);
2083 }
2084 
2085 /*
2086  * Schedule some work.
2087  */
2088 int
2089 dca_start(dca_t *dca, dca_request_t *reqp, int mcr, int dosched)
2090 {
2091 	dca_worklist_t	*wlp = WORKLIST(dca, mcr);
2092 
2093 	mutex_enter(&wlp->dwl_lock);
2094 
2095 	DBG(dca, DCHATTY, "req=%p, in=%p, out=%p, ctx=%p, ibuf=%p, obuf=%p",
2096 	    reqp, reqp->dr_in, reqp->dr_out, reqp->dr_ctx_kaddr,
2097 	    reqp->dr_ibuf_kaddr, reqp->dr_obuf_kaddr);
2098 	DBG(dca, DCHATTY, "ctx paddr = %x, ibuf paddr = %x, obuf paddr = %x",
2099 	    reqp->dr_ctx_paddr, reqp->dr_ibuf_paddr, reqp->dr_obuf_paddr);
2100 	/* sync out the entire context and descriptor chains */
2101 	(void) ddi_dma_sync(reqp->dr_ctx_dmah, 0, 0, DDI_DMA_SYNC_FORDEV);
2102 	if (dca_check_dma_handle(dca, reqp->dr_ctx_dmah,
2103 	    DCA_FM_ECLASS_NONE) != DDI_SUCCESS) {
2104 		reqp->destroy = TRUE;
2105 		mutex_exit(&wlp->dwl_lock);
2106 		return (CRYPTO_DEVICE_ERROR);
2107 	}
2108 
2109 	dca_enqueue(&wlp->dwl_waitq, (dca_listnode_t *)reqp);
2110 	wlp->dwl_count++;
2111 	wlp->dwl_lastsubmit = ddi_get_lbolt();
2112 	reqp->dr_wlp = wlp;
2113 
2114 	if ((wlp->dwl_count == wlp->dwl_hiwater) && (wlp->dwl_busy == 0)) {
2115 		/* we are fully loaded now, let kCF know */
2116 
2117 		wlp->dwl_flowctl++;
2118 		wlp->dwl_busy = 1;
2119 
2120 		crypto_prov_notify(wlp->dwl_prov, CRYPTO_PROVIDER_BUSY);
2121 	}
2122 
2123 	if (dosched) {
2124 #ifdef	SCHEDDELAY
2125 		/* possibly wait for more work to arrive */
2126 		if (wlp->dwl_count >= wlp->dwl_reqspermcr) {
2127 			dca_schedule(dca, mcr);
2128 		} else if (!wlp->dwl_schedtid) {
2129 			/* wait 1 msec for more work before doing it */
2130 			wlp->dwl_schedtid = timeout(dca_schedtimeout,
2131 			    (void *)wlp, drv_usectohz(MSEC));
2132 		}
2133 #else
2134 		dca_schedule(dca, mcr);
2135 #endif
2136 	}
2137 	mutex_exit(&wlp->dwl_lock);
2138 
2139 	return (CRYPTO_QUEUED);
2140 }
2141 
2142 void
2143 dca_schedule(dca_t *dca, int mcr)
2144 {
2145 	dca_worklist_t	*wlp = WORKLIST(dca, mcr);
2146 	int		csr;
2147 	int		full;
2148 	uint32_t	status;
2149 
2150 	ASSERT(mutex_owned(&wlp->dwl_lock));
2151 	/*
2152 	 * If the card is draining or has an outstanding failure,
2153 	 * don't schedule any more work on it right now
2154 	 */
2155 	if (wlp->dwl_drain || (dca->dca_flags & DCA_FAILED)) {
2156 		return;
2157 	}
2158 
2159 	if (mcr == MCR2) {
2160 		csr = CSR_MCR2;
2161 		full = DMASTAT_MCR2FULL;
2162 	} else {
2163 		csr = CSR_MCR1;
2164 		full = DMASTAT_MCR1FULL;
2165 	}
2166 
2167 	for (;;) {
2168 		dca_work_t	*workp;
2169 		uint32_t	offset;
2170 		int		nreqs;
2171 
2172 		status = GETCSR(dca, CSR_DMASTAT);
2173 		if (dca_check_acc_handle(dca, dca->dca_regs_handle,
2174 		    DCA_FM_ECLASS_NONE) != DDI_SUCCESS)
2175 			return;
2176 
2177 		if ((status & full) != 0)
2178 			break;
2179 
2180 #ifdef	SCHEDDELAY
2181 		/* if there isn't enough to do, don't bother now */
2182 		if ((wlp->dwl_count < wlp->dwl_reqspermcr) &&
2183 		    (ddi_get_lbolt() < (wlp->dwl_lastsubmit +
2184 		    drv_usectohz(MSEC)))) {
2185 			/* wait a bit longer... */
2186 			if (wlp->dwl_schedtid == 0) {
2187 				wlp->dwl_schedtid = timeout(dca_schedtimeout,
2188 				    (void *)wlp, drv_usectohz(MSEC));
2189 			}
2190 			return;
2191 		}
2192 #endif
2193 
2194 		/* grab a work structure */
2195 		workp = dca_getwork(dca, mcr);
2196 
2197 		if (workp == NULL) {
2198 			/*
2199 			 * There must be work ready to be reclaimed,
2200 			 * in this case, since the chip can only hold
2201 			 * less work outstanding than there are total.
2202 			 */
2203 			dca_reclaim(dca, mcr);
2204 			continue;
2205 		}
2206 
2207 		nreqs = 0;
2208 		offset = MCR_CTXADDR;
2209 
2210 		while (nreqs < wlp->dwl_reqspermcr) {
2211 			dca_request_t	*reqp;
2212 
2213 			reqp = (dca_request_t *)dca_dequeue(&wlp->dwl_waitq);
2214 			if (reqp == NULL) {
2215 				/* nothing left to process */
2216 				break;
2217 			}
2218 			/*
2219 			 * Update flow control.
2220 			 */
2221 			wlp->dwl_count--;
2222 			if ((wlp->dwl_count == wlp->dwl_lowater) &&
2223 			    (wlp->dwl_busy))  {
2224 				wlp->dwl_busy = 0;
2225 				crypto_prov_notify(wlp->dwl_prov,
2226 				    CRYPTO_PROVIDER_READY);
2227 			}
2228 
2229 			/*
2230 			 * Context address.
2231 			 */
2232 			PUTMCR32(workp, offset, reqp->dr_ctx_paddr);
2233 			offset += 4;
2234 
2235 			/*
2236 			 * Input chain.
2237 			 */
2238 			/* input buffer address */
2239 			PUTMCR32(workp, offset, reqp->dr_in_paddr);
2240 			offset += 4;
2241 			/* next input buffer entry */
2242 			PUTMCR32(workp, offset, reqp->dr_in_next);
2243 			offset += 4;
2244 			/* input buffer length */
2245 			PUTMCR16(workp, offset, reqp->dr_in_len);
2246 			offset += 2;
2247 			/* zero the reserved field */
2248 			PUTMCR16(workp, offset, 0);
2249 			offset += 2;
2250 
2251 			/*
2252 			 * Overall length.
2253 			 */
2254 			/* reserved field */
2255 			PUTMCR16(workp, offset, 0);
2256 			offset += 2;
2257 			/* total packet length */
2258 			PUTMCR16(workp, offset, reqp->dr_pkt_length);
2259 			offset += 2;
2260 
2261 			/*
2262 			 * Output chain.
2263 			 */
2264 			/* output buffer address */
2265 			PUTMCR32(workp, offset, reqp->dr_out_paddr);
2266 			offset += 4;
2267 			/* next output buffer entry */
2268 			PUTMCR32(workp, offset, reqp->dr_out_next);
2269 			offset += 4;
2270 			/* output buffer length */
2271 			PUTMCR16(workp, offset, reqp->dr_out_len);
2272 			offset += 2;
2273 			/* zero the reserved field */
2274 			PUTMCR16(workp, offset, 0);
2275 			offset += 2;
2276 
2277 			/*
2278 			 * Note submission.
2279 			 */
2280 			workp->dw_reqs[nreqs] = reqp;
2281 			nreqs++;
2282 		}
2283 
2284 		if (nreqs == 0) {
2285 			/* nothing in the queue! */
2286 			dca_freework(workp);
2287 			return;
2288 		}
2289 
2290 		wlp->dwl_submit++;
2291 
2292 		PUTMCR16(workp, MCR_FLAGS, 0);
2293 		PUTMCR16(workp, MCR_COUNT, nreqs);
2294 
2295 		DBG(dca, DCHATTY,
2296 		    "posting work (phys %x, virt 0x%p) (%d reqs) to MCR%d",
2297 		    workp->dw_mcr_paddr, workp->dw_mcr_kaddr,
2298 		    nreqs, mcr);
2299 
2300 		workp->dw_lbolt = ddi_get_lbolt();
2301 		/* Make sure MCR is synced out to device. */
2302 		(void) ddi_dma_sync(workp->dw_mcr_dmah, 0, 0,
2303 		    DDI_DMA_SYNC_FORDEV);
2304 		if (dca_check_dma_handle(dca, workp->dw_mcr_dmah,
2305 		    DCA_FM_ECLASS_NONE) != DDI_SUCCESS) {
2306 			dca_destroywork(workp);
2307 			return;
2308 		}
2309 
2310 		PUTCSR(dca, csr, workp->dw_mcr_paddr);
2311 		if (dca_check_acc_handle(dca, dca->dca_regs_handle,
2312 		    DCA_FM_ECLASS_NONE) != DDI_SUCCESS) {
2313 			dca_destroywork(workp);
2314 			return;
2315 		} else {
2316 			dca_enqueue(&wlp->dwl_runq, (dca_listnode_t *)workp);
2317 		}
2318 
2319 		DBG(dca, DCHATTY, "posted");
2320 	}
2321 }
2322 
2323 /*
2324  * Reclaim completed work, called in interrupt context.
2325  */
2326 void
2327 dca_reclaim(dca_t *dca, int mcr)
2328 {
2329 	dca_worklist_t	*wlp = WORKLIST(dca, mcr);
2330 	dca_work_t	*workp;
2331 	ushort_t	flags;
2332 	int		nreclaimed = 0;
2333 	int		i;
2334 
2335 	DBG(dca, DRECLAIM, "worklist = 0x%p (MCR%d)", wlp, mcr);
2336 	ASSERT(mutex_owned(&wlp->dwl_lock));
2337 	/*
2338 	 * For each MCR in the submitted (runq), we check to see if
2339 	 * it has been processed.  If so, then we note each individual
2340 	 * job in the MCR, and and do the completion processing for
2341 	 * each of such job.
2342 	 */
2343 	for (;;) {
2344 
2345 		workp = (dca_work_t *)dca_peekqueue(&wlp->dwl_runq);
2346 		if (workp == NULL) {
2347 			break;
2348 		}
2349 
2350 		/* only sync the MCR flags, since that's all we need */
2351 		(void) ddi_dma_sync(workp->dw_mcr_dmah, 0, 4,
2352 		    DDI_DMA_SYNC_FORKERNEL);
2353 		if (dca_check_dma_handle(dca, workp->dw_mcr_dmah,
2354 		    DCA_FM_ECLASS_NONE) != DDI_SUCCESS) {
2355 			dca_rmqueue((dca_listnode_t *)workp);
2356 			dca_destroywork(workp);
2357 			return;
2358 		}
2359 
2360 		flags = GETMCR16(workp, MCR_FLAGS);
2361 		if ((flags & MCRFLAG_FINISHED) == 0) {
2362 			/* chip is still working on it */
2363 			DBG(dca, DRECLAIM,
2364 			    "chip still working on it (MCR%d)", mcr);
2365 			break;
2366 		}
2367 
2368 		/* its really for us, so remove it from the queue */
2369 		dca_rmqueue((dca_listnode_t *)workp);
2370 
2371 		/* if we were draining, signal on the cv */
2372 		if (wlp->dwl_drain && QEMPTY(&wlp->dwl_runq)) {
2373 			cv_signal(&wlp->dwl_cv);
2374 		}
2375 
2376 		/* update statistics, done under the lock */
2377 		for (i = 0; i < wlp->dwl_reqspermcr; i++) {
2378 			dca_request_t *reqp = workp->dw_reqs[i];
2379 			if (reqp == NULL) {
2380 				continue;
2381 			}
2382 			if (reqp->dr_byte_stat >= 0) {
2383 				dca->dca_stats[reqp->dr_byte_stat] +=
2384 				    reqp->dr_pkt_length;
2385 			}
2386 			if (reqp->dr_job_stat >= 0) {
2387 				dca->dca_stats[reqp->dr_job_stat]++;
2388 			}
2389 		}
2390 		mutex_exit(&wlp->dwl_lock);
2391 
2392 		for (i = 0; i < wlp->dwl_reqspermcr; i++) {
2393 			dca_request_t *reqp = workp->dw_reqs[i];
2394 
2395 			if (reqp == NULL) {
2396 				continue;
2397 			}
2398 
2399 			/* Do the callback. */
2400 			workp->dw_reqs[i] = NULL;
2401 			dca_done(reqp, CRYPTO_SUCCESS);
2402 
2403 			nreclaimed++;
2404 		}
2405 
2406 		/* now we can release the work */
2407 		dca_freework(workp);
2408 
2409 		mutex_enter(&wlp->dwl_lock);
2410 	}
2411 	DBG(dca, DRECLAIM, "reclaimed %d cmds", nreclaimed);
2412 }
2413 
2414 int
2415 dca_length(crypto_data_t *cdata)
2416 {
2417 	return (cdata->cd_length);
2418 }
2419 
2420 /*
2421  * This is the callback function called from the interrupt when a kCF job
2422  * completes.  It does some driver-specific things, and then calls the
2423  * kCF-provided callback.  Finally, it cleans up the state for the work
2424  * request and drops the reference count to allow for DR.
2425  */
2426 void
2427 dca_done(dca_request_t *reqp, int err)
2428 {
2429 	uint64_t	ena = 0;
2430 
2431 	/* unbind any chains we were using */
2432 	if (dca_unbindchains(reqp) != DDI_SUCCESS) {
2433 		/* DMA failure */
2434 		ena = dca_ena(ena);
2435 		dca_failure(reqp->dr_dca, DDI_DATAPATH_FAULT,
2436 		    DCA_FM_ECLASS_NONE, ena, CRYPTO_DEVICE_ERROR,
2437 		    "fault on buffer DMA handle");
2438 		if (err == CRYPTO_SUCCESS) {
2439 			err = CRYPTO_DEVICE_ERROR;
2440 		}
2441 	}
2442 
2443 	if (reqp->dr_callback != NULL) {
2444 		reqp->dr_callback(reqp, err);
2445 	} else {
2446 		dca_freereq(reqp);
2447 	}
2448 }
2449 
2450 /*
2451  * Call this when a failure is detected.  It will reset the chip,
2452  * log a message, alert kCF, and mark jobs in the runq as failed.
2453  */
2454 /* ARGSUSED */
2455 void
2456 dca_failure(dca_t *dca, ddi_fault_location_t loc, dca_fma_eclass_t index,
2457     uint64_t ena, int errno, char *mess, ...)
2458 {
2459 	va_list	ap;
2460 	char	buf[256];
2461 	int	mcr;
2462 	char	*eclass;
2463 	int	have_mutex;
2464 
2465 	va_start(ap, mess);
2466 	(void) vsprintf(buf, mess, ap);
2467 	va_end(ap);
2468 
2469 	eclass = dca_fma_eclass_string(dca->dca_model, index);
2470 
2471 	if (DDI_FM_EREPORT_CAP(dca->fm_capabilities) &&
2472 	    index != DCA_FM_ECLASS_NONE) {
2473 		ddi_fm_ereport_post(dca->dca_dip, eclass, ena,
2474 		    DDI_NOSLEEP, FM_VERSION, DATA_TYPE_UINT8,
2475 		    FM_EREPORT_VERS0, NULL);
2476 
2477 		/* Report the impact of the failure to the DDI. */
2478 		ddi_fm_service_impact(dca->dca_dip, DDI_SERVICE_LOST);
2479 	} else {
2480 		/* Just log the error string to the message log */
2481 		dca_error(dca, buf);
2482 	}
2483 
2484 	/*
2485 	 * Indicate a failure (keeps schedule from running).
2486 	 */
2487 	dca->dca_flags |= DCA_FAILED;
2488 
2489 	/*
2490 	 * Reset the chip.  This should also have as a side effect, the
2491 	 * disabling of all interrupts from the device.
2492 	 */
2493 	(void) dca_reset(dca, 1);
2494 
2495 	/*
2496 	 * Report the failure to kCF.
2497 	 */
2498 	for (mcr = MCR1; mcr <= MCR2; mcr++) {
2499 		if (WORKLIST(dca, mcr)->dwl_prov) {
2500 			crypto_prov_notify(WORKLIST(dca, mcr)->dwl_prov,
2501 			    CRYPTO_PROVIDER_FAILED);
2502 		}
2503 	}
2504 
2505 	/*
2506 	 * Return jobs not sent to hardware back to kCF.
2507 	 */
2508 	dca_rejectjobs(dca);
2509 
2510 	/*
2511 	 * From this point on, no new work should be arriving, and the
2512 	 * chip should not be doing any active DMA.
2513 	 */
2514 
2515 	/*
2516 	 * Now find all the work submitted to the device and fail
2517 	 * them.
2518 	 */
2519 	for (mcr = MCR1; mcr <= MCR2; mcr++) {
2520 		dca_worklist_t	*wlp;
2521 		int		i;
2522 
2523 		wlp = WORKLIST(dca, mcr);
2524 
2525 		if (wlp == NULL || wlp->dwl_waitq.dl_prev == NULL) {
2526 			continue;
2527 		}
2528 		for (;;) {
2529 			dca_work_t	*workp;
2530 
2531 			have_mutex = mutex_tryenter(&wlp->dwl_lock);
2532 			workp = (dca_work_t *)dca_dequeue(&wlp->dwl_runq);
2533 			if (workp == NULL) {
2534 				if (have_mutex)
2535 					mutex_exit(&wlp->dwl_lock);
2536 				break;
2537 			}
2538 			mutex_exit(&wlp->dwl_lock);
2539 
2540 			/*
2541 			 * Free up requests
2542 			 */
2543 			for (i = 0; i < wlp->dwl_reqspermcr; i++) {
2544 				dca_request_t *reqp = workp->dw_reqs[i];
2545 				if (reqp) {
2546 					dca_done(reqp, errno);
2547 					workp->dw_reqs[i] = NULL;
2548 				}
2549 			}
2550 
2551 			mutex_enter(&wlp->dwl_lock);
2552 			/*
2553 			 * If waiting to drain, signal on the waiter.
2554 			 */
2555 			if (wlp->dwl_drain && QEMPTY(&wlp->dwl_runq)) {
2556 				cv_signal(&wlp->dwl_cv);
2557 			}
2558 
2559 			/*
2560 			 * Return the work and request structures to
2561 			 * the free pool.
2562 			 */
2563 			dca_freework(workp);
2564 			if (have_mutex)
2565 				mutex_exit(&wlp->dwl_lock);
2566 		}
2567 	}
2568 
2569 }
2570 
2571 #ifdef	SCHEDDELAY
2572 /*
2573  * Reschedule worklist as needed.
2574  */
2575 void
2576 dca_schedtimeout(void *arg)
2577 {
2578 	dca_worklist_t	*wlp = (dca_worklist_t *)arg;
2579 	mutex_enter(&wlp->dwl_lock);
2580 	wlp->dwl_schedtid = 0;
2581 	dca_schedule(wlp->dwl_dca, wlp->dwl_mcr);
2582 	mutex_exit(&wlp->dwl_lock);
2583 }
2584 #endif
2585 
2586 /*
2587  * Check for stalled jobs.
2588  */
2589 void
2590 dca_jobtimeout(void *arg)
2591 {
2592 	int		mcr;
2593 	dca_t		*dca = (dca_t *)arg;
2594 	int		hung = 0;
2595 
2596 	for (mcr = MCR1; mcr <= MCR2; mcr++) {
2597 		dca_worklist_t	*wlp = WORKLIST(dca, mcr);
2598 		dca_work_t	*workp;
2599 		clock_t		when;
2600 
2601 		mutex_enter(&wlp->dwl_lock);
2602 		when = ddi_get_lbolt();
2603 
2604 		workp = (dca_work_t *)dca_peekqueue(&wlp->dwl_runq);
2605 		if (workp == NULL) {
2606 			/* nothing sitting in the queue */
2607 			mutex_exit(&wlp->dwl_lock);
2608 			continue;
2609 		}
2610 
2611 		if ((when - workp->dw_lbolt) < drv_usectohz(STALETIME)) {
2612 			/* request has been queued for less than STALETIME */
2613 			mutex_exit(&wlp->dwl_lock);
2614 			continue;
2615 		}
2616 
2617 		/* job has been sitting around for over 1 second, badness */
2618 		DBG(dca, DWARN, "stale job (0x%p) found in MCR%d!", workp,
2619 		    mcr);
2620 
2621 		/* put it back in the queue, until we reset the chip */
2622 		hung++;
2623 		mutex_exit(&wlp->dwl_lock);
2624 	}
2625 
2626 	if (hung) {
2627 		dca_failure(dca, DDI_DEVICE_FAULT,
2628 		    DCA_FM_ECLASS_HW_TIMEOUT, dca_ena(0), CRYPTO_DEVICE_ERROR,
2629 		    "timeout processing job.)");
2630 	}
2631 
2632 	/* reschedule ourself */
2633 	mutex_enter(&dca->dca_intrlock);
2634 	if (dca->dca_jobtid == 0) {
2635 		/* timeout has been canceled, prior to DR */
2636 		mutex_exit(&dca->dca_intrlock);
2637 		return;
2638 	}
2639 
2640 	/* check again in 1 second */
2641 	dca->dca_jobtid = timeout(dca_jobtimeout, arg,
2642 	    drv_usectohz(SECOND));
2643 	mutex_exit(&dca->dca_intrlock);
2644 }
2645 
2646 /*
2647  * This returns all jobs back to kCF.  It assumes that processing
2648  * on the worklist has halted.
2649  */
2650 void
2651 dca_rejectjobs(dca_t *dca)
2652 {
2653 	int mcr;
2654 	int have_mutex;
2655 	for (mcr = MCR1; mcr <= MCR2; mcr++) {
2656 		dca_worklist_t	*wlp = WORKLIST(dca, mcr);
2657 		dca_request_t	*reqp;
2658 
2659 		if (wlp == NULL || wlp->dwl_waitq.dl_prev == NULL) {
2660 			continue;
2661 		}
2662 		have_mutex = mutex_tryenter(&wlp->dwl_lock);
2663 		for (;;) {
2664 			reqp = (dca_request_t *)dca_unqueue(&wlp->dwl_waitq);
2665 			if (reqp == NULL) {
2666 				break;
2667 			}
2668 			/* update flow control */
2669 			wlp->dwl_count--;
2670 			if ((wlp->dwl_count == wlp->dwl_lowater) &&
2671 			    (wlp->dwl_busy))  {
2672 				wlp->dwl_busy = 0;
2673 				crypto_prov_notify(wlp->dwl_prov,
2674 				    CRYPTO_PROVIDER_READY);
2675 			}
2676 			mutex_exit(&wlp->dwl_lock);
2677 
2678 			(void) dca_unbindchains(reqp);
2679 			reqp->dr_callback(reqp, EAGAIN);
2680 			mutex_enter(&wlp->dwl_lock);
2681 		}
2682 		if (have_mutex)
2683 			mutex_exit(&wlp->dwl_lock);
2684 	}
2685 }
2686 
2687 int
2688 dca_drain(dca_t *dca)
2689 {
2690 	int mcr;
2691 	for (mcr = MCR1; mcr <= MCR2; mcr++) {
2692 #ifdef	SCHEDDELAY
2693 		timeout_id_t	tid;
2694 #endif
2695 		dca_worklist_t *wlp = WORKLIST(dca, mcr);
2696 
2697 		mutex_enter(&wlp->dwl_lock);
2698 		wlp->dwl_drain = 1;
2699 
2700 		/* give it up to a second to drain from the chip */
2701 		if (!QEMPTY(&wlp->dwl_runq)) {
2702 			(void) cv_timedwait(&wlp->dwl_cv, &wlp->dwl_lock,
2703 			    ddi_get_time() + drv_usectohz(STALETIME));
2704 
2705 			if (!QEMPTY(&wlp->dwl_runq)) {
2706 				dca_error(dca, "unable to drain device");
2707 				mutex_exit(&wlp->dwl_lock);
2708 				dca_undrain(dca);
2709 				return (EBUSY);
2710 			}
2711 		}
2712 
2713 #ifdef	SCHEDDELAY
2714 		tid = wlp->dwl_schedtid;
2715 		mutex_exit(&wlp->dwl_lock);
2716 
2717 		/*
2718 		 * untimeout outside the lock -- this is safe because we
2719 		 * have set the drain flag, so dca_schedule() will not
2720 		 * reschedule another timeout
2721 		 */
2722 		if (tid) {
2723 			untimeout(tid);
2724 		}
2725 #else
2726 		mutex_exit(&wlp->dwl_lock);
2727 #endif
2728 	}
2729 	return (0);
2730 }
2731 
2732 void
2733 dca_undrain(dca_t *dca)
2734 {
2735 	int	mcr;
2736 
2737 	for (mcr = MCR1; mcr <= MCR2; mcr++) {
2738 		dca_worklist_t	*wlp = WORKLIST(dca, mcr);
2739 		mutex_enter(&wlp->dwl_lock);
2740 		wlp->dwl_drain = 0;
2741 		dca_schedule(dca, mcr);
2742 		mutex_exit(&wlp->dwl_lock);
2743 	}
2744 }
2745 
2746 /*
2747  * Duplicate the crypto_data_t structure, but point to the original
2748  * buffers.
2749  */
2750 int
2751 dca_dupcrypto(crypto_data_t *input, crypto_data_t *ninput)
2752 {
2753 	ninput->cd_format = input->cd_format;
2754 	ninput->cd_offset = input->cd_offset;
2755 	ninput->cd_length = input->cd_length;
2756 	ninput->cd_miscdata = input->cd_miscdata;
2757 
2758 	switch (input->cd_format) {
2759 	case CRYPTO_DATA_RAW:
2760 		ninput->cd_raw.iov_base = input->cd_raw.iov_base;
2761 		ninput->cd_raw.iov_len = input->cd_raw.iov_len;
2762 		break;
2763 
2764 	case CRYPTO_DATA_UIO:
2765 		ninput->cd_uio = input->cd_uio;
2766 		break;
2767 
2768 	case CRYPTO_DATA_MBLK:
2769 		ninput->cd_mp = input->cd_mp;
2770 		break;
2771 
2772 	default:
2773 		DBG(NULL, DWARN,
2774 		    "dca_dupcrypto: unrecognised crypto data format");
2775 		return (CRYPTO_FAILED);
2776 	}
2777 
2778 	return (CRYPTO_SUCCESS);
2779 }
2780 
2781 /*
2782  * Performs validation checks on the input and output data structures.
2783  */
2784 int
2785 dca_verifyio(crypto_data_t *input, crypto_data_t *output)
2786 {
2787 	int	rv = CRYPTO_SUCCESS;
2788 
2789 	switch (input->cd_format) {
2790 	case CRYPTO_DATA_RAW:
2791 		break;
2792 
2793 	case CRYPTO_DATA_UIO:
2794 		/* we support only kernel buffer */
2795 		if (input->cd_uio->uio_segflg != UIO_SYSSPACE) {
2796 			DBG(NULL, DWARN, "non kernel input uio buffer");
2797 			rv = CRYPTO_ARGUMENTS_BAD;
2798 		}
2799 		break;
2800 
2801 	case CRYPTO_DATA_MBLK:
2802 		break;
2803 
2804 	default:
2805 		DBG(NULL, DWARN, "unrecognised input crypto data format");
2806 		rv = CRYPTO_ARGUMENTS_BAD;
2807 	}
2808 
2809 	switch (output->cd_format) {
2810 	case CRYPTO_DATA_RAW:
2811 		break;
2812 
2813 	case CRYPTO_DATA_UIO:
2814 		/* we support only kernel buffer */
2815 		if (output->cd_uio->uio_segflg != UIO_SYSSPACE) {
2816 			DBG(NULL, DWARN, "non kernel output uio buffer");
2817 			rv = CRYPTO_ARGUMENTS_BAD;
2818 		}
2819 		break;
2820 
2821 	case CRYPTO_DATA_MBLK:
2822 		break;
2823 
2824 	default:
2825 		DBG(NULL, DWARN, "unrecognised output crypto data format");
2826 		rv = CRYPTO_ARGUMENTS_BAD;
2827 	}
2828 
2829 	return (rv);
2830 }
2831 
2832 /*
2833  * data: source crypto_data_t struct
2834  * off:	offset into the source before commencing copy
2835  * count: the amount of data to copy
2836  * dest: destination buffer
2837  */
2838 int
2839 dca_getbufbytes(crypto_data_t *data, size_t off, int count, uchar_t *dest)
2840 {
2841 	int rv = CRYPTO_SUCCESS;
2842 	uio_t *uiop;
2843 	uint_t vec_idx;
2844 	size_t cur_len;
2845 	mblk_t *mp;
2846 
2847 	if (count == 0) {
2848 		/* We don't want anything so we're done. */
2849 		return (rv);
2850 	}
2851 
2852 	/*
2853 	 * Sanity check that we haven't specified a length greater than the
2854 	 * offset adjusted size of the buffer.
2855 	 */
2856 	if (count > (data->cd_length - off)) {
2857 		return (CRYPTO_DATA_LEN_RANGE);
2858 	}
2859 
2860 	/* Add the internal crypto_data offset to the requested offset. */
2861 	off += data->cd_offset;
2862 
2863 	switch (data->cd_format) {
2864 	case CRYPTO_DATA_RAW:
2865 		bcopy(data->cd_raw.iov_base + off, dest, count);
2866 		break;
2867 
2868 	case CRYPTO_DATA_UIO:
2869 		/*
2870 		 * Jump to the first iovec containing data to be
2871 		 * processed.
2872 		 */
2873 		uiop = data->cd_uio;
2874 		for (vec_idx = 0; vec_idx < uiop->uio_iovcnt &&
2875 		    off >= uiop->uio_iov[vec_idx].iov_len;
2876 		    off -= uiop->uio_iov[vec_idx++].iov_len)
2877 			;
2878 		if (vec_idx == uiop->uio_iovcnt) {
2879 			/*
2880 			 * The caller specified an offset that is larger than
2881 			 * the total size of the buffers it provided.
2882 			 */
2883 			return (CRYPTO_DATA_LEN_RANGE);
2884 		}
2885 
2886 		/*
2887 		 * Now process the iovecs.
2888 		 */
2889 		while (vec_idx < uiop->uio_iovcnt && count > 0) {
2890 			cur_len = min(uiop->uio_iov[vec_idx].iov_len -
2891 			    off, count);
2892 			bcopy(uiop->uio_iov[vec_idx].iov_base + off, dest,
2893 			    cur_len);
2894 			count -= cur_len;
2895 			dest += cur_len;
2896 			vec_idx++;
2897 			off = 0;
2898 		}
2899 
2900 		if (vec_idx == uiop->uio_iovcnt && count > 0) {
2901 			/*
2902 			 * The end of the specified iovec's was reached but
2903 			 * the length requested could not be processed
2904 			 * (requested to digest more data than it provided).
2905 			 */
2906 			return (CRYPTO_DATA_LEN_RANGE);
2907 		}
2908 		break;
2909 
2910 	case CRYPTO_DATA_MBLK:
2911 		/*
2912 		 * Jump to the first mblk_t containing data to be processed.
2913 		 */
2914 		for (mp = data->cd_mp; mp != NULL && off >= MBLKL(mp);
2915 		    off -= MBLKL(mp), mp = mp->b_cont)
2916 			;
2917 		if (mp == NULL) {
2918 			/*
2919 			 * The caller specified an offset that is larger than
2920 			 * the total size of the buffers it provided.
2921 			 */
2922 			return (CRYPTO_DATA_LEN_RANGE);
2923 		}
2924 
2925 		/*
2926 		 * Now do the processing on the mblk chain.
2927 		 */
2928 		while (mp != NULL && count > 0) {
2929 			cur_len = min(MBLKL(mp) - off, count);
2930 			bcopy((char *)(mp->b_rptr + off), dest, cur_len);
2931 			count -= cur_len;
2932 			dest += cur_len;
2933 			mp = mp->b_cont;
2934 			off = 0;
2935 		}
2936 
2937 		if (mp == NULL && count > 0) {
2938 			/*
2939 			 * The end of the mblk was reached but the length
2940 			 * requested could not be processed, (requested to
2941 			 * digest more data than it provided).
2942 			 */
2943 			return (CRYPTO_DATA_LEN_RANGE);
2944 		}
2945 		break;
2946 
2947 	default:
2948 		DBG(NULL, DWARN, "unrecognised crypto data format");
2949 		rv = CRYPTO_ARGUMENTS_BAD;
2950 	}
2951 	return (rv);
2952 }
2953 
2954 
2955 /*
2956  * Performs the input, output or hard scatter/gather checks on the specified
2957  * crypto_data_t struct. Returns true if the data is scatter/gather in nature
2958  * ie fails the test.
2959  */
2960 int
2961 dca_sgcheck(dca_t *dca, crypto_data_t *data, dca_sg_param_t val)
2962 {
2963 	uio_t *uiop;
2964 	mblk_t *mp;
2965 	int rv = FALSE;
2966 
2967 	switch (val) {
2968 	case DCA_SG_CONTIG:
2969 		/*
2970 		 * Check for a contiguous data buffer.
2971 		 */
2972 		switch (data->cd_format) {
2973 		case CRYPTO_DATA_RAW:
2974 			/* Contiguous in nature */
2975 			break;
2976 
2977 		case CRYPTO_DATA_UIO:
2978 			if (data->cd_uio->uio_iovcnt > 1)
2979 				rv = TRUE;
2980 			break;
2981 
2982 		case CRYPTO_DATA_MBLK:
2983 			mp = data->cd_mp;
2984 			if (mp->b_cont != NULL)
2985 				rv = TRUE;
2986 			break;
2987 
2988 		default:
2989 			DBG(NULL, DWARN, "unrecognised crypto data format");
2990 		}
2991 		break;
2992 
2993 	case DCA_SG_WALIGN:
2994 		/*
2995 		 * Check for a contiguous data buffer that is 32-bit word
2996 		 * aligned and is of word multiples in size.
2997 		 */
2998 		switch (data->cd_format) {
2999 		case CRYPTO_DATA_RAW:
3000 			if ((data->cd_raw.iov_len % sizeof (uint32_t)) ||
3001 			    ((uintptr_t)data->cd_raw.iov_base %
3002 			    sizeof (uint32_t))) {
3003 				rv = TRUE;
3004 			}
3005 			break;
3006 
3007 		case CRYPTO_DATA_UIO:
3008 			uiop = data->cd_uio;
3009 			if (uiop->uio_iovcnt > 1) {
3010 				return (TRUE);
3011 			}
3012 			/* So there is only one iovec */
3013 			if ((uiop->uio_iov[0].iov_len % sizeof (uint32_t)) ||
3014 			    ((uintptr_t)uiop->uio_iov[0].iov_base %
3015 			    sizeof (uint32_t))) {
3016 				rv = TRUE;
3017 			}
3018 			break;
3019 
3020 		case CRYPTO_DATA_MBLK:
3021 			mp = data->cd_mp;
3022 			if (mp->b_cont != NULL) {
3023 				return (TRUE);
3024 			}
3025 			/* So there is only one mblk in the chain */
3026 			if ((MBLKL(mp) % sizeof (uint32_t)) ||
3027 			    ((uintptr_t)mp->b_rptr % sizeof (uint32_t))) {
3028 				rv = TRUE;
3029 			}
3030 			break;
3031 
3032 		default:
3033 			DBG(NULL, DWARN, "unrecognised crypto data format");
3034 		}
3035 		break;
3036 
3037 	case DCA_SG_PALIGN:
3038 		/*
3039 		 * Check that the data buffer is page aligned and is of
3040 		 * page multiples in size.
3041 		 */
3042 		switch (data->cd_format) {
3043 		case CRYPTO_DATA_RAW:
3044 			if ((data->cd_length % dca->dca_pagesize) ||
3045 			    ((uintptr_t)data->cd_raw.iov_base %
3046 			    dca->dca_pagesize)) {
3047 				rv = TRUE;
3048 			}
3049 			break;
3050 
3051 		case CRYPTO_DATA_UIO:
3052 			uiop = data->cd_uio;
3053 			if ((uiop->uio_iov[0].iov_len % dca->dca_pagesize) ||
3054 			    ((uintptr_t)uiop->uio_iov[0].iov_base %
3055 			    dca->dca_pagesize)) {
3056 				rv = TRUE;
3057 			}
3058 			break;
3059 
3060 		case CRYPTO_DATA_MBLK:
3061 			mp = data->cd_mp;
3062 			if ((MBLKL(mp) % dca->dca_pagesize) ||
3063 			    ((uintptr_t)mp->b_rptr % dca->dca_pagesize)) {
3064 				rv = TRUE;
3065 			}
3066 			break;
3067 
3068 		default:
3069 			DBG(NULL, DWARN, "unrecognised crypto data format");
3070 		}
3071 		break;
3072 
3073 	default:
3074 		DBG(NULL, DWARN, "unrecognised scatter/gather param type");
3075 	}
3076 
3077 	return (rv);
3078 }
3079 
3080 /*
3081  * Increments the cd_offset and decrements the cd_length as the data is
3082  * gathered from the crypto_data_t struct.
3083  * The data is reverse-copied into the dest buffer if the flag is true.
3084  */
3085 int
3086 dca_gather(crypto_data_t *in, char *dest, int count, int reverse)
3087 {
3088 	int	rv = CRYPTO_SUCCESS;
3089 	uint_t	vec_idx;
3090 	uio_t	*uiop;
3091 	off_t	off = in->cd_offset;
3092 	size_t	cur_len;
3093 	mblk_t	*mp;
3094 
3095 	switch (in->cd_format) {
3096 	case CRYPTO_DATA_RAW:
3097 		if (count > in->cd_length) {
3098 			/*
3099 			 * The caller specified a length greater than the
3100 			 * size of the buffer.
3101 			 */
3102 			return (CRYPTO_DATA_LEN_RANGE);
3103 		}
3104 		if (reverse)
3105 			dca_reverse(in->cd_raw.iov_base + off, dest, count,
3106 			    count);
3107 		else
3108 			bcopy(in->cd_raw.iov_base + in->cd_offset, dest, count);
3109 		in->cd_offset += count;
3110 		in->cd_length -= count;
3111 		break;
3112 
3113 	case CRYPTO_DATA_UIO:
3114 		/*
3115 		 * Jump to the first iovec containing data to be processed.
3116 		 */
3117 		uiop = in->cd_uio;
3118 		for (vec_idx = 0; vec_idx < uiop->uio_iovcnt &&
3119 		    off >= uiop->uio_iov[vec_idx].iov_len;
3120 		    off -= uiop->uio_iov[vec_idx++].iov_len)
3121 			;
3122 		if (vec_idx == uiop->uio_iovcnt) {
3123 			/*
3124 			 * The caller specified an offset that is larger than
3125 			 * the total size of the buffers it provided.
3126 			 */
3127 			return (CRYPTO_DATA_LEN_RANGE);
3128 		}
3129 
3130 		/*
3131 		 * Now process the iovecs.
3132 		 */
3133 		while (vec_idx < uiop->uio_iovcnt && count > 0) {
3134 			cur_len = min(uiop->uio_iov[vec_idx].iov_len -
3135 			    off, count);
3136 			count -= cur_len;
3137 			if (reverse) {
3138 				/* Fill the dest buffer from the end */
3139 				dca_reverse(uiop->uio_iov[vec_idx].iov_base +
3140 				    off, dest+count, cur_len, cur_len);
3141 			} else {
3142 				bcopy(uiop->uio_iov[vec_idx].iov_base + off,
3143 				    dest, cur_len);
3144 				dest += cur_len;
3145 			}
3146 			in->cd_offset += cur_len;
3147 			in->cd_length -= cur_len;
3148 			vec_idx++;
3149 			off = 0;
3150 		}
3151 
3152 		if (vec_idx == uiop->uio_iovcnt && count > 0) {
3153 			/*
3154 			 * The end of the specified iovec's was reached but
3155 			 * the length requested could not be processed
3156 			 * (requested to digest more data than it provided).
3157 			 */
3158 			return (CRYPTO_DATA_LEN_RANGE);
3159 		}
3160 		break;
3161 
3162 	case CRYPTO_DATA_MBLK:
3163 		/*
3164 		 * Jump to the first mblk_t containing data to be processed.
3165 		 */
3166 		for (mp = in->cd_mp; mp != NULL && off >= MBLKL(mp);
3167 		    off -= MBLKL(mp), mp = mp->b_cont)
3168 			;
3169 		if (mp == NULL) {
3170 			/*
3171 			 * The caller specified an offset that is larger than
3172 			 * the total size of the buffers it provided.
3173 			 */
3174 			return (CRYPTO_DATA_LEN_RANGE);
3175 		}
3176 
3177 		/*
3178 		 * Now do the processing on the mblk chain.
3179 		 */
3180 		while (mp != NULL && count > 0) {
3181 			cur_len = min(MBLKL(mp) - off, count);
3182 			count -= cur_len;
3183 			if (reverse) {
3184 				/* Fill the dest buffer from the end */
3185 				dca_reverse((char *)(mp->b_rptr + off),
3186 				    dest+count, cur_len, cur_len);
3187 			} else {
3188 				bcopy((char *)(mp->b_rptr + off), dest,
3189 				    cur_len);
3190 				dest += cur_len;
3191 			}
3192 			in->cd_offset += cur_len;
3193 			in->cd_length -= cur_len;
3194 			mp = mp->b_cont;
3195 			off = 0;
3196 		}
3197 
3198 		if (mp == NULL && count > 0) {
3199 			/*
3200 			 * The end of the mblk was reached but the length
3201 			 * requested could not be processed, (requested to
3202 			 * digest more data than it provided).
3203 			 */
3204 			return (CRYPTO_DATA_LEN_RANGE);
3205 		}
3206 		break;
3207 
3208 	default:
3209 		DBG(NULL, DWARN, "dca_gather: unrecognised crypto data format");
3210 		rv = CRYPTO_ARGUMENTS_BAD;
3211 	}
3212 	return (rv);
3213 }
3214 
3215 /*
3216  * Increments the cd_offset and decrements the cd_length as the data is
3217  * gathered from the crypto_data_t struct.
3218  */
3219 int
3220 dca_resid_gather(crypto_data_t *in, char *resid, int *residlen, char *dest,
3221     int count)
3222 {
3223 	int	rv = CRYPTO_SUCCESS;
3224 	caddr_t	baddr;
3225 	uint_t	vec_idx;
3226 	uio_t	*uiop;
3227 	off_t	off = in->cd_offset;
3228 	size_t	cur_len;
3229 	mblk_t	*mp;
3230 
3231 	/* Process the residual first */
3232 	if (*residlen > 0) {
3233 		uint_t	num = min(count, *residlen);
3234 		bcopy(resid, dest, num);
3235 		*residlen -= num;
3236 		if (*residlen > 0) {
3237 			/*
3238 			 * Requested amount 'count' is less than what's in
3239 			 * the residual, so shuffle any remaining resid to
3240 			 * the front.
3241 			 */
3242 			baddr = resid + num;
3243 			bcopy(baddr, resid, *residlen);
3244 		}
3245 		dest += num;
3246 		count -= num;
3247 	}
3248 
3249 	/* Now process what's in the crypto_data_t structs */
3250 	switch (in->cd_format) {
3251 	case CRYPTO_DATA_RAW:
3252 		if (count > in->cd_length) {
3253 			/*
3254 			 * The caller specified a length greater than the
3255 			 * size of the buffer.
3256 			 */
3257 			return (CRYPTO_DATA_LEN_RANGE);
3258 		}
3259 		bcopy(in->cd_raw.iov_base + in->cd_offset, dest, count);
3260 		in->cd_offset += count;
3261 		in->cd_length -= count;
3262 		break;
3263 
3264 	case CRYPTO_DATA_UIO:
3265 		/*
3266 		 * Jump to the first iovec containing data to be processed.
3267 		 */
3268 		uiop = in->cd_uio;
3269 		for (vec_idx = 0; vec_idx < uiop->uio_iovcnt &&
3270 		    off >= uiop->uio_iov[vec_idx].iov_len;
3271 		    off -= uiop->uio_iov[vec_idx++].iov_len)
3272 			;
3273 		if (vec_idx == uiop->uio_iovcnt) {
3274 			/*
3275 			 * The caller specified an offset that is larger than
3276 			 * the total size of the buffers it provided.
3277 			 */
3278 			return (CRYPTO_DATA_LEN_RANGE);
3279 		}
3280 
3281 		/*
3282 		 * Now process the iovecs.
3283 		 */
3284 		while (vec_idx < uiop->uio_iovcnt && count > 0) {
3285 			cur_len = min(uiop->uio_iov[vec_idx].iov_len -
3286 			    off, count);
3287 			bcopy(uiop->uio_iov[vec_idx].iov_base + off, dest,
3288 			    cur_len);
3289 			count -= cur_len;
3290 			dest += cur_len;
3291 			in->cd_offset += cur_len;
3292 			in->cd_length -= cur_len;
3293 			vec_idx++;
3294 			off = 0;
3295 		}
3296 
3297 		if (vec_idx == uiop->uio_iovcnt && count > 0) {
3298 			/*
3299 			 * The end of the specified iovec's was reached but
3300 			 * the length requested could not be processed
3301 			 * (requested to digest more data than it provided).
3302 			 */
3303 			return (CRYPTO_DATA_LEN_RANGE);
3304 		}
3305 		break;
3306 
3307 	case CRYPTO_DATA_MBLK:
3308 		/*
3309 		 * Jump to the first mblk_t containing data to be processed.
3310 		 */
3311 		for (mp = in->cd_mp; mp != NULL && off >= MBLKL(mp);
3312 		    off -= MBLKL(mp), mp = mp->b_cont)
3313 			;
3314 		if (mp == NULL) {
3315 			/*
3316 			 * The caller specified an offset that is larger than
3317 			 * the total size of the buffers it provided.
3318 			 */
3319 			return (CRYPTO_DATA_LEN_RANGE);
3320 		}
3321 
3322 		/*
3323 		 * Now do the processing on the mblk chain.
3324 		 */
3325 		while (mp != NULL && count > 0) {
3326 			cur_len = min(MBLKL(mp) - off, count);
3327 			bcopy((char *)(mp->b_rptr + off), dest, cur_len);
3328 			count -= cur_len;
3329 			dest += cur_len;
3330 			in->cd_offset += cur_len;
3331 			in->cd_length -= cur_len;
3332 			mp = mp->b_cont;
3333 			off = 0;
3334 		}
3335 
3336 		if (mp == NULL && count > 0) {
3337 			/*
3338 			 * The end of the mblk was reached but the length
3339 			 * requested could not be processed, (requested to
3340 			 * digest more data than it provided).
3341 			 */
3342 			return (CRYPTO_DATA_LEN_RANGE);
3343 		}
3344 		break;
3345 
3346 	default:
3347 		DBG(NULL, DWARN,
3348 		    "dca_resid_gather: unrecognised crypto data format");
3349 		rv = CRYPTO_ARGUMENTS_BAD;
3350 	}
3351 	return (rv);
3352 }
3353 
3354 /*
3355  * Appends the data to the crypto_data_t struct increasing cd_length.
3356  * cd_offset is left unchanged.
3357  * Data is reverse-copied if the flag is TRUE.
3358  */
3359 int
3360 dca_scatter(const char *src, crypto_data_t *out, int count, int reverse)
3361 {
3362 	int	rv = CRYPTO_SUCCESS;
3363 	off_t	offset = out->cd_offset + out->cd_length;
3364 	uint_t	vec_idx;
3365 	uio_t	*uiop;
3366 	size_t	cur_len;
3367 	mblk_t	*mp;
3368 
3369 	switch (out->cd_format) {
3370 	case CRYPTO_DATA_RAW:
3371 		if (out->cd_raw.iov_len - offset < count) {
3372 			/* Trying to write out more than space available. */
3373 			return (CRYPTO_DATA_LEN_RANGE);
3374 		}
3375 		if (reverse)
3376 			dca_reverse((void*) src, out->cd_raw.iov_base + offset,
3377 			    count, count);
3378 		else
3379 			bcopy(src, out->cd_raw.iov_base + offset, count);
3380 		out->cd_length += count;
3381 		break;
3382 
3383 	case CRYPTO_DATA_UIO:
3384 		/*
3385 		 * Jump to the first iovec that can be written to.
3386 		 */
3387 		uiop = out->cd_uio;
3388 		for (vec_idx = 0; vec_idx < uiop->uio_iovcnt &&
3389 		    offset >= uiop->uio_iov[vec_idx].iov_len;
3390 		    offset -= uiop->uio_iov[vec_idx++].iov_len)
3391 			;
3392 		if (vec_idx == uiop->uio_iovcnt) {
3393 			/*
3394 			 * The caller specified an offset that is larger than
3395 			 * the total size of the buffers it provided.
3396 			 */
3397 			return (CRYPTO_DATA_LEN_RANGE);
3398 		}
3399 
3400 		/*
3401 		 * Now process the iovecs.
3402 		 */
3403 		while (vec_idx < uiop->uio_iovcnt && count > 0) {
3404 			cur_len = min(uiop->uio_iov[vec_idx].iov_len -
3405 			    offset, count);
3406 			count -= cur_len;
3407 			if (reverse) {
3408 				dca_reverse((void*) (src+count),
3409 				    uiop->uio_iov[vec_idx].iov_base +
3410 				    offset, cur_len, cur_len);
3411 			} else {
3412 				bcopy(src, uiop->uio_iov[vec_idx].iov_base +
3413 				    offset, cur_len);
3414 				src += cur_len;
3415 			}
3416 			out->cd_length += cur_len;
3417 			vec_idx++;
3418 			offset = 0;
3419 		}
3420 
3421 		if (vec_idx == uiop->uio_iovcnt && count > 0) {
3422 			/*
3423 			 * The end of the specified iovec's was reached but
3424 			 * the length requested could not be processed
3425 			 * (requested to write more data than space provided).
3426 			 */
3427 			return (CRYPTO_DATA_LEN_RANGE);
3428 		}
3429 		break;
3430 
3431 	case CRYPTO_DATA_MBLK:
3432 		/*
3433 		 * Jump to the first mblk_t that can be written to.
3434 		 */
3435 		for (mp = out->cd_mp; mp != NULL && offset >= MBLKL(mp);
3436 		    offset -= MBLKL(mp), mp = mp->b_cont)
3437 			;
3438 		if (mp == NULL) {
3439 			/*
3440 			 * The caller specified an offset that is larger than
3441 			 * the total size of the buffers it provided.
3442 			 */
3443 			return (CRYPTO_DATA_LEN_RANGE);
3444 		}
3445 
3446 		/*
3447 		 * Now do the processing on the mblk chain.
3448 		 */
3449 		while (mp != NULL && count > 0) {
3450 			cur_len = min(MBLKL(mp) - offset, count);
3451 			count -= cur_len;
3452 			if (reverse) {
3453 				dca_reverse((void*) (src+count),
3454 				    (char *)(mp->b_rptr + offset), cur_len,
3455 				    cur_len);
3456 			} else {
3457 				bcopy(src, (char *)(mp->b_rptr + offset),
3458 				    cur_len);
3459 				src += cur_len;
3460 			}
3461 			out->cd_length += cur_len;
3462 			mp = mp->b_cont;
3463 			offset = 0;
3464 		}
3465 
3466 		if (mp == NULL && count > 0) {
3467 			/*
3468 			 * The end of the mblk was reached but the length
3469 			 * requested could not be processed, (requested to
3470 			 * digest more data than it provided).
3471 			 */
3472 			return (CRYPTO_DATA_LEN_RANGE);
3473 		}
3474 		break;
3475 
3476 	default:
3477 		DBG(NULL, DWARN, "unrecognised crypto data format");
3478 		rv = CRYPTO_ARGUMENTS_BAD;
3479 	}
3480 	return (rv);
3481 }
3482 
3483 /*
3484  * Compare two byte arrays in reverse order.
3485  * Return 0 if they are identical, 1 otherwise.
3486  */
3487 int
3488 dca_bcmp_reverse(const void *s1, const void *s2, size_t n)
3489 {
3490 	int i;
3491 	caddr_t src, dst;
3492 
3493 	if (!n)
3494 		return (0);
3495 
3496 	src = ((caddr_t)s1) + n - 1;
3497 	dst = (caddr_t)s2;
3498 	for (i = 0; i < n; i++) {
3499 		if (*src != *dst)
3500 			return (1);
3501 		src--;
3502 		dst++;
3503 	}
3504 
3505 	return (0);
3506 }
3507 
3508 
3509 /*
3510  * This calculates the size of a bignum in bits, specifically not counting
3511  * leading zero bits.  This size calculation must be done *before* any
3512  * endian reversal takes place (i.e. the numbers are in absolute big-endian
3513  * order.)
3514  */
3515 int
3516 dca_bitlen(unsigned char *bignum, int bytelen)
3517 {
3518 	unsigned char	msbyte;
3519 	int		i, j;
3520 
3521 	for (i = 0; i < bytelen - 1; i++) {
3522 		if (bignum[i] != 0) {
3523 			break;
3524 		}
3525 	}
3526 	msbyte = bignum[i];
3527 	for (j = 8; j > 1; j--) {
3528 		if (msbyte & 0x80) {
3529 			break;
3530 		}
3531 		msbyte <<= 1;
3532 	}
3533 	return ((8 * (bytelen - i - 1)) + j);
3534 }
3535 
3536 /*
3537  * This compares to bignums (in big-endian order).  It ignores leading
3538  * null bytes.  The result semantics follow bcmp, mempcmp, strcmp, etc.
3539  */
3540 int
3541 dca_numcmp(caddr_t n1, int n1len, caddr_t n2, int n2len)
3542 {
3543 	while ((n1len > 1) && (*n1 == 0)) {
3544 		n1len--;
3545 		n1++;
3546 	}
3547 	while ((n2len > 1) && (*n2 == 0)) {
3548 		n2len--;
3549 		n2++;
3550 	}
3551 	if (n1len != n2len) {
3552 		return (n1len - n2len);
3553 	}
3554 	while ((n1len > 1) && (*n1 == *n2)) {
3555 		n1++;
3556 		n2++;
3557 		n1len--;
3558 	}
3559 	return ((int)(*(uchar_t *)n1) - (int)(*(uchar_t *)n2));
3560 }
3561 
3562 /*
3563  * Return array of key attributes.
3564  */
3565 crypto_object_attribute_t *
3566 dca_get_key_attr(crypto_key_t *key)
3567 {
3568 	if ((key->ck_format != CRYPTO_KEY_ATTR_LIST) ||
3569 	    (key->ck_count == 0)) {
3570 		return (NULL);
3571 	}
3572 
3573 	return (key->ck_attrs);
3574 }
3575 
3576 /*
3577  * If attribute type exists valp points to it's 32-bit value.
3578  */
3579 int
3580 dca_attr_lookup_uint32(crypto_object_attribute_t *attrp, uint_t atnum,
3581     uint64_t atype, uint32_t *valp)
3582 {
3583 	crypto_object_attribute_t	*bap;
3584 
3585 	bap = dca_find_attribute(attrp, atnum, atype);
3586 	if (bap == NULL) {
3587 		return (CRYPTO_ATTRIBUTE_TYPE_INVALID);
3588 	}
3589 
3590 	*valp = *bap->oa_value;
3591 
3592 	return (CRYPTO_SUCCESS);
3593 }
3594 
3595 /*
3596  * If attribute type exists data contains the start address of the value,
3597  * and numelems contains it's length.
3598  */
3599 int
3600 dca_attr_lookup_uint8_array(crypto_object_attribute_t *attrp, uint_t atnum,
3601     uint64_t atype, void **data, unsigned int *numelems)
3602 {
3603 	crypto_object_attribute_t	*bap;
3604 
3605 	bap = dca_find_attribute(attrp, atnum, atype);
3606 	if (bap == NULL) {
3607 		return (CRYPTO_ATTRIBUTE_TYPE_INVALID);
3608 	}
3609 
3610 	*data = bap->oa_value;
3611 	*numelems = bap->oa_value_len;
3612 
3613 	return (CRYPTO_SUCCESS);
3614 }
3615 
3616 /*
3617  * Finds entry of specified name. If it is not found dca_find_attribute returns
3618  * NULL.
3619  */
3620 crypto_object_attribute_t *
3621 dca_find_attribute(crypto_object_attribute_t *attrp, uint_t atnum,
3622     uint64_t atype)
3623 {
3624 	while (atnum) {
3625 		if (attrp->oa_type == atype)
3626 			return (attrp);
3627 		atnum--;
3628 		attrp++;
3629 	}
3630 	return (NULL);
3631 }
3632 
3633 /*
3634  * Return the address of the first data buffer. If the data format is
3635  * unrecognised return NULL.
3636  */
3637 caddr_t
3638 dca_bufdaddr(crypto_data_t *data)
3639 {
3640 	switch (data->cd_format) {
3641 	case CRYPTO_DATA_RAW:
3642 		return (data->cd_raw.iov_base + data->cd_offset);
3643 	case CRYPTO_DATA_UIO:
3644 		return (data->cd_uio->uio_iov[0].iov_base + data->cd_offset);
3645 	case CRYPTO_DATA_MBLK:
3646 		return ((char *)data->cd_mp->b_rptr + data->cd_offset);
3647 	default:
3648 		DBG(NULL, DWARN,
3649 		    "dca_bufdaddr: unrecognised crypto data format");
3650 		return (NULL);
3651 	}
3652 }
3653 
3654 static caddr_t
3655 dca_bufdaddr_out(crypto_data_t *data)
3656 {
3657 	size_t offset = data->cd_offset + data->cd_length;
3658 
3659 	switch (data->cd_format) {
3660 	case CRYPTO_DATA_RAW:
3661 		return (data->cd_raw.iov_base + offset);
3662 	case CRYPTO_DATA_UIO:
3663 		return (data->cd_uio->uio_iov[0].iov_base + offset);
3664 	case CRYPTO_DATA_MBLK:
3665 		return ((char *)data->cd_mp->b_rptr + offset);
3666 	default:
3667 		DBG(NULL, DWARN,
3668 		    "dca_bufdaddr_out: unrecognised crypto data format");
3669 		return (NULL);
3670 	}
3671 }
3672 
3673 /*
3674  * Control entry points.
3675  */
3676 
3677 /* ARGSUSED */
3678 static void
3679 dca_provider_status(crypto_provider_handle_t provider, uint_t *status)
3680 {
3681 	*status = CRYPTO_PROVIDER_READY;
3682 }
3683 
3684 /*
3685  * Cipher (encrypt/decrypt) entry points.
3686  */
3687 
3688 /* ARGSUSED */
3689 static int
3690 dca_encrypt_init(crypto_ctx_t *ctx, crypto_mechanism_t *mechanism,
3691     crypto_key_t *key, crypto_spi_ctx_template_t ctx_template,
3692     crypto_req_handle_t req)
3693 {
3694 	int error = CRYPTO_FAILED;
3695 	dca_t *softc;
3696 	/* LINTED E_FUNC_SET_NOT_USED */
3697 	int instance;
3698 
3699 	/* extract softc and instance number from context */
3700 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
3701 	DBG(softc, DENTRY, "dca_encrypt_init: started");
3702 
3703 	/* check mechanism */
3704 	switch (mechanism->cm_type) {
3705 	case DES_CBC_MECH_INFO_TYPE:
3706 		error = dca_3desctxinit(ctx, mechanism, key, KM_SLEEP,
3707 		    DR_ENCRYPT);
3708 		break;
3709 	case DES3_CBC_MECH_INFO_TYPE:
3710 		error = dca_3desctxinit(ctx, mechanism, key, KM_SLEEP,
3711 		    DR_ENCRYPT | DR_TRIPLE);
3712 		break;
3713 	case RSA_PKCS_MECH_INFO_TYPE:
3714 	case RSA_X_509_MECH_INFO_TYPE:
3715 		error = dca_rsainit(ctx, mechanism, key, KM_SLEEP);
3716 		break;
3717 	default:
3718 		cmn_err(CE_WARN, "dca_encrypt_init: unexpected mech type "
3719 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
3720 		error = CRYPTO_MECHANISM_INVALID;
3721 	}
3722 
3723 	DBG(softc, DENTRY, "dca_encrypt_init: done, err = 0x%x", error);
3724 
3725 	if (error == CRYPTO_SUCCESS)
3726 		dca_enlist2(&softc->dca_ctx_list, ctx->cc_provider_private,
3727 		    &softc->dca_ctx_list_lock);
3728 
3729 	return (error);
3730 }
3731 
3732 /* ARGSUSED */
3733 static int
3734 dca_encrypt(crypto_ctx_t *ctx, crypto_data_t *plaintext,
3735     crypto_data_t *ciphertext, crypto_req_handle_t req)
3736 {
3737 	int error = CRYPTO_FAILED;
3738 	dca_t *softc;
3739 	/* LINTED E_FUNC_SET_NOT_USED */
3740 	int instance;
3741 
3742 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
3743 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
3744 
3745 	/* extract softc and instance number from context */
3746 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
3747 	DBG(softc, DENTRY, "dca_encrypt: started");
3748 
3749 	/* handle inplace ops */
3750 	if (!ciphertext) {
3751 		dca_request_t *reqp = ctx->cc_provider_private;
3752 		reqp->dr_flags |= DR_INPLACE;
3753 		ciphertext = plaintext;
3754 	}
3755 
3756 	/* check mechanism */
3757 	switch (DCA_MECH_FROM_CTX(ctx)) {
3758 	case DES_CBC_MECH_INFO_TYPE:
3759 		error = dca_3des(ctx, plaintext, ciphertext, req, DR_ENCRYPT);
3760 		break;
3761 	case DES3_CBC_MECH_INFO_TYPE:
3762 		error = dca_3des(ctx, plaintext, ciphertext, req,
3763 		    DR_ENCRYPT | DR_TRIPLE);
3764 		break;
3765 	case RSA_PKCS_MECH_INFO_TYPE:
3766 	case RSA_X_509_MECH_INFO_TYPE:
3767 		error = dca_rsastart(ctx, plaintext, ciphertext, req,
3768 		    DCA_RSA_ENC);
3769 		break;
3770 	default:
3771 		/* Should never reach here */
3772 		cmn_err(CE_WARN, "dca_encrypt: unexpected mech type "
3773 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
3774 		error = CRYPTO_MECHANISM_INVALID;
3775 	}
3776 
3777 	if ((error != CRYPTO_QUEUED) && (error != CRYPTO_SUCCESS) &&
3778 	    (error != CRYPTO_BUFFER_TOO_SMALL)) {
3779 		ciphertext->cd_length = 0;
3780 	}
3781 
3782 	DBG(softc, DENTRY, "dca_encrypt: done, err = 0x%x", error);
3783 
3784 	return (error);
3785 }
3786 
3787 /* ARGSUSED */
3788 static int
3789 dca_encrypt_update(crypto_ctx_t *ctx, crypto_data_t *plaintext,
3790     crypto_data_t *ciphertext, crypto_req_handle_t req)
3791 {
3792 	int error = CRYPTO_FAILED;
3793 	dca_t *softc;
3794 	/* LINTED E_FUNC_SET_NOT_USED */
3795 	int instance;
3796 
3797 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
3798 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
3799 
3800 	/* extract softc and instance number from context */
3801 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
3802 	DBG(softc, DENTRY, "dca_encrypt_update: started");
3803 
3804 	/* handle inplace ops */
3805 	if (!ciphertext) {
3806 		dca_request_t *reqp = ctx->cc_provider_private;
3807 		reqp->dr_flags |= DR_INPLACE;
3808 		ciphertext = plaintext;
3809 	}
3810 
3811 	/* check mechanism */
3812 	switch (DCA_MECH_FROM_CTX(ctx)) {
3813 	case DES_CBC_MECH_INFO_TYPE:
3814 		error = dca_3desupdate(ctx, plaintext, ciphertext, req,
3815 		    DR_ENCRYPT);
3816 		break;
3817 	case DES3_CBC_MECH_INFO_TYPE:
3818 		error = dca_3desupdate(ctx, plaintext, ciphertext, req,
3819 		    DR_ENCRYPT | DR_TRIPLE);
3820 		break;
3821 	default:
3822 		/* Should never reach here */
3823 		cmn_err(CE_WARN, "dca_encrypt_update: unexpected mech type "
3824 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
3825 		error = CRYPTO_MECHANISM_INVALID;
3826 	}
3827 
3828 	DBG(softc, DENTRY, "dca_encrypt_update: done, err = 0x%x", error);
3829 
3830 	return (error);
3831 }
3832 
3833 /* ARGSUSED */
3834 static int
3835 dca_encrypt_final(crypto_ctx_t *ctx, crypto_data_t *ciphertext,
3836     crypto_req_handle_t req)
3837 {
3838 	int error = CRYPTO_FAILED;
3839 	dca_t *softc;
3840 	/* LINTED E_FUNC_SET_NOT_USED */
3841 	int instance;
3842 
3843 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
3844 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
3845 
3846 	/* extract softc and instance number from context */
3847 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
3848 	DBG(softc, DENTRY, "dca_encrypt_final: started");
3849 
3850 	/* check mechanism */
3851 	switch (DCA_MECH_FROM_CTX(ctx)) {
3852 	case DES_CBC_MECH_INFO_TYPE:
3853 		error = dca_3desfinal(ctx, ciphertext, DR_ENCRYPT);
3854 		break;
3855 	case DES3_CBC_MECH_INFO_TYPE:
3856 		error = dca_3desfinal(ctx, ciphertext, DR_ENCRYPT | DR_TRIPLE);
3857 		break;
3858 	default:
3859 		/* Should never reach here */
3860 		cmn_err(CE_WARN, "dca_encrypt_final: unexpected mech type "
3861 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
3862 		error = CRYPTO_MECHANISM_INVALID;
3863 	}
3864 
3865 	DBG(softc, DENTRY, "dca_encrypt_final: done, err = 0x%x", error);
3866 
3867 	return (error);
3868 }
3869 
3870 /* ARGSUSED */
3871 static int
3872 dca_encrypt_atomic(crypto_provider_handle_t provider,
3873     crypto_session_id_t session_id, crypto_mechanism_t *mechanism,
3874     crypto_key_t *key, crypto_data_t *plaintext, crypto_data_t *ciphertext,
3875     crypto_spi_ctx_template_t ctx_template, crypto_req_handle_t req)
3876 {
3877 	int error = CRYPTO_FAILED;
3878 	dca_t *softc = (dca_t *)provider;
3879 
3880 	DBG(softc, DENTRY, "dca_encrypt_atomic: started");
3881 
3882 	if (ctx_template != NULL)
3883 		return (CRYPTO_ARGUMENTS_BAD);
3884 
3885 	/* handle inplace ops */
3886 	if (!ciphertext) {
3887 		ciphertext = plaintext;
3888 	}
3889 
3890 	/* check mechanism */
3891 	switch (mechanism->cm_type) {
3892 	case DES_CBC_MECH_INFO_TYPE:
3893 		error = dca_3desatomic(provider, session_id, mechanism, key,
3894 		    plaintext, ciphertext, KM_SLEEP, req,
3895 		    DR_ENCRYPT | DR_ATOMIC);
3896 		break;
3897 	case DES3_CBC_MECH_INFO_TYPE:
3898 		error = dca_3desatomic(provider, session_id, mechanism, key,
3899 		    plaintext, ciphertext, KM_SLEEP, req,
3900 		    DR_ENCRYPT | DR_TRIPLE | DR_ATOMIC);
3901 		break;
3902 	case RSA_PKCS_MECH_INFO_TYPE:
3903 	case RSA_X_509_MECH_INFO_TYPE:
3904 		error = dca_rsaatomic(provider, session_id, mechanism, key,
3905 		    plaintext, ciphertext, KM_SLEEP, req, DCA_RSA_ENC);
3906 		break;
3907 	default:
3908 		cmn_err(CE_WARN, "dca_encrypt_atomic: unexpected mech type "
3909 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
3910 		error = CRYPTO_MECHANISM_INVALID;
3911 	}
3912 
3913 	if ((error != CRYPTO_QUEUED) && (error != CRYPTO_SUCCESS)) {
3914 		ciphertext->cd_length = 0;
3915 	}
3916 
3917 	DBG(softc, DENTRY, "dca_encrypt_atomic: done, err = 0x%x", error);
3918 
3919 	return (error);
3920 }
3921 
3922 /* ARGSUSED */
3923 static int
3924 dca_decrypt_init(crypto_ctx_t *ctx, crypto_mechanism_t *mechanism,
3925     crypto_key_t *key, crypto_spi_ctx_template_t ctx_template,
3926     crypto_req_handle_t req)
3927 {
3928 	int error = CRYPTO_FAILED;
3929 	dca_t *softc;
3930 	/* LINTED E_FUNC_SET_NOT_USED */
3931 	int instance;
3932 
3933 	/* extract softc and instance number from context */
3934 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
3935 	DBG(softc, DENTRY, "dca_decrypt_init: started");
3936 
3937 	/* check mechanism */
3938 	switch (mechanism->cm_type) {
3939 	case DES_CBC_MECH_INFO_TYPE:
3940 		error = dca_3desctxinit(ctx, mechanism, key, KM_SLEEP,
3941 		    DR_DECRYPT);
3942 		break;
3943 	case DES3_CBC_MECH_INFO_TYPE:
3944 		error = dca_3desctxinit(ctx, mechanism, key, KM_SLEEP,
3945 		    DR_DECRYPT | DR_TRIPLE);
3946 		break;
3947 	case RSA_PKCS_MECH_INFO_TYPE:
3948 	case RSA_X_509_MECH_INFO_TYPE:
3949 		error = dca_rsainit(ctx, mechanism, key, KM_SLEEP);
3950 		break;
3951 	default:
3952 		cmn_err(CE_WARN, "dca_decrypt_init: unexpected mech type "
3953 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
3954 		error = CRYPTO_MECHANISM_INVALID;
3955 	}
3956 
3957 	DBG(softc, DENTRY, "dca_decrypt_init: done, err = 0x%x", error);
3958 
3959 	if (error == CRYPTO_SUCCESS)
3960 		dca_enlist2(&softc->dca_ctx_list, ctx->cc_provider_private,
3961 		    &softc->dca_ctx_list_lock);
3962 
3963 	return (error);
3964 }
3965 
3966 /* ARGSUSED */
3967 static int
3968 dca_decrypt(crypto_ctx_t *ctx, crypto_data_t *ciphertext,
3969     crypto_data_t *plaintext, crypto_req_handle_t req)
3970 {
3971 	int error = CRYPTO_FAILED;
3972 	dca_t *softc;
3973 	/* LINTED E_FUNC_SET_NOT_USED */
3974 	int instance;
3975 
3976 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
3977 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
3978 
3979 	/* extract softc and instance number from context */
3980 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
3981 	DBG(softc, DENTRY, "dca_decrypt: started");
3982 
3983 	/* handle inplace ops */
3984 	if (!plaintext) {
3985 		dca_request_t *reqp = ctx->cc_provider_private;
3986 		reqp->dr_flags |= DR_INPLACE;
3987 		plaintext = ciphertext;
3988 	}
3989 
3990 	/* check mechanism */
3991 	switch (DCA_MECH_FROM_CTX(ctx)) {
3992 	case DES_CBC_MECH_INFO_TYPE:
3993 		error = dca_3des(ctx, ciphertext, plaintext, req, DR_DECRYPT);
3994 		break;
3995 	case DES3_CBC_MECH_INFO_TYPE:
3996 		error = dca_3des(ctx, ciphertext, plaintext, req,
3997 		    DR_DECRYPT | DR_TRIPLE);
3998 		break;
3999 	case RSA_PKCS_MECH_INFO_TYPE:
4000 	case RSA_X_509_MECH_INFO_TYPE:
4001 		error = dca_rsastart(ctx, ciphertext, plaintext, req,
4002 		    DCA_RSA_DEC);
4003 		break;
4004 	default:
4005 		/* Should never reach here */
4006 		cmn_err(CE_WARN, "dca_decrypt: unexpected mech type "
4007 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4008 		error = CRYPTO_MECHANISM_INVALID;
4009 	}
4010 
4011 	if ((error != CRYPTO_QUEUED) && (error != CRYPTO_SUCCESS) &&
4012 	    (error != CRYPTO_BUFFER_TOO_SMALL)) {
4013 		if (plaintext)
4014 			plaintext->cd_length = 0;
4015 	}
4016 
4017 	DBG(softc, DENTRY, "dca_decrypt: done, err = 0x%x", error);
4018 
4019 	return (error);
4020 }
4021 
4022 /* ARGSUSED */
4023 static int
4024 dca_decrypt_update(crypto_ctx_t *ctx, crypto_data_t *ciphertext,
4025     crypto_data_t *plaintext, crypto_req_handle_t req)
4026 {
4027 	int error = CRYPTO_FAILED;
4028 	dca_t *softc;
4029 	/* LINTED E_FUNC_SET_NOT_USED */
4030 	int instance;
4031 
4032 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4033 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4034 
4035 	/* extract softc and instance number from context */
4036 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4037 	DBG(softc, DENTRY, "dca_decrypt_update: started");
4038 
4039 	/* handle inplace ops */
4040 	if (!plaintext) {
4041 		dca_request_t *reqp = ctx->cc_provider_private;
4042 		reqp->dr_flags |= DR_INPLACE;
4043 		plaintext = ciphertext;
4044 	}
4045 
4046 	/* check mechanism */
4047 	switch (DCA_MECH_FROM_CTX(ctx)) {
4048 	case DES_CBC_MECH_INFO_TYPE:
4049 		error = dca_3desupdate(ctx, ciphertext, plaintext, req,
4050 		    DR_DECRYPT);
4051 		break;
4052 	case DES3_CBC_MECH_INFO_TYPE:
4053 		error = dca_3desupdate(ctx, ciphertext, plaintext, req,
4054 		    DR_DECRYPT | DR_TRIPLE);
4055 		break;
4056 	default:
4057 		/* Should never reach here */
4058 		cmn_err(CE_WARN, "dca_decrypt_update: unexpected mech type "
4059 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4060 		error = CRYPTO_MECHANISM_INVALID;
4061 	}
4062 
4063 	DBG(softc, DENTRY, "dca_decrypt_update: done, err = 0x%x", error);
4064 
4065 	return (error);
4066 }
4067 
4068 /* ARGSUSED */
4069 static int
4070 dca_decrypt_final(crypto_ctx_t *ctx, crypto_data_t *plaintext,
4071     crypto_req_handle_t req)
4072 {
4073 	int error = CRYPTO_FAILED;
4074 	dca_t *softc;
4075 	/* LINTED E_FUNC_SET_NOT_USED */
4076 	int instance;
4077 
4078 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4079 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4080 
4081 	/* extract softc and instance number from context */
4082 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4083 	DBG(softc, DENTRY, "dca_decrypt_final: started");
4084 
4085 	/* check mechanism */
4086 	switch (DCA_MECH_FROM_CTX(ctx)) {
4087 	case DES_CBC_MECH_INFO_TYPE:
4088 		error = dca_3desfinal(ctx, plaintext, DR_DECRYPT);
4089 		break;
4090 	case DES3_CBC_MECH_INFO_TYPE:
4091 		error = dca_3desfinal(ctx, plaintext, DR_DECRYPT | DR_TRIPLE);
4092 		break;
4093 	default:
4094 		/* Should never reach here */
4095 		cmn_err(CE_WARN, "dca_decrypt_final: unexpected mech type "
4096 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4097 		error = CRYPTO_MECHANISM_INVALID;
4098 	}
4099 
4100 	DBG(softc, DENTRY, "dca_decrypt_final: done, err = 0x%x", error);
4101 
4102 	return (error);
4103 }
4104 
4105 /* ARGSUSED */
4106 static int
4107 dca_decrypt_atomic(crypto_provider_handle_t provider,
4108     crypto_session_id_t session_id, crypto_mechanism_t *mechanism,
4109     crypto_key_t *key, crypto_data_t *ciphertext, crypto_data_t *plaintext,
4110     crypto_spi_ctx_template_t ctx_template, crypto_req_handle_t req)
4111 {
4112 	int error = CRYPTO_FAILED;
4113 	dca_t *softc = (dca_t *)provider;
4114 
4115 	DBG(softc, DENTRY, "dca_decrypt_atomic: started");
4116 
4117 	if (ctx_template != NULL)
4118 		return (CRYPTO_ARGUMENTS_BAD);
4119 
4120 	/* handle inplace ops */
4121 	if (!plaintext) {
4122 		plaintext = ciphertext;
4123 	}
4124 
4125 	/* check mechanism */
4126 	switch (mechanism->cm_type) {
4127 	case DES_CBC_MECH_INFO_TYPE:
4128 		error = dca_3desatomic(provider, session_id, mechanism, key,
4129 		    ciphertext, plaintext, KM_SLEEP, req,
4130 		    DR_DECRYPT | DR_ATOMIC);
4131 		break;
4132 	case DES3_CBC_MECH_INFO_TYPE:
4133 		error = dca_3desatomic(provider, session_id, mechanism, key,
4134 		    ciphertext, plaintext, KM_SLEEP, req,
4135 		    DR_DECRYPT | DR_TRIPLE | DR_ATOMIC);
4136 		break;
4137 	case RSA_PKCS_MECH_INFO_TYPE:
4138 	case RSA_X_509_MECH_INFO_TYPE:
4139 		error = dca_rsaatomic(provider, session_id, mechanism, key,
4140 		    ciphertext, plaintext, KM_SLEEP, req, DCA_RSA_DEC);
4141 		break;
4142 	default:
4143 		cmn_err(CE_WARN, "dca_decrypt_atomic: unexpected mech type "
4144 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
4145 		error = CRYPTO_MECHANISM_INVALID;
4146 	}
4147 
4148 	if ((error != CRYPTO_QUEUED) && (error != CRYPTO_SUCCESS)) {
4149 		plaintext->cd_length = 0;
4150 	}
4151 
4152 	DBG(softc, DENTRY, "dca_decrypt_atomic: done, err = 0x%x", error);
4153 
4154 	return (error);
4155 }
4156 
4157 /*
4158  * Sign entry points.
4159  */
4160 
4161 /* ARGSUSED */
4162 static int
4163 dca_sign_init(crypto_ctx_t *ctx, crypto_mechanism_t *mechanism,
4164     crypto_key_t *key, crypto_spi_ctx_template_t ctx_template,
4165     crypto_req_handle_t req)
4166 {
4167 	int error = CRYPTO_FAILED;
4168 	dca_t *softc;
4169 	/* LINTED E_FUNC_SET_NOT_USED */
4170 	int instance;
4171 
4172 	/* extract softc and instance number from context */
4173 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4174 	DBG(softc, DENTRY, "dca_sign_init: started\n");
4175 
4176 	if (ctx_template != NULL)
4177 		return (CRYPTO_ARGUMENTS_BAD);
4178 
4179 	/* check mechanism */
4180 	switch (mechanism->cm_type) {
4181 	case RSA_PKCS_MECH_INFO_TYPE:
4182 	case RSA_X_509_MECH_INFO_TYPE:
4183 		error = dca_rsainit(ctx, mechanism, key, KM_SLEEP);
4184 		break;
4185 	case DSA_MECH_INFO_TYPE:
4186 		error = dca_dsainit(ctx, mechanism, key, KM_SLEEP,
4187 		    DCA_DSA_SIGN);
4188 		break;
4189 	default:
4190 		cmn_err(CE_WARN, "dca_sign_init: unexpected mech type "
4191 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
4192 		error = CRYPTO_MECHANISM_INVALID;
4193 	}
4194 
4195 	DBG(softc, DENTRY, "dca_sign_init: done, err = 0x%x", error);
4196 
4197 	if (error == CRYPTO_SUCCESS)
4198 		dca_enlist2(&softc->dca_ctx_list, ctx->cc_provider_private,
4199 		    &softc->dca_ctx_list_lock);
4200 
4201 	return (error);
4202 }
4203 
4204 static int
4205 dca_sign(crypto_ctx_t *ctx, crypto_data_t *data,
4206     crypto_data_t *signature, crypto_req_handle_t req)
4207 {
4208 	int error = CRYPTO_FAILED;
4209 	dca_t *softc;
4210 	/* LINTED E_FUNC_SET_NOT_USED */
4211 	int instance;
4212 
4213 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4214 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4215 
4216 	/* extract softc and instance number from context */
4217 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4218 	DBG(softc, DENTRY, "dca_sign: started\n");
4219 
4220 	/* check mechanism */
4221 	switch (DCA_MECH_FROM_CTX(ctx)) {
4222 	case RSA_PKCS_MECH_INFO_TYPE:
4223 	case RSA_X_509_MECH_INFO_TYPE:
4224 		error = dca_rsastart(ctx, data, signature, req, DCA_RSA_SIGN);
4225 		break;
4226 	case DSA_MECH_INFO_TYPE:
4227 		error = dca_dsa_sign(ctx, data, signature, req);
4228 		break;
4229 	default:
4230 		cmn_err(CE_WARN, "dca_sign: unexpected mech type "
4231 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4232 		error = CRYPTO_MECHANISM_INVALID;
4233 	}
4234 
4235 	DBG(softc, DENTRY, "dca_sign: done, err = 0x%x", error);
4236 
4237 	return (error);
4238 }
4239 
4240 /* ARGSUSED */
4241 static int
4242 dca_sign_update(crypto_ctx_t *ctx, crypto_data_t *data,
4243     crypto_req_handle_t req)
4244 {
4245 	int error = CRYPTO_MECHANISM_INVALID;
4246 	dca_t *softc;
4247 	/* LINTED E_FUNC_SET_NOT_USED */
4248 	int instance;
4249 
4250 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4251 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4252 
4253 	/* extract softc and instance number from context */
4254 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4255 	DBG(softc, DENTRY, "dca_sign_update: started\n");
4256 
4257 	cmn_err(CE_WARN, "dca_sign_update: unexpected mech type "
4258 	    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4259 
4260 	DBG(softc, DENTRY, "dca_sign_update: done, err = 0x%x", error);
4261 
4262 	return (error);
4263 }
4264 
4265 /* ARGSUSED */
4266 static int
4267 dca_sign_final(crypto_ctx_t *ctx, crypto_data_t *signature,
4268     crypto_req_handle_t req)
4269 {
4270 	int error = CRYPTO_MECHANISM_INVALID;
4271 	dca_t *softc;
4272 	/* LINTED E_FUNC_SET_NOT_USED */
4273 	int instance;
4274 
4275 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4276 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4277 
4278 	/* extract softc and instance number from context */
4279 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4280 	DBG(softc, DENTRY, "dca_sign_final: started\n");
4281 
4282 	cmn_err(CE_WARN, "dca_sign_final: unexpected mech type "
4283 	    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4284 
4285 	DBG(softc, DENTRY, "dca_sign_final: done, err = 0x%x", error);
4286 
4287 	return (error);
4288 }
4289 
4290 static int
4291 dca_sign_atomic(crypto_provider_handle_t provider,
4292     crypto_session_id_t session_id, crypto_mechanism_t *mechanism,
4293     crypto_key_t *key, crypto_data_t *data, crypto_data_t *signature,
4294     crypto_spi_ctx_template_t ctx_template, crypto_req_handle_t req)
4295 {
4296 	int error = CRYPTO_FAILED;
4297 	dca_t *softc = (dca_t *)provider;
4298 
4299 	DBG(softc, DENTRY, "dca_sign_atomic: started\n");
4300 
4301 	if (ctx_template != NULL)
4302 		return (CRYPTO_ARGUMENTS_BAD);
4303 
4304 	/* check mechanism */
4305 	switch (mechanism->cm_type) {
4306 	case RSA_PKCS_MECH_INFO_TYPE:
4307 	case RSA_X_509_MECH_INFO_TYPE:
4308 		error = dca_rsaatomic(provider, session_id, mechanism, key,
4309 		    data, signature, KM_SLEEP, req, DCA_RSA_SIGN);
4310 		break;
4311 	case DSA_MECH_INFO_TYPE:
4312 		error = dca_dsaatomic(provider, session_id, mechanism, key,
4313 		    data, signature, KM_SLEEP, req, DCA_DSA_SIGN);
4314 		break;
4315 	default:
4316 		cmn_err(CE_WARN, "dca_sign_atomic: unexpected mech type "
4317 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
4318 		error = CRYPTO_MECHANISM_INVALID;
4319 	}
4320 
4321 	DBG(softc, DENTRY, "dca_sign_atomic: done, err = 0x%x", error);
4322 
4323 	return (error);
4324 }
4325 
4326 /* ARGSUSED */
4327 static int
4328 dca_sign_recover_init(crypto_ctx_t *ctx, crypto_mechanism_t *mechanism,
4329     crypto_key_t *key, crypto_spi_ctx_template_t ctx_template,
4330     crypto_req_handle_t req)
4331 {
4332 	int error = CRYPTO_FAILED;
4333 	dca_t *softc;
4334 	/* LINTED E_FUNC_SET_NOT_USED */
4335 	int instance;
4336 
4337 	/* extract softc and instance number from context */
4338 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4339 	DBG(softc, DENTRY, "dca_sign_recover_init: started\n");
4340 
4341 	if (ctx_template != NULL)
4342 		return (CRYPTO_ARGUMENTS_BAD);
4343 
4344 	/* check mechanism */
4345 	switch (mechanism->cm_type) {
4346 	case RSA_PKCS_MECH_INFO_TYPE:
4347 	case RSA_X_509_MECH_INFO_TYPE:
4348 		error = dca_rsainit(ctx, mechanism, key, KM_SLEEP);
4349 		break;
4350 	default:
4351 		cmn_err(CE_WARN, "dca_sign_recover_init: unexpected mech type "
4352 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
4353 		error = CRYPTO_MECHANISM_INVALID;
4354 	}
4355 
4356 	DBG(softc, DENTRY, "dca_sign_recover_init: done, err = 0x%x", error);
4357 
4358 	if (error == CRYPTO_SUCCESS)
4359 		dca_enlist2(&softc->dca_ctx_list, ctx->cc_provider_private,
4360 		    &softc->dca_ctx_list_lock);
4361 
4362 	return (error);
4363 }
4364 
4365 static int
4366 dca_sign_recover(crypto_ctx_t *ctx, crypto_data_t *data,
4367     crypto_data_t *signature, crypto_req_handle_t req)
4368 {
4369 	int error = CRYPTO_FAILED;
4370 	dca_t *softc;
4371 	/* LINTED E_FUNC_SET_NOT_USED */
4372 	int instance;
4373 
4374 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4375 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4376 
4377 	/* extract softc and instance number from context */
4378 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4379 	DBG(softc, DENTRY, "dca_sign_recover: started\n");
4380 
4381 	/* check mechanism */
4382 	switch (DCA_MECH_FROM_CTX(ctx)) {
4383 	case RSA_PKCS_MECH_INFO_TYPE:
4384 	case RSA_X_509_MECH_INFO_TYPE:
4385 		error = dca_rsastart(ctx, data, signature, req, DCA_RSA_SIGNR);
4386 		break;
4387 	default:
4388 		cmn_err(CE_WARN, "dca_sign_recover: unexpected mech type "
4389 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4390 		error = CRYPTO_MECHANISM_INVALID;
4391 	}
4392 
4393 	DBG(softc, DENTRY, "dca_sign_recover: done, err = 0x%x", error);
4394 
4395 	return (error);
4396 }
4397 
4398 static int
4399 dca_sign_recover_atomic(crypto_provider_handle_t provider,
4400     crypto_session_id_t session_id, crypto_mechanism_t *mechanism,
4401     crypto_key_t *key, crypto_data_t *data, crypto_data_t *signature,
4402     crypto_spi_ctx_template_t ctx_template, crypto_req_handle_t req)
4403 {
4404 	int error = CRYPTO_FAILED;
4405 	dca_t *softc = (dca_t *)provider;
4406 	/* LINTED E_FUNC_SET_NOT_USED */
4407 	int instance;
4408 
4409 	instance = ddi_get_instance(softc->dca_dip);
4410 	DBG(softc, DENTRY, "dca_sign_recover_atomic: started\n");
4411 
4412 	if (ctx_template != NULL)
4413 		return (CRYPTO_ARGUMENTS_BAD);
4414 
4415 	/* check mechanism */
4416 	switch (mechanism->cm_type) {
4417 	case RSA_PKCS_MECH_INFO_TYPE:
4418 	case RSA_X_509_MECH_INFO_TYPE:
4419 		error = dca_rsaatomic(provider, session_id, mechanism, key,
4420 		    data, signature, KM_SLEEP, req, DCA_RSA_SIGNR);
4421 		break;
4422 	default:
4423 		cmn_err(CE_WARN, "dca_sign_recover_atomic: unexpected mech type"
4424 		    " 0x%llx\n", (unsigned long long)mechanism->cm_type);
4425 		error = CRYPTO_MECHANISM_INVALID;
4426 	}
4427 
4428 	DBG(softc, DENTRY, "dca_sign_recover_atomic: done, err = 0x%x", error);
4429 
4430 	return (error);
4431 }
4432 
4433 /*
4434  * Verify entry points.
4435  */
4436 
4437 /* ARGSUSED */
4438 static int
4439 dca_verify_init(crypto_ctx_t *ctx, crypto_mechanism_t *mechanism,
4440     crypto_key_t *key, crypto_spi_ctx_template_t ctx_template,
4441     crypto_req_handle_t req)
4442 {
4443 	int error = CRYPTO_FAILED;
4444 	dca_t *softc;
4445 	/* LINTED E_FUNC_SET_NOT_USED */
4446 	int instance;
4447 
4448 	/* extract softc and instance number from context */
4449 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4450 	DBG(softc, DENTRY, "dca_verify_init: started\n");
4451 
4452 	if (ctx_template != NULL)
4453 		return (CRYPTO_ARGUMENTS_BAD);
4454 
4455 	/* check mechanism */
4456 	switch (mechanism->cm_type) {
4457 	case RSA_PKCS_MECH_INFO_TYPE:
4458 	case RSA_X_509_MECH_INFO_TYPE:
4459 		error = dca_rsainit(ctx, mechanism, key, KM_SLEEP);
4460 		break;
4461 	case DSA_MECH_INFO_TYPE:
4462 		error = dca_dsainit(ctx, mechanism, key, KM_SLEEP,
4463 		    DCA_DSA_VRFY);
4464 		break;
4465 	default:
4466 		cmn_err(CE_WARN, "dca_verify_init: unexpected mech type "
4467 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
4468 		error = CRYPTO_MECHANISM_INVALID;
4469 	}
4470 
4471 	DBG(softc, DENTRY, "dca_verify_init: done, err = 0x%x", error);
4472 
4473 	if (error == CRYPTO_SUCCESS)
4474 		dca_enlist2(&softc->dca_ctx_list, ctx->cc_provider_private,
4475 		    &softc->dca_ctx_list_lock);
4476 
4477 	return (error);
4478 }
4479 
4480 static int
4481 dca_verify(crypto_ctx_t *ctx, crypto_data_t *data, crypto_data_t *signature,
4482     crypto_req_handle_t req)
4483 {
4484 	int error = CRYPTO_FAILED;
4485 	dca_t *softc;
4486 	/* LINTED E_FUNC_SET_NOT_USED */
4487 	int instance;
4488 
4489 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4490 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4491 
4492 	/* extract softc and instance number from context */
4493 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4494 	DBG(softc, DENTRY, "dca_verify: started\n");
4495 
4496 	/* check mechanism */
4497 	switch (DCA_MECH_FROM_CTX(ctx)) {
4498 	case RSA_PKCS_MECH_INFO_TYPE:
4499 	case RSA_X_509_MECH_INFO_TYPE:
4500 		error = dca_rsastart(ctx, signature, data, req, DCA_RSA_VRFY);
4501 		break;
4502 	case DSA_MECH_INFO_TYPE:
4503 		error = dca_dsa_verify(ctx, data, signature, req);
4504 		break;
4505 	default:
4506 		cmn_err(CE_WARN, "dca_verify: unexpected mech type "
4507 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4508 		error = CRYPTO_MECHANISM_INVALID;
4509 	}
4510 
4511 	DBG(softc, DENTRY, "dca_verify: done, err = 0x%x", error);
4512 
4513 	return (error);
4514 }
4515 
4516 /* ARGSUSED */
4517 static int
4518 dca_verify_update(crypto_ctx_t *ctx, crypto_data_t *data,
4519     crypto_req_handle_t req)
4520 {
4521 	int error = CRYPTO_MECHANISM_INVALID;
4522 	dca_t *softc;
4523 	/* LINTED E_FUNC_SET_NOT_USED */
4524 	int instance;
4525 
4526 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4527 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4528 
4529 	/* extract softc and instance number from context */
4530 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4531 	DBG(softc, DENTRY, "dca_verify_update: started\n");
4532 
4533 	cmn_err(CE_WARN, "dca_verify_update: unexpected mech type "
4534 	    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4535 
4536 	DBG(softc, DENTRY, "dca_verify_update: done, err = 0x%x", error);
4537 
4538 	return (error);
4539 }
4540 
4541 /* ARGSUSED */
4542 static int
4543 dca_verify_final(crypto_ctx_t *ctx, crypto_data_t *signature,
4544     crypto_req_handle_t req)
4545 {
4546 	int error = CRYPTO_MECHANISM_INVALID;
4547 	dca_t *softc;
4548 	/* LINTED E_FUNC_SET_NOT_USED */
4549 	int instance;
4550 
4551 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4552 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4553 
4554 	/* extract softc and instance number from context */
4555 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4556 	DBG(softc, DENTRY, "dca_verify_final: started\n");
4557 
4558 	cmn_err(CE_WARN, "dca_verify_final: unexpected mech type "
4559 	    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4560 
4561 	DBG(softc, DENTRY, "dca_verify_final: done, err = 0x%x", error);
4562 
4563 	return (error);
4564 }
4565 
4566 static int
4567 dca_verify_atomic(crypto_provider_handle_t provider,
4568     crypto_session_id_t session_id, crypto_mechanism_t *mechanism,
4569     crypto_key_t *key, crypto_data_t *data, crypto_data_t *signature,
4570     crypto_spi_ctx_template_t ctx_template, crypto_req_handle_t req)
4571 {
4572 	int error = CRYPTO_FAILED;
4573 	dca_t *softc = (dca_t *)provider;
4574 
4575 	DBG(softc, DENTRY, "dca_verify_atomic: started\n");
4576 
4577 	if (ctx_template != NULL)
4578 		return (CRYPTO_ARGUMENTS_BAD);
4579 
4580 	/* check mechanism */
4581 	switch (mechanism->cm_type) {
4582 	case RSA_PKCS_MECH_INFO_TYPE:
4583 	case RSA_X_509_MECH_INFO_TYPE:
4584 		error = dca_rsaatomic(provider, session_id, mechanism, key,
4585 		    signature, data, KM_SLEEP, req, DCA_RSA_VRFY);
4586 		break;
4587 	case DSA_MECH_INFO_TYPE:
4588 		error = dca_dsaatomic(provider, session_id, mechanism, key,
4589 		    data, signature, KM_SLEEP, req, DCA_DSA_VRFY);
4590 		break;
4591 	default:
4592 		cmn_err(CE_WARN, "dca_verify_atomic: unexpected mech type "
4593 		    "0x%llx\n", (unsigned long long)mechanism->cm_type);
4594 		error = CRYPTO_MECHANISM_INVALID;
4595 	}
4596 
4597 	DBG(softc, DENTRY, "dca_verify_atomic: done, err = 0x%x", error);
4598 
4599 	return (error);
4600 }
4601 
4602 /* ARGSUSED */
4603 static int
4604 dca_verify_recover_init(crypto_ctx_t *ctx, crypto_mechanism_t *mechanism,
4605     crypto_key_t *key, crypto_spi_ctx_template_t ctx_template,
4606     crypto_req_handle_t req)
4607 {
4608 	int error = CRYPTO_MECHANISM_INVALID;
4609 	dca_t *softc;
4610 	/* LINTED E_FUNC_SET_NOT_USED */
4611 	int instance;
4612 
4613 	/* extract softc and instance number from context */
4614 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4615 	DBG(softc, DENTRY, "dca_verify_recover_init: started\n");
4616 
4617 	if (ctx_template != NULL)
4618 		return (CRYPTO_ARGUMENTS_BAD);
4619 
4620 	/* check mechanism */
4621 	switch (mechanism->cm_type) {
4622 	case RSA_PKCS_MECH_INFO_TYPE:
4623 	case RSA_X_509_MECH_INFO_TYPE:
4624 		error = dca_rsainit(ctx, mechanism, key, KM_SLEEP);
4625 		break;
4626 	default:
4627 		cmn_err(CE_WARN, "dca_verify_recover_init: unexpected mech type"
4628 		    " 0x%llx\n", (unsigned long long)mechanism->cm_type);
4629 	}
4630 
4631 	DBG(softc, DENTRY, "dca_verify_recover_init: done, err = 0x%x", error);
4632 
4633 	if (error == CRYPTO_SUCCESS)
4634 		dca_enlist2(&softc->dca_ctx_list, ctx->cc_provider_private,
4635 		    &softc->dca_ctx_list_lock);
4636 
4637 	return (error);
4638 }
4639 
4640 static int
4641 dca_verify_recover(crypto_ctx_t *ctx, crypto_data_t *signature,
4642     crypto_data_t *data, crypto_req_handle_t req)
4643 {
4644 	int error = CRYPTO_MECHANISM_INVALID;
4645 	dca_t *softc;
4646 	/* LINTED E_FUNC_SET_NOT_USED */
4647 	int instance;
4648 
4649 	if (!ctx || !ctx->cc_provider || !ctx->cc_provider_private)
4650 		return (CRYPTO_OPERATION_NOT_INITIALIZED);
4651 
4652 	/* extract softc and instance number from context */
4653 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4654 	DBG(softc, DENTRY, "dca_verify_recover: started\n");
4655 
4656 	/* check mechanism */
4657 	switch (DCA_MECH_FROM_CTX(ctx)) {
4658 	case RSA_PKCS_MECH_INFO_TYPE:
4659 	case RSA_X_509_MECH_INFO_TYPE:
4660 		error = dca_rsastart(ctx, signature, data, req, DCA_RSA_VRFYR);
4661 		break;
4662 	default:
4663 		cmn_err(CE_WARN, "dca_verify_recover: unexpected mech type "
4664 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4665 	}
4666 
4667 	DBG(softc, DENTRY, "dca_verify_recover: done, err = 0x%x", error);
4668 
4669 	return (error);
4670 }
4671 
4672 static int
4673 dca_verify_recover_atomic(crypto_provider_handle_t provider,
4674     crypto_session_id_t session_id, crypto_mechanism_t *mechanism,
4675     crypto_key_t *key, crypto_data_t *data, crypto_data_t *signature,
4676     crypto_spi_ctx_template_t ctx_template, crypto_req_handle_t req)
4677 {
4678 	int error = CRYPTO_MECHANISM_INVALID;
4679 	dca_t *softc = (dca_t *)provider;
4680 
4681 	DBG(softc, DENTRY, "dca_verify_recover_atomic: started\n");
4682 
4683 	if (ctx_template != NULL)
4684 		return (CRYPTO_ARGUMENTS_BAD);
4685 
4686 	/* check mechanism */
4687 	switch (mechanism->cm_type) {
4688 	case RSA_PKCS_MECH_INFO_TYPE:
4689 	case RSA_X_509_MECH_INFO_TYPE:
4690 		error = dca_rsaatomic(provider, session_id, mechanism, key,
4691 		    signature, data, KM_SLEEP, req, DCA_RSA_VRFYR);
4692 		break;
4693 	default:
4694 		cmn_err(CE_WARN, "dca_verify_recover_atomic: unexpected mech "
4695 		    "type 0x%llx\n", (unsigned long long)mechanism->cm_type);
4696 		error = CRYPTO_MECHANISM_INVALID;
4697 	}
4698 
4699 	DBG(softc, DENTRY,
4700 	    "dca_verify_recover_atomic: done, err = 0x%x", error);
4701 
4702 	return (error);
4703 }
4704 
4705 /*
4706  * Random number entry points.
4707  */
4708 
4709 /* ARGSUSED */
4710 static int
4711 dca_generate_random(crypto_provider_handle_t provider,
4712     crypto_session_id_t session_id,
4713     uchar_t *buf, size_t len, crypto_req_handle_t req)
4714 {
4715 	int error = CRYPTO_FAILED;
4716 	dca_t *softc = (dca_t *)provider;
4717 	/* LINTED E_FUNC_SET_NOT_USED */
4718 	int instance;
4719 
4720 	instance = ddi_get_instance(softc->dca_dip);
4721 	DBG(softc, DENTRY, "dca_generate_random: started");
4722 
4723 	error = dca_rng(softc, buf, len, req);
4724 
4725 	DBG(softc, DENTRY, "dca_generate_random: done, err = 0x%x", error);
4726 
4727 	return (error);
4728 }
4729 
4730 /*
4731  * Context management entry points.
4732  */
4733 
4734 int
4735 dca_free_context(crypto_ctx_t *ctx)
4736 {
4737 	int error = CRYPTO_SUCCESS;
4738 	dca_t *softc;
4739 	/* LINTED E_FUNC_SET_NOT_USED */
4740 	int instance;
4741 
4742 	/* extract softc and instance number from context */
4743 	DCA_SOFTC_FROM_CTX(ctx, softc, instance);
4744 	DBG(softc, DENTRY, "dca_free_context: entered");
4745 
4746 	if (ctx->cc_provider_private == NULL)
4747 		return (error);
4748 
4749 	dca_rmlist2(ctx->cc_provider_private, &softc->dca_ctx_list_lock);
4750 
4751 	error = dca_free_context_low(ctx);
4752 
4753 	DBG(softc, DENTRY, "dca_free_context: done, err = 0x%x", error);
4754 
4755 	return (error);
4756 }
4757 
4758 static int
4759 dca_free_context_low(crypto_ctx_t *ctx)
4760 {
4761 	int error = CRYPTO_SUCCESS;
4762 
4763 	/* check mechanism */
4764 	switch (DCA_MECH_FROM_CTX(ctx)) {
4765 	case DES_CBC_MECH_INFO_TYPE:
4766 	case DES3_CBC_MECH_INFO_TYPE:
4767 		dca_3desctxfree(ctx);
4768 		break;
4769 	case RSA_PKCS_MECH_INFO_TYPE:
4770 	case RSA_X_509_MECH_INFO_TYPE:
4771 		dca_rsactxfree(ctx);
4772 		break;
4773 	case DSA_MECH_INFO_TYPE:
4774 		dca_dsactxfree(ctx);
4775 		break;
4776 	default:
4777 		/* Should never reach here */
4778 		cmn_err(CE_WARN, "dca_free_context_low: unexpected mech type "
4779 		    "0x%llx\n", (unsigned long long)DCA_MECH_FROM_CTX(ctx));
4780 		error = CRYPTO_MECHANISM_INVALID;
4781 	}
4782 
4783 	return (error);
4784 }
4785 
4786 
4787 /* Free any unfreed private context. It is called in detach. */
4788 static void
4789 dca_free_context_list(dca_t *dca)
4790 {
4791 	dca_listnode_t	*node;
4792 	crypto_ctx_t	ctx;
4793 
4794 	(void) memset(&ctx, 0, sizeof (ctx));
4795 	ctx.cc_provider = dca;
4796 
4797 	while ((node = dca_delist2(&dca->dca_ctx_list,
4798 	    &dca->dca_ctx_list_lock)) != NULL) {
4799 		ctx.cc_provider_private = node;
4800 		(void) dca_free_context_low(&ctx);
4801 	}
4802 }
4803 
4804 static int
4805 ext_info_sym(crypto_provider_handle_t prov,
4806     crypto_provider_ext_info_t *ext_info, crypto_req_handle_t cfreq)
4807 {
4808 	return (ext_info_base(prov, ext_info, cfreq, IDENT_SYM));
4809 }
4810 
4811 static int
4812 ext_info_asym(crypto_provider_handle_t prov,
4813     crypto_provider_ext_info_t *ext_info, crypto_req_handle_t cfreq)
4814 {
4815 	int rv;
4816 
4817 	rv = ext_info_base(prov, ext_info, cfreq, IDENT_ASYM);
4818 	/* The asymmetric cipher slot supports random */
4819 	ext_info->ei_flags |= CRYPTO_EXTF_RNG;
4820 
4821 	return (rv);
4822 }
4823 
4824 /* ARGSUSED */
4825 static int
4826 ext_info_base(crypto_provider_handle_t prov,
4827     crypto_provider_ext_info_t *ext_info, crypto_req_handle_t cfreq, char *id)
4828 {
4829 	dca_t   *dca = (dca_t *)prov;
4830 	int len;
4831 
4832 	/* Label */
4833 	(void) sprintf((char *)ext_info->ei_label, "%s/%d %s",
4834 	    ddi_driver_name(dca->dca_dip), ddi_get_instance(dca->dca_dip), id);
4835 	len = strlen((char *)ext_info->ei_label);
4836 	(void) memset(ext_info->ei_label + len, ' ',
4837 	    CRYPTO_EXT_SIZE_LABEL - len);
4838 
4839 	/* Manufacturer ID */
4840 	(void) sprintf((char *)ext_info->ei_manufacturerID, "%s",
4841 	    DCA_MANUFACTURER_ID);
4842 	len = strlen((char *)ext_info->ei_manufacturerID);
4843 	(void) memset(ext_info->ei_manufacturerID + len, ' ',
4844 	    CRYPTO_EXT_SIZE_MANUF - len);
4845 
4846 	/* Model */
4847 	(void) sprintf((char *)ext_info->ei_model, dca->dca_model);
4848 
4849 	DBG(dca, DWARN, "kCF MODEL: %s", (char *)ext_info->ei_model);
4850 
4851 	len = strlen((char *)ext_info->ei_model);
4852 	(void) memset(ext_info->ei_model + len, ' ',
4853 	    CRYPTO_EXT_SIZE_MODEL - len);
4854 
4855 	/* Serial Number. Blank for Deimos */
4856 	(void) memset(ext_info->ei_serial_number, ' ', CRYPTO_EXT_SIZE_SERIAL);
4857 
4858 	ext_info->ei_flags = CRYPTO_EXTF_WRITE_PROTECTED;
4859 
4860 	ext_info->ei_max_session_count = CRYPTO_UNAVAILABLE_INFO;
4861 	ext_info->ei_max_pin_len = CRYPTO_UNAVAILABLE_INFO;
4862 	ext_info->ei_min_pin_len = CRYPTO_UNAVAILABLE_INFO;
4863 	ext_info->ei_total_public_memory = CRYPTO_UNAVAILABLE_INFO;
4864 	ext_info->ei_free_public_memory = CRYPTO_UNAVAILABLE_INFO;
4865 	ext_info->ei_total_private_memory = CRYPTO_UNAVAILABLE_INFO;
4866 	ext_info->ei_free_private_memory = CRYPTO_UNAVAILABLE_INFO;
4867 	ext_info->ei_hardware_version.cv_major = 0;
4868 	ext_info->ei_hardware_version.cv_minor = 0;
4869 	ext_info->ei_firmware_version.cv_major = 0;
4870 	ext_info->ei_firmware_version.cv_minor = 0;
4871 
4872 	/* Time. No need to be supplied for token without a clock */
4873 	ext_info->ei_time[0] = '\000';
4874 
4875 	return (CRYPTO_SUCCESS);
4876 }
4877 
4878 static void
4879 dca_fma_init(dca_t *dca)
4880 {
4881 	ddi_iblock_cookie_t fm_ibc;
4882 	int fm_capabilities = DDI_FM_EREPORT_CAPABLE |
4883 	    DDI_FM_ACCCHK_CAPABLE | DDI_FM_DMACHK_CAPABLE |
4884 	    DDI_FM_ERRCB_CAPABLE;
4885 
4886 	/* Read FMA capabilities from dca.conf file (if present) */
4887 	dca->fm_capabilities = ddi_getprop(DDI_DEV_T_ANY, dca->dca_dip,
4888 	    DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS, "fm-capable",
4889 	    fm_capabilities);
4890 
4891 	DBG(dca, DWARN, "dca->fm_capabilities = 0x%x", dca->fm_capabilities);
4892 
4893 	/* Only register with IO Fault Services if we have some capability */
4894 	if (dca->fm_capabilities) {
4895 		dca_regsattr.devacc_attr_access = DDI_FLAGERR_ACC;
4896 		dca_devattr.devacc_attr_access = DDI_FLAGERR_ACC;
4897 		dca_dmaattr.dma_attr_flags = DDI_DMA_FLAGERR;
4898 
4899 		/* Register capabilities with IO Fault Services */
4900 		ddi_fm_init(dca->dca_dip, &dca->fm_capabilities, &fm_ibc);
4901 		DBG(dca, DWARN, "fm_capable() =  0x%x",
4902 		    ddi_fm_capable(dca->dca_dip));
4903 
4904 		/*
4905 		 * Initialize pci ereport capabilities if ereport capable
4906 		 */
4907 		if (DDI_FM_EREPORT_CAP(dca->fm_capabilities) ||
4908 		    DDI_FM_ERRCB_CAP(dca->fm_capabilities))
4909 			pci_ereport_setup(dca->dca_dip);
4910 
4911 		/*
4912 		 * Initialize callback mutex and register error callback if
4913 		 * error callback capable.
4914 		 */
4915 		if (DDI_FM_ERRCB_CAP(dca->fm_capabilities)) {
4916 			ddi_fm_handler_register(dca->dca_dip, dca_fm_error_cb,
4917 			    (void *)dca);
4918 		}
4919 	} else {
4920 		/*
4921 		 * These fields have to be cleared of FMA if there are no
4922 		 * FMA capabilities at runtime.
4923 		 */
4924 		dca_regsattr.devacc_attr_access = DDI_DEFAULT_ACC;
4925 		dca_devattr.devacc_attr_access = DDI_DEFAULT_ACC;
4926 		dca_dmaattr.dma_attr_flags = 0;
4927 	}
4928 }
4929 
4930 
4931 static void
4932 dca_fma_fini(dca_t *dca)
4933 {
4934 	/* Only unregister FMA capabilities if we registered some */
4935 	if (dca->fm_capabilities) {
4936 
4937 		/*
4938 		 * Release any resources allocated by pci_ereport_setup()
4939 		 */
4940 		if (DDI_FM_EREPORT_CAP(dca->fm_capabilities) ||
4941 		    DDI_FM_ERRCB_CAP(dca->fm_capabilities)) {
4942 			pci_ereport_teardown(dca->dca_dip);
4943 		}
4944 
4945 		/*
4946 		 * Free callback mutex and un-register error callback if
4947 		 * error callback capable.
4948 		 */
4949 		if (DDI_FM_ERRCB_CAP(dca->fm_capabilities)) {
4950 			ddi_fm_handler_unregister(dca->dca_dip);
4951 		}
4952 
4953 		/* Unregister from IO Fault Services */
4954 		ddi_fm_fini(dca->dca_dip);
4955 		DBG(dca, DWARN, "fm_capable() = 0x%x",
4956 		    ddi_fm_capable(dca->dca_dip));
4957 	}
4958 }
4959 
4960 
4961 /*
4962  * The IO fault service error handling callback function
4963  */
4964 /*ARGSUSED*/
4965 static int
4966 dca_fm_error_cb(dev_info_t *dip, ddi_fm_error_t *err, const void *impl_data)
4967 {
4968 	dca_t		*dca = (dca_t *)impl_data;
4969 
4970 	pci_ereport_post(dip, err, NULL);
4971 	if (err->fme_status == DDI_FM_FATAL) {
4972 		dca_failure(dca, DDI_DATAPATH_FAULT,
4973 		    DCA_FM_ECLASS_NONE, dca_ena(0), CRYPTO_DEVICE_ERROR,
4974 		    "fault PCI in FMA callback.");
4975 	}
4976 	return (err->fme_status);
4977 }
4978 
4979 
4980 static int
4981 dca_check_acc_handle(dca_t *dca, ddi_acc_handle_t handle,
4982     dca_fma_eclass_t eclass_index)
4983 {
4984 	ddi_fm_error_t	de;
4985 	int		version = 0;
4986 
4987 	ddi_fm_acc_err_get(handle, &de, version);
4988 	if (de.fme_status != DDI_FM_OK) {
4989 		dca_failure(dca, DDI_DATAPATH_FAULT,
4990 		    eclass_index, fm_ena_increment(de.fme_ena),
4991 		    CRYPTO_DEVICE_ERROR, "");
4992 		return (DDI_FAILURE);
4993 	}
4994 
4995 	return (DDI_SUCCESS);
4996 }
4997 
4998 int
4999 dca_check_dma_handle(dca_t *dca, ddi_dma_handle_t handle,
5000     dca_fma_eclass_t eclass_index)
5001 {
5002 	ddi_fm_error_t	de;
5003 	int		version = 0;
5004 
5005 	ddi_fm_dma_err_get(handle, &de, version);
5006 	if (de.fme_status != DDI_FM_OK) {
5007 		dca_failure(dca, DDI_DATAPATH_FAULT,
5008 		    eclass_index, fm_ena_increment(de.fme_ena),
5009 		    CRYPTO_DEVICE_ERROR, "");
5010 		return (DDI_FAILURE);
5011 	}
5012 	return (DDI_SUCCESS);
5013 }
5014 
5015 static uint64_t
5016 dca_ena(uint64_t ena)
5017 {
5018 	if (ena == 0)
5019 		ena = fm_ena_generate(0, FM_ENA_FMT1);
5020 	else
5021 		ena = fm_ena_increment(ena);
5022 	return (ena);
5023 }
5024 
5025 static char *
5026 dca_fma_eclass_string(char *model, dca_fma_eclass_t index)
5027 {
5028 	if (strstr(model, "500"))
5029 		return (dca_fma_eclass_sca500[index]);
5030 	else
5031 		return (dca_fma_eclass_sca1000[index]);
5032 }
5033