1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright (c) 2001, 2010, Oracle and/or its affiliates. All rights reserved. 23 */ 24 25 #include <sys/param.h> 26 #include <sys/types.h> 27 #include <sys/stream.h> 28 #include <sys/strsubr.h> 29 #include <sys/strsun.h> 30 #include <sys/stropts.h> 31 #include <sys/zone.h> 32 #include <sys/vnode.h> 33 #include <sys/sysmacros.h> 34 #define _SUN_TPI_VERSION 2 35 #include <sys/tihdr.h> 36 #include <sys/ddi.h> 37 #include <sys/sunddi.h> 38 #include <sys/mkdev.h> 39 #include <sys/debug.h> 40 #include <sys/kmem.h> 41 #include <sys/cmn_err.h> 42 #include <sys/suntpi.h> 43 #include <sys/policy.h> 44 #include <sys/dls.h> 45 46 #include <sys/socket.h> 47 #include <netinet/in.h> 48 #include <net/pfkeyv2.h> 49 #include <net/pfpolicy.h> 50 51 #include <inet/common.h> 52 #include <netinet/ip6.h> 53 #include <inet/ip.h> 54 #include <inet/ip6.h> 55 #include <inet/mi.h> 56 #include <inet/proto_set.h> 57 #include <inet/nd.h> 58 #include <inet/ip_if.h> 59 #include <inet/optcom.h> 60 #include <inet/ipsec_impl.h> 61 #include <inet/spdsock.h> 62 #include <inet/sadb.h> 63 #include <inet/iptun.h> 64 #include <inet/iptun/iptun_impl.h> 65 66 #include <sys/isa_defs.h> 67 68 #include <c2/audit.h> 69 70 /* 71 * This is a transport provider for the PF_POLICY IPsec policy 72 * management socket, which provides a management interface into the 73 * SPD, allowing policy rules to be added, deleted, and queried. 74 * 75 * This effectively replaces the old private SIOC*IPSECONFIG ioctls 76 * with an extensible interface which will hopefully be public some 77 * day. 78 * 79 * See <net/pfpolicy.h> for more details on the protocol. 80 * 81 * We link against drv/ip and call directly into it to manipulate the 82 * SPD; see ipsec_impl.h for the policy data structures and spd.c for 83 * the code which maintains them. 84 * 85 * The MT model of this is QPAIR with the addition of some explicit 86 * locking to protect system-wide policy data structures. 87 */ 88 89 static vmem_t *spdsock_vmem; /* for minor numbers. */ 90 91 #define ALIGNED64(x) IS_P2ALIGNED((x), sizeof (uint64_t)) 92 93 /* Default structure copied into T_INFO_ACK messages (from rts.c...) */ 94 static struct T_info_ack spdsock_g_t_info_ack = { 95 T_INFO_ACK, 96 T_INFINITE, /* TSDU_size. Maximum size messages. */ 97 T_INVALID, /* ETSDU_size. No expedited data. */ 98 T_INVALID, /* CDATA_size. No connect data. */ 99 T_INVALID, /* DDATA_size. No disconnect data. */ 100 0, /* ADDR_size. */ 101 0, /* OPT_size. No user-settable options */ 102 64 * 1024, /* TIDU_size. spdsock allows maximum size messages. */ 103 T_COTS, /* SERV_type. spdsock supports connection oriented. */ 104 TS_UNBND, /* CURRENT_state. This is set from spdsock_state. */ 105 (XPG4_1) /* Provider flags */ 106 }; 107 108 /* Named Dispatch Parameter Management Structure */ 109 typedef struct spdsockparam_s { 110 uint_t spdsock_param_min; 111 uint_t spdsock_param_max; 112 uint_t spdsock_param_value; 113 char *spdsock_param_name; 114 } spdsockparam_t; 115 116 /* 117 * Table of NDD variables supported by spdsock. These are loaded into 118 * spdsock_g_nd in spdsock_init_nd. 119 * All of these are alterable, within the min/max values given, at run time. 120 */ 121 static spdsockparam_t lcl_param_arr[] = { 122 /* min max value name */ 123 { 4096, 65536, 8192, "spdsock_xmit_hiwat"}, 124 { 0, 65536, 1024, "spdsock_xmit_lowat"}, 125 { 4096, 65536, 8192, "spdsock_recv_hiwat"}, 126 { 65536, 1024*1024*1024, 256*1024, "spdsock_max_buf"}, 127 { 0, 3, 0, "spdsock_debug"}, 128 }; 129 #define spds_xmit_hiwat spds_params[0].spdsock_param_value 130 #define spds_xmit_lowat spds_params[1].spdsock_param_value 131 #define spds_recv_hiwat spds_params[2].spdsock_param_value 132 #define spds_max_buf spds_params[3].spdsock_param_value 133 #define spds_debug spds_params[4].spdsock_param_value 134 135 #define ss0dbg(a) printf a 136 /* NOTE: != 0 instead of > 0 so lint doesn't complain. */ 137 #define ss1dbg(spds, a) if (spds->spds_debug != 0) printf a 138 #define ss2dbg(spds, a) if (spds->spds_debug > 1) printf a 139 #define ss3dbg(spds, a) if (spds->spds_debug > 2) printf a 140 141 #define RESET_SPDSOCK_DUMP_POLHEAD(ss, iph) { \ 142 ASSERT(RW_READ_HELD(&(iph)->iph_lock)); \ 143 (ss)->spdsock_dump_head = (iph); \ 144 (ss)->spdsock_dump_gen = (iph)->iph_gen; \ 145 (ss)->spdsock_dump_cur_type = 0; \ 146 (ss)->spdsock_dump_cur_af = IPSEC_AF_V4; \ 147 (ss)->spdsock_dump_cur_rule = NULL; \ 148 (ss)->spdsock_dump_count = 0; \ 149 (ss)->spdsock_dump_cur_chain = 0; \ 150 } 151 152 static int spdsock_close(queue_t *); 153 static int spdsock_open(queue_t *, dev_t *, int, int, cred_t *); 154 static void spdsock_wput(queue_t *, mblk_t *); 155 static void spdsock_wsrv(queue_t *); 156 static void spdsock_rsrv(queue_t *); 157 static void *spdsock_stack_init(netstackid_t stackid, netstack_t *ns); 158 static void spdsock_stack_shutdown(netstackid_t stackid, void *arg); 159 static void spdsock_stack_fini(netstackid_t stackid, void *arg); 160 static void spdsock_loadcheck(void *); 161 static void spdsock_merge_algs(spd_stack_t *); 162 static void spdsock_flush_one(ipsec_policy_head_t *, netstack_t *); 163 static mblk_t *spdsock_dump_next_record(spdsock_t *); 164 static void update_iptun_policy(ipsec_tun_pol_t *); 165 166 static struct module_info info = { 167 5138, "spdsock", 1, INFPSZ, 512, 128 168 }; 169 170 static struct qinit rinit = { 171 NULL, (pfi_t)spdsock_rsrv, spdsock_open, spdsock_close, 172 NULL, &info 173 }; 174 175 static struct qinit winit = { 176 (pfi_t)spdsock_wput, (pfi_t)spdsock_wsrv, NULL, NULL, NULL, &info 177 }; 178 179 struct streamtab spdsockinfo = { 180 &rinit, &winit 181 }; 182 183 /* mapping from alg type to protocol number, as per RFC 2407 */ 184 static const uint_t algproto[] = { 185 PROTO_IPSEC_AH, 186 PROTO_IPSEC_ESP, 187 }; 188 189 #define NALGPROTOS (sizeof (algproto) / sizeof (algproto[0])) 190 191 /* mapping from kernel exec mode to spdsock exec mode */ 192 static const uint_t execmodes[] = { 193 SPD_ALG_EXEC_MODE_SYNC, 194 SPD_ALG_EXEC_MODE_ASYNC 195 }; 196 197 #define NEXECMODES (sizeof (execmodes) / sizeof (execmodes[0])) 198 199 #define ALL_ACTIVE_POLHEADS ((ipsec_policy_head_t *)-1) 200 #define ALL_INACTIVE_POLHEADS ((ipsec_policy_head_t *)-2) 201 202 #define ITP_NAME(itp) (itp != NULL ? itp->itp_name : NULL) 203 204 /* ARGSUSED */ 205 static int 206 spdsock_param_get(q, mp, cp, cr) 207 queue_t *q; 208 mblk_t *mp; 209 caddr_t cp; 210 cred_t *cr; 211 { 212 spdsockparam_t *spdsockpa = (spdsockparam_t *)cp; 213 uint_t value; 214 spdsock_t *ss = (spdsock_t *)q->q_ptr; 215 spd_stack_t *spds = ss->spdsock_spds; 216 217 mutex_enter(&spds->spds_param_lock); 218 value = spdsockpa->spdsock_param_value; 219 mutex_exit(&spds->spds_param_lock); 220 221 (void) mi_mpprintf(mp, "%u", value); 222 return (0); 223 } 224 225 /* This routine sets an NDD variable in a spdsockparam_t structure. */ 226 /* ARGSUSED */ 227 static int 228 spdsock_param_set(q, mp, value, cp, cr) 229 queue_t *q; 230 mblk_t *mp; 231 char *value; 232 caddr_t cp; 233 cred_t *cr; 234 { 235 ulong_t new_value; 236 spdsockparam_t *spdsockpa = (spdsockparam_t *)cp; 237 spdsock_t *ss = (spdsock_t *)q->q_ptr; 238 spd_stack_t *spds = ss->spdsock_spds; 239 240 /* Convert the value from a string into a long integer. */ 241 if (ddi_strtoul(value, NULL, 10, &new_value) != 0) 242 return (EINVAL); 243 244 mutex_enter(&spds->spds_param_lock); 245 /* 246 * Fail the request if the new value does not lie within the 247 * required bounds. 248 */ 249 if (new_value < spdsockpa->spdsock_param_min || 250 new_value > spdsockpa->spdsock_param_max) { 251 mutex_exit(&spds->spds_param_lock); 252 return (EINVAL); 253 } 254 255 /* Set the new value */ 256 spdsockpa->spdsock_param_value = new_value; 257 mutex_exit(&spds->spds_param_lock); 258 259 return (0); 260 } 261 262 /* 263 * Initialize at module load time 264 */ 265 boolean_t 266 spdsock_ddi_init(void) 267 { 268 spdsock_max_optsize = optcom_max_optsize( 269 spdsock_opt_obj.odb_opt_des_arr, spdsock_opt_obj.odb_opt_arr_cnt); 270 271 spdsock_vmem = vmem_create("spdsock", (void *)1, MAXMIN, 1, 272 NULL, NULL, NULL, 1, VM_SLEEP | VMC_IDENTIFIER); 273 274 /* 275 * We want to be informed each time a stack is created or 276 * destroyed in the kernel, so we can maintain the 277 * set of spd_stack_t's. 278 */ 279 netstack_register(NS_SPDSOCK, spdsock_stack_init, 280 spdsock_stack_shutdown, spdsock_stack_fini); 281 282 return (B_TRUE); 283 } 284 285 /* 286 * Walk through the param array specified registering each element with the 287 * named dispatch handler. 288 */ 289 static boolean_t 290 spdsock_param_register(IDP *ndp, spdsockparam_t *ssp, int cnt) 291 { 292 for (; cnt-- > 0; ssp++) { 293 if (ssp->spdsock_param_name != NULL && 294 ssp->spdsock_param_name[0]) { 295 if (!nd_load(ndp, 296 ssp->spdsock_param_name, 297 spdsock_param_get, spdsock_param_set, 298 (caddr_t)ssp)) { 299 nd_free(ndp); 300 return (B_FALSE); 301 } 302 } 303 } 304 return (B_TRUE); 305 } 306 307 /* 308 * Initialize for each stack instance 309 */ 310 /* ARGSUSED */ 311 static void * 312 spdsock_stack_init(netstackid_t stackid, netstack_t *ns) 313 { 314 spd_stack_t *spds; 315 spdsockparam_t *ssp; 316 317 spds = (spd_stack_t *)kmem_zalloc(sizeof (*spds), KM_SLEEP); 318 spds->spds_netstack = ns; 319 320 ASSERT(spds->spds_g_nd == NULL); 321 322 ssp = (spdsockparam_t *)kmem_alloc(sizeof (lcl_param_arr), KM_SLEEP); 323 spds->spds_params = ssp; 324 bcopy(lcl_param_arr, ssp, sizeof (lcl_param_arr)); 325 326 (void) spdsock_param_register(&spds->spds_g_nd, ssp, 327 A_CNT(lcl_param_arr)); 328 329 mutex_init(&spds->spds_param_lock, NULL, MUTEX_DEFAULT, NULL); 330 mutex_init(&spds->spds_alg_lock, NULL, MUTEX_DEFAULT, NULL); 331 332 return (spds); 333 } 334 335 void 336 spdsock_ddi_destroy(void) 337 { 338 vmem_destroy(spdsock_vmem); 339 340 netstack_unregister(NS_SPDSOCK); 341 } 342 343 /* 344 * Do pre-removal cleanup. 345 */ 346 /* ARGSUSED */ 347 static void 348 spdsock_stack_shutdown(netstackid_t stackid, void *arg) 349 { 350 spd_stack_t *spds = (spd_stack_t *)arg; 351 352 if (spds->spds_mp_algs != NULL) { 353 freemsg(spds->spds_mp_algs); 354 spds->spds_mp_algs = NULL; 355 } 356 } 357 358 /* ARGSUSED */ 359 static void 360 spdsock_stack_fini(netstackid_t stackid, void *arg) 361 { 362 spd_stack_t *spds = (spd_stack_t *)arg; 363 364 ASSERT(spds->spds_mp_algs == NULL); 365 mutex_destroy(&spds->spds_param_lock); 366 mutex_destroy(&spds->spds_alg_lock); 367 nd_free(&spds->spds_g_nd); 368 kmem_free(spds->spds_params, sizeof (lcl_param_arr)); 369 spds->spds_params = NULL; 370 371 kmem_free(spds, sizeof (*spds)); 372 } 373 374 /* 375 * NOTE: large quantities of this should be shared with keysock. 376 * Would be nice to combine some of this into a common module, but 377 * not possible given time pressures. 378 */ 379 380 /* 381 * High-level reality checking of extensions. 382 */ 383 /* ARGSUSED */ /* XXX */ 384 static boolean_t 385 ext_check(spd_ext_t *ext) 386 { 387 spd_if_t *tunname = (spd_if_t *)ext; 388 int i; 389 char *idstr; 390 391 if (ext->spd_ext_type == SPD_EXT_TUN_NAME) { 392 /* (NOTE: Modified from SADB_EXT_IDENTITY..) */ 393 394 /* 395 * Make sure the strings in these identities are 396 * null-terminated. Let's "proactively" null-terminate the 397 * string at the last byte if it's not terminated sooner. 398 */ 399 i = SPD_64TO8(tunname->spd_if_len) - sizeof (spd_if_t); 400 idstr = (char *)(tunname + 1); 401 while (*idstr != '\0' && i > 0) { 402 i--; 403 idstr++; 404 } 405 if (i == 0) { 406 /* 407 * I.e., if the bozo user didn't NULL-terminate the 408 * string... 409 */ 410 idstr--; 411 *idstr = '\0'; 412 } 413 } 414 return (B_TRUE); /* For now... */ 415 } 416 417 418 419 /* Return values for spdsock_get_ext(). */ 420 #define KGE_OK 0 421 #define KGE_DUP 1 422 #define KGE_UNK 2 423 #define KGE_LEN 3 424 #define KGE_CHK 4 425 426 /* 427 * Parse basic extension headers and return in the passed-in pointer vector. 428 * Return values include: 429 * 430 * KGE_OK Everything's nice and parsed out. 431 * If there are no extensions, place NULL in extv[0]. 432 * KGE_DUP There is a duplicate extension. 433 * First instance in appropriate bin. First duplicate in 434 * extv[0]. 435 * KGE_UNK Unknown extension type encountered. extv[0] contains 436 * unknown header. 437 * KGE_LEN Extension length error. 438 * KGE_CHK High-level reality check failed on specific extension. 439 * 440 * My apologies for some of the pointer arithmetic in here. I'm thinking 441 * like an assembly programmer, yet trying to make the compiler happy. 442 */ 443 static int 444 spdsock_get_ext(spd_ext_t *extv[], spd_msg_t *basehdr, uint_t msgsize) 445 { 446 bzero(extv, sizeof (spd_ext_t *) * (SPD_EXT_MAX + 1)); 447 448 /* Use extv[0] as the "current working pointer". */ 449 450 extv[0] = (spd_ext_t *)(basehdr + 1); 451 452 while (extv[0] < (spd_ext_t *)(((uint8_t *)basehdr) + msgsize)) { 453 /* Check for unknown headers. */ 454 if (extv[0]->spd_ext_type == 0 || 455 extv[0]->spd_ext_type > SPD_EXT_MAX) 456 return (KGE_UNK); 457 458 /* 459 * Check length. Use uint64_t because extlen is in units 460 * of 64-bit words. If length goes beyond the msgsize, 461 * return an error. (Zero length also qualifies here.) 462 */ 463 if (extv[0]->spd_ext_len == 0 || 464 (void *)((uint64_t *)extv[0] + extv[0]->spd_ext_len) > 465 (void *)((uint8_t *)basehdr + msgsize)) 466 return (KGE_LEN); 467 468 /* Check for redundant headers. */ 469 if (extv[extv[0]->spd_ext_type] != NULL) 470 return (KGE_DUP); 471 472 /* 473 * Reality check the extension if possible at the spdsock 474 * level. 475 */ 476 if (!ext_check(extv[0])) 477 return (KGE_CHK); 478 479 /* If I make it here, assign the appropriate bin. */ 480 extv[extv[0]->spd_ext_type] = extv[0]; 481 482 /* Advance pointer (See above for uint64_t ptr reasoning.) */ 483 extv[0] = (spd_ext_t *) 484 ((uint64_t *)extv[0] + extv[0]->spd_ext_len); 485 } 486 487 /* Everything's cool. */ 488 489 /* 490 * If extv[0] == NULL, then there are no extension headers in this 491 * message. Ensure that this is the case. 492 */ 493 if (extv[0] == (spd_ext_t *)(basehdr + 1)) 494 extv[0] = NULL; 495 496 return (KGE_OK); 497 } 498 499 static const int bad_ext_diag[] = { 500 SPD_DIAGNOSTIC_MALFORMED_LCLPORT, 501 SPD_DIAGNOSTIC_MALFORMED_REMPORT, 502 SPD_DIAGNOSTIC_MALFORMED_PROTO, 503 SPD_DIAGNOSTIC_MALFORMED_LCLADDR, 504 SPD_DIAGNOSTIC_MALFORMED_REMADDR, 505 SPD_DIAGNOSTIC_MALFORMED_ACTION, 506 SPD_DIAGNOSTIC_MALFORMED_RULE, 507 SPD_DIAGNOSTIC_MALFORMED_RULESET, 508 SPD_DIAGNOSTIC_MALFORMED_ICMP_TYPECODE 509 }; 510 511 static const int dup_ext_diag[] = { 512 SPD_DIAGNOSTIC_DUPLICATE_LCLPORT, 513 SPD_DIAGNOSTIC_DUPLICATE_REMPORT, 514 SPD_DIAGNOSTIC_DUPLICATE_PROTO, 515 SPD_DIAGNOSTIC_DUPLICATE_LCLADDR, 516 SPD_DIAGNOSTIC_DUPLICATE_REMADDR, 517 SPD_DIAGNOSTIC_DUPLICATE_ACTION, 518 SPD_DIAGNOSTIC_DUPLICATE_RULE, 519 SPD_DIAGNOSTIC_DUPLICATE_RULESET, 520 SPD_DIAGNOSTIC_DUPLICATE_ICMP_TYPECODE 521 }; 522 523 /* 524 * Transmit a PF_POLICY error message to the instance either pointed to 525 * by ks, the instance with serial number serial, or more, depending. 526 * 527 * The faulty message (or a reasonable facsimile thereof) is in mp. 528 * This function will free mp or recycle it for delivery, thereby causing 529 * the stream head to free it. 530 */ 531 static void 532 spdsock_error(queue_t *q, mblk_t *mp, int error, int diagnostic) 533 { 534 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 535 536 ASSERT(mp->b_datap->db_type == M_DATA); 537 538 if (spmsg->spd_msg_type < SPD_MIN || 539 spmsg->spd_msg_type > SPD_MAX) 540 spmsg->spd_msg_type = SPD_RESERVED; 541 542 /* 543 * Strip out extension headers. 544 */ 545 ASSERT(mp->b_rptr + sizeof (*spmsg) <= mp->b_datap->db_lim); 546 mp->b_wptr = mp->b_rptr + sizeof (*spmsg); 547 spmsg->spd_msg_len = SPD_8TO64(sizeof (spd_msg_t)); 548 spmsg->spd_msg_errno = (uint8_t)error; 549 spmsg->spd_msg_diagnostic = (uint16_t)diagnostic; 550 551 qreply(q, mp); 552 } 553 554 static void 555 spdsock_diag(queue_t *q, mblk_t *mp, int diagnostic) 556 { 557 spdsock_error(q, mp, EINVAL, diagnostic); 558 } 559 560 static void 561 spd_echo(queue_t *q, mblk_t *mp) 562 { 563 qreply(q, mp); 564 } 565 566 /* 567 * Do NOT consume a reference to itp. 568 */ 569 /*ARGSUSED*/ 570 static void 571 spdsock_flush_node(ipsec_tun_pol_t *itp, void *cookie, netstack_t *ns) 572 { 573 boolean_t active = (boolean_t)cookie; 574 ipsec_policy_head_t *iph; 575 576 iph = active ? itp->itp_policy : itp->itp_inactive; 577 IPPH_REFHOLD(iph); 578 mutex_enter(&itp->itp_lock); 579 spdsock_flush_one(iph, ns); /* Releases iph refhold. */ 580 if (active) 581 itp->itp_flags &= ~ITPF_PFLAGS; 582 else 583 itp->itp_flags &= ~ITPF_IFLAGS; 584 mutex_exit(&itp->itp_lock); 585 /* SPD_FLUSH is worth a tunnel MTU check. */ 586 update_iptun_policy(itp); 587 } 588 589 /* 590 * Clear out one polhead. 591 */ 592 static void 593 spdsock_flush_one(ipsec_policy_head_t *iph, netstack_t *ns) 594 { 595 rw_enter(&iph->iph_lock, RW_WRITER); 596 ipsec_polhead_flush(iph, ns); 597 rw_exit(&iph->iph_lock); 598 IPPH_REFRELE(iph, ns); 599 } 600 601 static void 602 spdsock_flush(queue_t *q, ipsec_policy_head_t *iph, ipsec_tun_pol_t *itp, 603 mblk_t *mp) 604 { 605 boolean_t active; 606 spdsock_t *ss = (spdsock_t *)q->q_ptr; 607 netstack_t *ns = ss->spdsock_spds->spds_netstack; 608 uint32_t auditing = AU_AUDITING(); 609 610 if (iph != ALL_ACTIVE_POLHEADS && iph != ALL_INACTIVE_POLHEADS) { 611 spdsock_flush_one(iph, ns); 612 if (auditing) { 613 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 614 cred_t *cr; 615 pid_t cpid; 616 617 cr = msg_getcred(mp, &cpid); 618 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 619 audit_pf_policy(SPD_FLUSH, cr, ns, 620 ITP_NAME(itp), active, 0, cpid); 621 } 622 } else { 623 active = (iph == ALL_ACTIVE_POLHEADS); 624 625 /* First flush the global policy. */ 626 spdsock_flush_one(active ? ipsec_system_policy(ns) : 627 ipsec_inactive_policy(ns), ns); 628 if (auditing) { 629 cred_t *cr; 630 pid_t cpid; 631 632 cr = msg_getcred(mp, &cpid); 633 audit_pf_policy(SPD_FLUSH, cr, ns, NULL, 634 active, 0, cpid); 635 } 636 /* Then flush every tunnel's appropriate one. */ 637 itp_walk(spdsock_flush_node, (void *)active, ns); 638 if (auditing) { 639 cred_t *cr; 640 pid_t cpid; 641 642 cr = msg_getcred(mp, &cpid); 643 audit_pf_policy(SPD_FLUSH, cr, ns, 644 "all tunnels", active, 0, cpid); 645 } 646 } 647 648 spd_echo(q, mp); 649 } 650 651 static boolean_t 652 spdsock_ext_to_sel(spd_ext_t **extv, ipsec_selkey_t *sel, int *diag) 653 { 654 bzero(sel, sizeof (*sel)); 655 656 if (extv[SPD_EXT_PROTO] != NULL) { 657 struct spd_proto *pr = 658 (struct spd_proto *)extv[SPD_EXT_PROTO]; 659 sel->ipsl_proto = pr->spd_proto_number; 660 sel->ipsl_valid |= IPSL_PROTOCOL; 661 } 662 if (extv[SPD_EXT_LCLPORT] != NULL) { 663 struct spd_portrange *pr = 664 (struct spd_portrange *)extv[SPD_EXT_LCLPORT]; 665 sel->ipsl_lport = pr->spd_ports_minport; 666 sel->ipsl_valid |= IPSL_LOCAL_PORT; 667 } 668 if (extv[SPD_EXT_REMPORT] != NULL) { 669 struct spd_portrange *pr = 670 (struct spd_portrange *)extv[SPD_EXT_REMPORT]; 671 sel->ipsl_rport = pr->spd_ports_minport; 672 sel->ipsl_valid |= IPSL_REMOTE_PORT; 673 } 674 675 if (extv[SPD_EXT_ICMP_TYPECODE] != NULL) { 676 struct spd_typecode *tc= 677 (struct spd_typecode *)extv[SPD_EXT_ICMP_TYPECODE]; 678 679 sel->ipsl_valid |= IPSL_ICMP_TYPE; 680 sel->ipsl_icmp_type = tc->spd_typecode_type; 681 if (tc->spd_typecode_type_end < tc->spd_typecode_type) 682 sel->ipsl_icmp_type_end = tc->spd_typecode_type; 683 else 684 sel->ipsl_icmp_type_end = tc->spd_typecode_type_end; 685 686 if (tc->spd_typecode_code != 255) { 687 sel->ipsl_valid |= IPSL_ICMP_CODE; 688 sel->ipsl_icmp_code = tc->spd_typecode_code; 689 if (tc->spd_typecode_code_end < tc->spd_typecode_code) 690 sel->ipsl_icmp_code_end = tc->spd_typecode_code; 691 else 692 sel->ipsl_icmp_code_end = 693 tc->spd_typecode_code_end; 694 } 695 } 696 #define ADDR2SEL(sel, extv, field, pfield, extn, bit) \ 697 if ((extv)[(extn)] != NULL) { \ 698 uint_t addrlen; \ 699 struct spd_address *ap = \ 700 (struct spd_address *)((extv)[(extn)]); \ 701 addrlen = (ap->spd_address_af == AF_INET6) ? \ 702 IPV6_ADDR_LEN : IP_ADDR_LEN; \ 703 if (SPD_64TO8(ap->spd_address_len) < \ 704 (addrlen + sizeof (*ap))) { \ 705 *diag = SPD_DIAGNOSTIC_BAD_ADDR_LEN; \ 706 return (B_FALSE); \ 707 } \ 708 bcopy((ap+1), &((sel)->field), addrlen); \ 709 (sel)->pfield = ap->spd_address_prefixlen; \ 710 (sel)->ipsl_valid |= (bit); \ 711 (sel)->ipsl_valid |= (ap->spd_address_af == AF_INET6) ? \ 712 IPSL_IPV6 : IPSL_IPV4; \ 713 } 714 715 ADDR2SEL(sel, extv, ipsl_local, ipsl_local_pfxlen, 716 SPD_EXT_LCLADDR, IPSL_LOCAL_ADDR); 717 ADDR2SEL(sel, extv, ipsl_remote, ipsl_remote_pfxlen, 718 SPD_EXT_REMADDR, IPSL_REMOTE_ADDR); 719 720 if ((sel->ipsl_valid & (IPSL_IPV6|IPSL_IPV4)) == 721 (IPSL_IPV6|IPSL_IPV4)) { 722 *diag = SPD_DIAGNOSTIC_MIXED_AF; 723 return (B_FALSE); 724 } 725 726 #undef ADDR2SEL 727 728 return (B_TRUE); 729 } 730 731 static boolean_t 732 spd_convert_type(uint32_t type, ipsec_act_t *act) 733 { 734 switch (type) { 735 case SPD_ACTTYPE_DROP: 736 act->ipa_type = IPSEC_ACT_DISCARD; 737 return (B_TRUE); 738 739 case SPD_ACTTYPE_PASS: 740 act->ipa_type = IPSEC_ACT_CLEAR; 741 return (B_TRUE); 742 743 case SPD_ACTTYPE_IPSEC: 744 act->ipa_type = IPSEC_ACT_APPLY; 745 return (B_TRUE); 746 } 747 return (B_FALSE); 748 } 749 750 static boolean_t 751 spd_convert_flags(uint32_t flags, ipsec_act_t *act) 752 { 753 /* 754 * Note use of !! for boolean canonicalization. 755 */ 756 act->ipa_apply.ipp_use_ah = !!(flags & SPD_APPLY_AH); 757 act->ipa_apply.ipp_use_esp = !!(flags & SPD_APPLY_ESP); 758 act->ipa_apply.ipp_use_espa = !!(flags & SPD_APPLY_ESPA); 759 act->ipa_apply.ipp_use_se = !!(flags & SPD_APPLY_SE); 760 act->ipa_apply.ipp_use_unique = !!(flags & SPD_APPLY_UNIQUE); 761 return (B_TRUE); 762 } 763 764 static void 765 spdsock_reset_act(ipsec_act_t *act) 766 { 767 bzero(act, sizeof (*act)); 768 act->ipa_apply.ipp_espe_maxbits = IPSEC_MAX_KEYBITS; 769 act->ipa_apply.ipp_espa_maxbits = IPSEC_MAX_KEYBITS; 770 act->ipa_apply.ipp_ah_maxbits = IPSEC_MAX_KEYBITS; 771 } 772 773 /* 774 * Sanity check action against reality, and shrink-wrap key sizes.. 775 */ 776 static boolean_t 777 spdsock_check_action(ipsec_act_t *act, boolean_t tunnel_polhead, int *diag, 778 spd_stack_t *spds) 779 { 780 if (tunnel_polhead && act->ipa_apply.ipp_use_unique) { 781 *diag = SPD_DIAGNOSTIC_ADD_INCON_FLAGS; 782 return (B_FALSE); 783 } 784 if ((act->ipa_type != IPSEC_ACT_APPLY) && 785 (act->ipa_apply.ipp_use_ah || 786 act->ipa_apply.ipp_use_esp || 787 act->ipa_apply.ipp_use_espa || 788 act->ipa_apply.ipp_use_se || 789 act->ipa_apply.ipp_use_unique)) { 790 *diag = SPD_DIAGNOSTIC_ADD_INCON_FLAGS; 791 return (B_FALSE); 792 } 793 if ((act->ipa_type == IPSEC_ACT_APPLY) && 794 !act->ipa_apply.ipp_use_ah && 795 !act->ipa_apply.ipp_use_esp) { 796 *diag = SPD_DIAGNOSTIC_ADD_INCON_FLAGS; 797 return (B_FALSE); 798 } 799 return (ipsec_check_action(act, diag, spds->spds_netstack)); 800 } 801 802 /* 803 * We may be short a few error checks here.. 804 */ 805 static boolean_t 806 spdsock_ext_to_actvec(spd_ext_t **extv, ipsec_act_t **actpp, uint_t *nactp, 807 int *diag, spd_stack_t *spds) 808 { 809 struct spd_ext_actions *sactp = 810 (struct spd_ext_actions *)extv[SPD_EXT_ACTION]; 811 ipsec_act_t act, *actp, *endactp; 812 struct spd_attribute *attrp, *endattrp; 813 uint64_t *endp; 814 int nact; 815 boolean_t tunnel_polhead; 816 817 tunnel_polhead = (extv[SPD_EXT_TUN_NAME] != NULL && 818 (((struct spd_rule *)extv[SPD_EXT_RULE])->spd_rule_flags & 819 SPD_RULE_FLAG_TUNNEL)); 820 821 *actpp = NULL; 822 *nactp = 0; 823 824 if (sactp == NULL) { 825 *diag = SPD_DIAGNOSTIC_NO_ACTION_EXT; 826 return (B_FALSE); 827 } 828 829 /* 830 * Parse the "action" extension and convert into an action chain. 831 */ 832 833 nact = sactp->spd_actions_count; 834 835 endp = (uint64_t *)sactp; 836 endp += sactp->spd_actions_len; 837 endattrp = (struct spd_attribute *)endp; 838 839 actp = kmem_alloc(sizeof (*actp) * nact, KM_NOSLEEP); 840 if (actp == NULL) { 841 *diag = SPD_DIAGNOSTIC_ADD_NO_MEM; 842 return (B_FALSE); 843 } 844 *actpp = actp; 845 *nactp = nact; 846 endactp = actp + nact; 847 848 spdsock_reset_act(&act); 849 attrp = (struct spd_attribute *)(&sactp[1]); 850 851 for (; attrp < endattrp; attrp++) { 852 switch (attrp->spd_attr_tag) { 853 case SPD_ATTR_NOP: 854 break; 855 856 case SPD_ATTR_EMPTY: 857 spdsock_reset_act(&act); 858 break; 859 860 case SPD_ATTR_END: 861 attrp = endattrp; 862 /* FALLTHRU */ 863 case SPD_ATTR_NEXT: 864 if (actp >= endactp) { 865 *diag = SPD_DIAGNOSTIC_ADD_WRONG_ACT_COUNT; 866 goto fail; 867 } 868 if (!spdsock_check_action(&act, tunnel_polhead, 869 diag, spds)) 870 goto fail; 871 *actp++ = act; 872 spdsock_reset_act(&act); 873 break; 874 875 case SPD_ATTR_TYPE: 876 if (!spd_convert_type(attrp->spd_attr_value, &act)) { 877 *diag = SPD_DIAGNOSTIC_ADD_BAD_TYPE; 878 goto fail; 879 } 880 break; 881 882 case SPD_ATTR_FLAGS: 883 if (!tunnel_polhead && extv[SPD_EXT_TUN_NAME] != NULL) { 884 /* 885 * Set "sa unique" for transport-mode 886 * tunnels whether we want to or not. 887 */ 888 attrp->spd_attr_value |= SPD_APPLY_UNIQUE; 889 } 890 if (!spd_convert_flags(attrp->spd_attr_value, &act)) { 891 *diag = SPD_DIAGNOSTIC_ADD_BAD_FLAGS; 892 goto fail; 893 } 894 break; 895 896 case SPD_ATTR_AH_AUTH: 897 if (attrp->spd_attr_value == 0) { 898 *diag = SPD_DIAGNOSTIC_UNSUPP_AH_ALG; 899 goto fail; 900 } 901 act.ipa_apply.ipp_auth_alg = attrp->spd_attr_value; 902 break; 903 904 case SPD_ATTR_ESP_ENCR: 905 if (attrp->spd_attr_value == 0) { 906 *diag = SPD_DIAGNOSTIC_UNSUPP_ESP_ENCR_ALG; 907 goto fail; 908 } 909 act.ipa_apply.ipp_encr_alg = attrp->spd_attr_value; 910 break; 911 912 case SPD_ATTR_ESP_AUTH: 913 if (attrp->spd_attr_value == 0) { 914 *diag = SPD_DIAGNOSTIC_UNSUPP_ESP_AUTH_ALG; 915 goto fail; 916 } 917 act.ipa_apply.ipp_esp_auth_alg = attrp->spd_attr_value; 918 break; 919 920 case SPD_ATTR_ENCR_MINBITS: 921 act.ipa_apply.ipp_espe_minbits = attrp->spd_attr_value; 922 break; 923 924 case SPD_ATTR_ENCR_MAXBITS: 925 act.ipa_apply.ipp_espe_maxbits = attrp->spd_attr_value; 926 break; 927 928 case SPD_ATTR_AH_MINBITS: 929 act.ipa_apply.ipp_ah_minbits = attrp->spd_attr_value; 930 break; 931 932 case SPD_ATTR_AH_MAXBITS: 933 act.ipa_apply.ipp_ah_maxbits = attrp->spd_attr_value; 934 break; 935 936 case SPD_ATTR_ESPA_MINBITS: 937 act.ipa_apply.ipp_espa_minbits = attrp->spd_attr_value; 938 break; 939 940 case SPD_ATTR_ESPA_MAXBITS: 941 act.ipa_apply.ipp_espa_maxbits = attrp->spd_attr_value; 942 break; 943 944 case SPD_ATTR_LIFE_SOFT_TIME: 945 case SPD_ATTR_LIFE_HARD_TIME: 946 case SPD_ATTR_LIFE_SOFT_BYTES: 947 case SPD_ATTR_LIFE_HARD_BYTES: 948 break; 949 950 case SPD_ATTR_KM_PROTO: 951 act.ipa_apply.ipp_km_proto = attrp->spd_attr_value; 952 break; 953 954 case SPD_ATTR_KM_COOKIE: 955 act.ipa_apply.ipp_km_cookie = attrp->spd_attr_value; 956 break; 957 958 case SPD_ATTR_REPLAY_DEPTH: 959 act.ipa_apply.ipp_replay_depth = attrp->spd_attr_value; 960 break; 961 } 962 } 963 if (actp != endactp) { 964 *diag = SPD_DIAGNOSTIC_ADD_WRONG_ACT_COUNT; 965 goto fail; 966 } 967 968 return (B_TRUE); 969 fail: 970 ipsec_actvec_free(*actpp, nact); 971 *actpp = NULL; 972 return (B_FALSE); 973 } 974 975 typedef struct 976 { 977 ipsec_policy_t *pol; 978 int dir; 979 } tmprule_t; 980 981 static int 982 mkrule(ipsec_policy_head_t *iph, struct spd_rule *rule, 983 ipsec_selkey_t *sel, ipsec_act_t *actp, int nact, uint_t dir, uint_t af, 984 tmprule_t **rp, uint64_t *index, spd_stack_t *spds) 985 { 986 ipsec_policy_t *pol; 987 988 sel->ipsl_valid &= ~(IPSL_IPV6|IPSL_IPV4); 989 sel->ipsl_valid |= af; 990 991 pol = ipsec_policy_create(sel, actp, nact, rule->spd_rule_priority, 992 index, spds->spds_netstack); 993 if (pol == NULL) 994 return (ENOMEM); 995 996 (*rp)->pol = pol; 997 (*rp)->dir = dir; 998 (*rp)++; 999 1000 if (!ipsec_check_policy(iph, pol, dir)) 1001 return (EEXIST); 1002 1003 rule->spd_rule_index = pol->ipsp_index; 1004 return (0); 1005 } 1006 1007 static int 1008 mkrulepair(ipsec_policy_head_t *iph, struct spd_rule *rule, 1009 ipsec_selkey_t *sel, ipsec_act_t *actp, int nact, uint_t dir, uint_t afs, 1010 tmprule_t **rp, uint64_t *index, spd_stack_t *spds) 1011 { 1012 int error; 1013 1014 if (afs & IPSL_IPV4) { 1015 error = mkrule(iph, rule, sel, actp, nact, dir, IPSL_IPV4, rp, 1016 index, spds); 1017 if (error != 0) 1018 return (error); 1019 } 1020 if (afs & IPSL_IPV6) { 1021 error = mkrule(iph, rule, sel, actp, nact, dir, IPSL_IPV6, rp, 1022 index, spds); 1023 if (error != 0) 1024 return (error); 1025 } 1026 return (0); 1027 } 1028 1029 1030 static void 1031 spdsock_addrule(queue_t *q, ipsec_policy_head_t *iph, mblk_t *mp, 1032 spd_ext_t **extv, ipsec_tun_pol_t *itp) 1033 { 1034 ipsec_selkey_t sel; 1035 ipsec_act_t *actp; 1036 uint_t nact; 1037 int diag = 0, error, afs; 1038 struct spd_rule *rule = (struct spd_rule *)extv[SPD_EXT_RULE]; 1039 tmprule_t rules[4], *rulep = &rules[0]; 1040 boolean_t tunnel_mode, empty_itp, active; 1041 uint64_t *index = (itp == NULL) ? NULL : &itp->itp_next_policy_index; 1042 spdsock_t *ss = (spdsock_t *)q->q_ptr; 1043 spd_stack_t *spds = ss->spdsock_spds; 1044 uint32_t auditing = AU_AUDITING(); 1045 1046 if (rule == NULL) { 1047 spdsock_diag(q, mp, SPD_DIAGNOSTIC_NO_RULE_EXT); 1048 if (auditing) { 1049 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1050 cred_t *cr; 1051 pid_t cpid; 1052 1053 cr = msg_getcred(mp, &cpid); 1054 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1055 audit_pf_policy(SPD_ADDRULE, cr, 1056 spds->spds_netstack, ITP_NAME(itp), active, 1057 SPD_DIAGNOSTIC_NO_RULE_EXT, cpid); 1058 } 1059 return; 1060 } 1061 1062 tunnel_mode = (rule->spd_rule_flags & SPD_RULE_FLAG_TUNNEL); 1063 1064 if (itp != NULL) { 1065 mutex_enter(&itp->itp_lock); 1066 ASSERT(itp->itp_policy == iph || itp->itp_inactive == iph); 1067 active = (itp->itp_policy == iph); 1068 if (ITP_P_ISACTIVE(itp, iph)) { 1069 /* Check for mix-and-match of tunnel/transport. */ 1070 if ((tunnel_mode && !ITP_P_ISTUNNEL(itp, iph)) || 1071 (!tunnel_mode && ITP_P_ISTUNNEL(itp, iph))) { 1072 mutex_exit(&itp->itp_lock); 1073 spdsock_error(q, mp, EBUSY, 0); 1074 return; 1075 } 1076 empty_itp = B_FALSE; 1077 } else { 1078 empty_itp = B_TRUE; 1079 itp->itp_flags = active ? ITPF_P_ACTIVE : ITPF_I_ACTIVE; 1080 if (tunnel_mode) 1081 itp->itp_flags |= active ? ITPF_P_TUNNEL : 1082 ITPF_I_TUNNEL; 1083 } 1084 } else { 1085 empty_itp = B_FALSE; 1086 } 1087 1088 if (rule->spd_rule_index != 0) { 1089 diag = SPD_DIAGNOSTIC_INVALID_RULE_INDEX; 1090 error = EINVAL; 1091 goto fail2; 1092 } 1093 1094 if (!spdsock_ext_to_sel(extv, &sel, &diag)) { 1095 error = EINVAL; 1096 goto fail2; 1097 } 1098 1099 if (itp != NULL) { 1100 if (tunnel_mode) { 1101 if (sel.ipsl_valid & 1102 (IPSL_REMOTE_PORT | IPSL_LOCAL_PORT)) { 1103 itp->itp_flags |= active ? 1104 ITPF_P_PER_PORT_SECURITY : 1105 ITPF_I_PER_PORT_SECURITY; 1106 } 1107 } else { 1108 /* 1109 * For now, we don't allow transport-mode on a tunnel 1110 * with ANY specific selectors. Bail if we have such 1111 * a request. 1112 */ 1113 if (sel.ipsl_valid & IPSL_WILDCARD) { 1114 diag = SPD_DIAGNOSTIC_NO_TUNNEL_SELECTORS; 1115 error = EINVAL; 1116 goto fail2; 1117 } 1118 } 1119 } 1120 1121 if (!spdsock_ext_to_actvec(extv, &actp, &nact, &diag, spds)) { 1122 error = EINVAL; 1123 goto fail2; 1124 } 1125 /* 1126 * If no addresses were specified, add both. 1127 */ 1128 afs = sel.ipsl_valid & (IPSL_IPV6|IPSL_IPV4); 1129 if (afs == 0) 1130 afs = (IPSL_IPV6|IPSL_IPV4); 1131 1132 rw_enter(&iph->iph_lock, RW_WRITER); 1133 1134 if (rule->spd_rule_flags & SPD_RULE_FLAG_OUTBOUND) { 1135 error = mkrulepair(iph, rule, &sel, actp, nact, 1136 IPSEC_TYPE_OUTBOUND, afs, &rulep, index, spds); 1137 if (error != 0) 1138 goto fail; 1139 } 1140 1141 if (rule->spd_rule_flags & SPD_RULE_FLAG_INBOUND) { 1142 error = mkrulepair(iph, rule, &sel, actp, nact, 1143 IPSEC_TYPE_INBOUND, afs, &rulep, index, spds); 1144 if (error != 0) 1145 goto fail; 1146 } 1147 1148 while ((--rulep) >= &rules[0]) { 1149 ipsec_enter_policy(iph, rulep->pol, rulep->dir, 1150 spds->spds_netstack); 1151 } 1152 rw_exit(&iph->iph_lock); 1153 if (itp != NULL) 1154 mutex_exit(&itp->itp_lock); 1155 1156 ipsec_actvec_free(actp, nact); 1157 spd_echo(q, mp); 1158 if (auditing) { 1159 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1160 cred_t *cr; 1161 pid_t cpid; 1162 1163 cr = msg_getcred(mp, &cpid); 1164 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1165 audit_pf_policy(SPD_ADDRULE, cr, spds->spds_netstack, 1166 ITP_NAME(itp), active, 0, cpid); 1167 } 1168 return; 1169 1170 fail: 1171 rw_exit(&iph->iph_lock); 1172 while ((--rulep) >= &rules[0]) 1173 IPPOL_REFRELE(rulep->pol); 1174 ipsec_actvec_free(actp, nact); 1175 fail2: 1176 if (itp != NULL) { 1177 if (empty_itp) 1178 itp->itp_flags = 0; 1179 mutex_exit(&itp->itp_lock); 1180 } 1181 spdsock_error(q, mp, error, diag); 1182 if (auditing) { 1183 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1184 cred_t *cr; 1185 pid_t cpid; 1186 1187 cr = msg_getcred(mp, &cpid); 1188 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1189 audit_pf_policy(SPD_ADDRULE, cr, spds->spds_netstack, 1190 ITP_NAME(itp), active, error, cpid); 1191 } 1192 } 1193 1194 void 1195 spdsock_deleterule(queue_t *q, ipsec_policy_head_t *iph, mblk_t *mp, 1196 spd_ext_t **extv, ipsec_tun_pol_t *itp) 1197 { 1198 ipsec_selkey_t sel; 1199 struct spd_rule *rule = (struct spd_rule *)extv[SPD_EXT_RULE]; 1200 int err, diag = 0; 1201 spdsock_t *ss = (spdsock_t *)q->q_ptr; 1202 netstack_t *ns = ss->spdsock_spds->spds_netstack; 1203 uint32_t auditing = AU_AUDITING(); 1204 1205 if (rule == NULL) { 1206 spdsock_diag(q, mp, SPD_DIAGNOSTIC_NO_RULE_EXT); 1207 if (auditing) { 1208 boolean_t active; 1209 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1210 cred_t *cr; 1211 pid_t cpid; 1212 1213 cr = msg_getcred(mp, &cpid); 1214 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1215 audit_pf_policy(SPD_DELETERULE, cr, ns, 1216 ITP_NAME(itp), active, SPD_DIAGNOSTIC_NO_RULE_EXT, 1217 cpid); 1218 } 1219 return; 1220 } 1221 1222 /* 1223 * Must enter itp_lock first to avoid deadlock. See tun.c's 1224 * set_sec_simple() for the other case of itp_lock and iph_lock. 1225 */ 1226 if (itp != NULL) 1227 mutex_enter(&itp->itp_lock); 1228 1229 if (rule->spd_rule_index != 0) { 1230 if (ipsec_policy_delete_index(iph, rule->spd_rule_index, ns) != 1231 0) { 1232 err = ESRCH; 1233 goto fail; 1234 } 1235 } else { 1236 if (!spdsock_ext_to_sel(extv, &sel, &diag)) { 1237 err = EINVAL; /* diag already set... */ 1238 goto fail; 1239 } 1240 1241 if ((rule->spd_rule_flags & SPD_RULE_FLAG_INBOUND) && 1242 !ipsec_policy_delete(iph, &sel, IPSEC_TYPE_INBOUND, ns)) { 1243 err = ESRCH; 1244 goto fail; 1245 } 1246 1247 if ((rule->spd_rule_flags & SPD_RULE_FLAG_OUTBOUND) && 1248 !ipsec_policy_delete(iph, &sel, IPSEC_TYPE_OUTBOUND, ns)) { 1249 err = ESRCH; 1250 goto fail; 1251 } 1252 } 1253 1254 if (itp != NULL) { 1255 ASSERT(iph == itp->itp_policy || iph == itp->itp_inactive); 1256 rw_enter(&iph->iph_lock, RW_READER); 1257 if (avl_numnodes(&iph->iph_rulebyid) == 0) { 1258 if (iph == itp->itp_policy) 1259 itp->itp_flags &= ~ITPF_PFLAGS; 1260 else 1261 itp->itp_flags &= ~ITPF_IFLAGS; 1262 } 1263 /* Can exit locks in any order. */ 1264 rw_exit(&iph->iph_lock); 1265 mutex_exit(&itp->itp_lock); 1266 } 1267 spd_echo(q, mp); 1268 if (auditing) { 1269 boolean_t active; 1270 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1271 cred_t *cr; 1272 pid_t cpid; 1273 1274 cr = msg_getcred(mp, &cpid); 1275 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1276 audit_pf_policy(SPD_DELETERULE, cr, ns, ITP_NAME(itp), 1277 active, 0, cpid); 1278 } 1279 return; 1280 fail: 1281 if (itp != NULL) 1282 mutex_exit(&itp->itp_lock); 1283 spdsock_error(q, mp, err, diag); 1284 if (auditing) { 1285 boolean_t active; 1286 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1287 cred_t *cr; 1288 pid_t cpid; 1289 1290 cr = msg_getcred(mp, &cpid); 1291 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1292 audit_pf_policy(SPD_DELETERULE, cr, ns, ITP_NAME(itp), 1293 active, err, cpid); 1294 } 1295 } 1296 1297 /* Do NOT consume a reference to itp. */ 1298 /* ARGSUSED */ 1299 static void 1300 spdsock_flip_node(ipsec_tun_pol_t *itp, void *ignoreme, netstack_t *ns) 1301 { 1302 mutex_enter(&itp->itp_lock); 1303 ITPF_SWAP(itp->itp_flags); 1304 ipsec_swap_policy(itp->itp_policy, itp->itp_inactive, ns); 1305 mutex_exit(&itp->itp_lock); 1306 /* SPD_FLIP is worth a tunnel MTU check. */ 1307 update_iptun_policy(itp); 1308 } 1309 1310 void 1311 spdsock_flip(queue_t *q, mblk_t *mp, spd_if_t *tunname) 1312 { 1313 char *tname; 1314 ipsec_tun_pol_t *itp; 1315 spdsock_t *ss = (spdsock_t *)q->q_ptr; 1316 netstack_t *ns = ss->spdsock_spds->spds_netstack; 1317 uint32_t auditing = AU_AUDITING(); 1318 1319 if (tunname != NULL) { 1320 tname = (char *)tunname->spd_if_name; 1321 if (*tname == '\0') { 1322 /* can't fail */ 1323 ipsec_swap_global_policy(ns); 1324 if (auditing) { 1325 boolean_t active; 1326 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1327 cred_t *cr; 1328 pid_t cpid; 1329 1330 cr = msg_getcred(mp, &cpid); 1331 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1332 audit_pf_policy(SPD_FLIP, cr, ns, 1333 NULL, active, 0, cpid); 1334 } 1335 itp_walk(spdsock_flip_node, NULL, ns); 1336 if (auditing) { 1337 boolean_t active; 1338 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1339 cred_t *cr; 1340 pid_t cpid; 1341 1342 cr = msg_getcred(mp, &cpid); 1343 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1344 audit_pf_policy(SPD_FLIP, cr, ns, 1345 "all tunnels", active, 0, cpid); 1346 } 1347 } else { 1348 itp = get_tunnel_policy(tname, ns); 1349 if (itp == NULL) { 1350 /* Better idea for "tunnel not found"? */ 1351 spdsock_error(q, mp, ESRCH, 0); 1352 if (auditing) { 1353 boolean_t active; 1354 spd_msg_t *spmsg = 1355 (spd_msg_t *)mp->b_rptr; 1356 cred_t *cr; 1357 pid_t cpid; 1358 1359 cr = msg_getcred(mp, &cpid); 1360 active = (spmsg->spd_msg_spdid == 1361 SPD_ACTIVE); 1362 audit_pf_policy(SPD_FLIP, cr, ns, 1363 ITP_NAME(itp), active, 1364 ESRCH, cpid); 1365 } 1366 return; 1367 } 1368 spdsock_flip_node(itp, NULL, ns); 1369 if (auditing) { 1370 boolean_t active; 1371 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1372 cred_t *cr; 1373 pid_t cpid; 1374 1375 cr = msg_getcred(mp, &cpid); 1376 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1377 audit_pf_policy(SPD_FLIP, cr, ns, 1378 ITP_NAME(itp), active, 0, cpid); 1379 } 1380 ITP_REFRELE(itp, ns); 1381 } 1382 } else { 1383 ipsec_swap_global_policy(ns); /* can't fail */ 1384 if (auditing) { 1385 boolean_t active; 1386 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 1387 cred_t *cr; 1388 pid_t cpid; 1389 1390 cr = msg_getcred(mp, &cpid); 1391 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 1392 audit_pf_policy(SPD_FLIP, cr, 1393 ns, NULL, active, 0, cpid); 1394 } 1395 } 1396 spd_echo(q, mp); 1397 } 1398 1399 /* 1400 * Unimplemented feature 1401 */ 1402 /* ARGSUSED */ 1403 static void 1404 spdsock_lookup(queue_t *q, ipsec_policy_head_t *iph, mblk_t *mp, 1405 spd_ext_t **extv, ipsec_tun_pol_t *itp) 1406 { 1407 spdsock_error(q, mp, EINVAL, 0); 1408 } 1409 1410 1411 static mblk_t * 1412 spdsock_dump_ruleset(mblk_t *req, ipsec_policy_head_t *iph, 1413 uint32_t count, uint16_t error) 1414 { 1415 size_t len = sizeof (spd_ruleset_ext_t) + sizeof (spd_msg_t); 1416 spd_msg_t *msg; 1417 spd_ruleset_ext_t *ruleset; 1418 mblk_t *m = allocb(len, BPRI_HI); 1419 1420 ASSERT(RW_READ_HELD(&iph->iph_lock)); 1421 1422 if (m == NULL) { 1423 return (NULL); 1424 } 1425 msg = (spd_msg_t *)m->b_rptr; 1426 ruleset = (spd_ruleset_ext_t *)(&msg[1]); 1427 1428 m->b_wptr = (uint8_t *)&ruleset[1]; 1429 1430 *msg = *(spd_msg_t *)(req->b_rptr); 1431 msg->spd_msg_len = SPD_8TO64(len); 1432 msg->spd_msg_errno = error; 1433 1434 ruleset->spd_ruleset_len = SPD_8TO64(sizeof (*ruleset)); 1435 ruleset->spd_ruleset_type = SPD_EXT_RULESET; 1436 ruleset->spd_ruleset_count = count; 1437 ruleset->spd_ruleset_version = iph->iph_gen; 1438 return (m); 1439 } 1440 1441 static mblk_t * 1442 spdsock_dump_finish(spdsock_t *ss, int error) 1443 { 1444 mblk_t *m; 1445 ipsec_policy_head_t *iph = ss->spdsock_dump_head; 1446 mblk_t *req = ss->spdsock_dump_req; 1447 netstack_t *ns = ss->spdsock_spds->spds_netstack; 1448 1449 rw_enter(&iph->iph_lock, RW_READER); 1450 m = spdsock_dump_ruleset(req, iph, ss->spdsock_dump_count, error); 1451 rw_exit(&iph->iph_lock); 1452 IPPH_REFRELE(iph, ns); 1453 if (ss->spdsock_itp != NULL) { 1454 ITP_REFRELE(ss->spdsock_itp, ns); 1455 ss->spdsock_itp = NULL; 1456 } 1457 ss->spdsock_dump_req = NULL; 1458 freemsg(req); 1459 1460 return (m); 1461 } 1462 1463 /* 1464 * Rule encoding functions. 1465 * We do a two-pass encode. 1466 * If base != NULL, fill in encoded rule part starting at base+offset. 1467 * Always return "offset" plus length of to-be-encoded data. 1468 */ 1469 static uint_t 1470 spdsock_encode_typecode(uint8_t *base, uint_t offset, uint8_t type, 1471 uint8_t type_end, uint8_t code, uint8_t code_end) 1472 { 1473 struct spd_typecode *tcp; 1474 1475 ASSERT(ALIGNED64(offset)); 1476 1477 if (base != NULL) { 1478 tcp = (struct spd_typecode *)(base + offset); 1479 tcp->spd_typecode_len = SPD_8TO64(sizeof (*tcp)); 1480 tcp->spd_typecode_exttype = SPD_EXT_ICMP_TYPECODE; 1481 tcp->spd_typecode_code = code; 1482 tcp->spd_typecode_type = type; 1483 tcp->spd_typecode_type_end = type_end; 1484 tcp->spd_typecode_code_end = code_end; 1485 } 1486 offset += sizeof (*tcp); 1487 1488 ASSERT(ALIGNED64(offset)); 1489 1490 return (offset); 1491 } 1492 1493 static uint_t 1494 spdsock_encode_proto(uint8_t *base, uint_t offset, uint8_t proto) 1495 { 1496 struct spd_proto *spp; 1497 1498 ASSERT(ALIGNED64(offset)); 1499 1500 if (base != NULL) { 1501 spp = (struct spd_proto *)(base + offset); 1502 spp->spd_proto_len = SPD_8TO64(sizeof (*spp)); 1503 spp->spd_proto_exttype = SPD_EXT_PROTO; 1504 spp->spd_proto_number = proto; 1505 spp->spd_proto_reserved1 = 0; 1506 spp->spd_proto_reserved2 = 0; 1507 } 1508 offset += sizeof (*spp); 1509 1510 ASSERT(ALIGNED64(offset)); 1511 1512 return (offset); 1513 } 1514 1515 static uint_t 1516 spdsock_encode_port(uint8_t *base, uint_t offset, uint16_t ext, uint16_t port) 1517 { 1518 struct spd_portrange *spp; 1519 1520 ASSERT(ALIGNED64(offset)); 1521 1522 if (base != NULL) { 1523 spp = (struct spd_portrange *)(base + offset); 1524 spp->spd_ports_len = SPD_8TO64(sizeof (*spp)); 1525 spp->spd_ports_exttype = ext; 1526 spp->spd_ports_minport = port; 1527 spp->spd_ports_maxport = port; 1528 } 1529 offset += sizeof (*spp); 1530 1531 ASSERT(ALIGNED64(offset)); 1532 1533 return (offset); 1534 } 1535 1536 static uint_t 1537 spdsock_encode_addr(uint8_t *base, uint_t offset, uint16_t ext, 1538 const ipsec_selkey_t *sel, const ipsec_addr_t *addr, uint_t pfxlen) 1539 { 1540 struct spd_address *sae; 1541 ipsec_addr_t *spdaddr; 1542 uint_t start = offset; 1543 uint_t addrlen; 1544 uint_t af; 1545 1546 if (sel->ipsl_valid & IPSL_IPV4) { 1547 af = AF_INET; 1548 addrlen = IP_ADDR_LEN; 1549 } else { 1550 af = AF_INET6; 1551 addrlen = IPV6_ADDR_LEN; 1552 } 1553 1554 ASSERT(ALIGNED64(offset)); 1555 1556 if (base != NULL) { 1557 sae = (struct spd_address *)(base + offset); 1558 sae->spd_address_exttype = ext; 1559 sae->spd_address_af = af; 1560 sae->spd_address_prefixlen = pfxlen; 1561 sae->spd_address_reserved2 = 0; 1562 1563 spdaddr = (ipsec_addr_t *)(&sae[1]); 1564 bcopy(addr, spdaddr, addrlen); 1565 } 1566 offset += sizeof (*sae); 1567 addrlen = roundup(addrlen, sizeof (uint64_t)); 1568 offset += addrlen; 1569 1570 ASSERT(ALIGNED64(offset)); 1571 1572 if (base != NULL) 1573 sae->spd_address_len = SPD_8TO64(offset - start); 1574 return (offset); 1575 } 1576 1577 static uint_t 1578 spdsock_encode_sel(uint8_t *base, uint_t offset, const ipsec_sel_t *sel) 1579 { 1580 const ipsec_selkey_t *selkey = &sel->ipsl_key; 1581 1582 if (selkey->ipsl_valid & IPSL_PROTOCOL) 1583 offset = spdsock_encode_proto(base, offset, selkey->ipsl_proto); 1584 if (selkey->ipsl_valid & IPSL_LOCAL_PORT) 1585 offset = spdsock_encode_port(base, offset, SPD_EXT_LCLPORT, 1586 selkey->ipsl_lport); 1587 if (selkey->ipsl_valid & IPSL_REMOTE_PORT) 1588 offset = spdsock_encode_port(base, offset, SPD_EXT_REMPORT, 1589 selkey->ipsl_rport); 1590 if (selkey->ipsl_valid & IPSL_REMOTE_ADDR) 1591 offset = spdsock_encode_addr(base, offset, SPD_EXT_REMADDR, 1592 selkey, &selkey->ipsl_remote, selkey->ipsl_remote_pfxlen); 1593 if (selkey->ipsl_valid & IPSL_LOCAL_ADDR) 1594 offset = spdsock_encode_addr(base, offset, SPD_EXT_LCLADDR, 1595 selkey, &selkey->ipsl_local, selkey->ipsl_local_pfxlen); 1596 if (selkey->ipsl_valid & IPSL_ICMP_TYPE) { 1597 offset = spdsock_encode_typecode(base, offset, 1598 selkey->ipsl_icmp_type, selkey->ipsl_icmp_type_end, 1599 (selkey->ipsl_valid & IPSL_ICMP_CODE) ? 1600 selkey->ipsl_icmp_code : 255, 1601 (selkey->ipsl_valid & IPSL_ICMP_CODE) ? 1602 selkey->ipsl_icmp_code_end : 255); 1603 } 1604 return (offset); 1605 } 1606 1607 static uint_t 1608 spdsock_encode_actattr(uint8_t *base, uint_t offset, uint32_t tag, 1609 uint32_t value) 1610 { 1611 struct spd_attribute *attr; 1612 1613 ASSERT(ALIGNED64(offset)); 1614 1615 if (base != NULL) { 1616 attr = (struct spd_attribute *)(base + offset); 1617 attr->spd_attr_tag = tag; 1618 attr->spd_attr_value = value; 1619 } 1620 offset += sizeof (struct spd_attribute); 1621 1622 ASSERT(ALIGNED64(offset)); 1623 1624 return (offset); 1625 } 1626 1627 1628 #define EMIT(t, v) offset = spdsock_encode_actattr(base, offset, (t), (v)) 1629 1630 static uint_t 1631 spdsock_encode_action(uint8_t *base, uint_t offset, const ipsec_action_t *ap) 1632 { 1633 const struct ipsec_act *act = &(ap->ipa_act); 1634 uint_t flags; 1635 1636 EMIT(SPD_ATTR_EMPTY, 0); 1637 switch (act->ipa_type) { 1638 case IPSEC_ACT_DISCARD: 1639 case IPSEC_ACT_REJECT: 1640 EMIT(SPD_ATTR_TYPE, SPD_ACTTYPE_DROP); 1641 break; 1642 case IPSEC_ACT_BYPASS: 1643 case IPSEC_ACT_CLEAR: 1644 EMIT(SPD_ATTR_TYPE, SPD_ACTTYPE_PASS); 1645 break; 1646 1647 case IPSEC_ACT_APPLY: 1648 EMIT(SPD_ATTR_TYPE, SPD_ACTTYPE_IPSEC); 1649 flags = 0; 1650 if (act->ipa_apply.ipp_use_ah) 1651 flags |= SPD_APPLY_AH; 1652 if (act->ipa_apply.ipp_use_esp) 1653 flags |= SPD_APPLY_ESP; 1654 if (act->ipa_apply.ipp_use_espa) 1655 flags |= SPD_APPLY_ESPA; 1656 if (act->ipa_apply.ipp_use_se) 1657 flags |= SPD_APPLY_SE; 1658 if (act->ipa_apply.ipp_use_unique) 1659 flags |= SPD_APPLY_UNIQUE; 1660 EMIT(SPD_ATTR_FLAGS, flags); 1661 if (flags & SPD_APPLY_AH) { 1662 EMIT(SPD_ATTR_AH_AUTH, act->ipa_apply.ipp_auth_alg); 1663 EMIT(SPD_ATTR_AH_MINBITS, 1664 act->ipa_apply.ipp_ah_minbits); 1665 EMIT(SPD_ATTR_AH_MAXBITS, 1666 act->ipa_apply.ipp_ah_maxbits); 1667 } 1668 if (flags & SPD_APPLY_ESP) { 1669 EMIT(SPD_ATTR_ESP_ENCR, act->ipa_apply.ipp_encr_alg); 1670 EMIT(SPD_ATTR_ENCR_MINBITS, 1671 act->ipa_apply.ipp_espe_minbits); 1672 EMIT(SPD_ATTR_ENCR_MAXBITS, 1673 act->ipa_apply.ipp_espe_maxbits); 1674 if (flags & SPD_APPLY_ESPA) { 1675 EMIT(SPD_ATTR_ESP_AUTH, 1676 act->ipa_apply.ipp_esp_auth_alg); 1677 EMIT(SPD_ATTR_ESPA_MINBITS, 1678 act->ipa_apply.ipp_espa_minbits); 1679 EMIT(SPD_ATTR_ESPA_MAXBITS, 1680 act->ipa_apply.ipp_espa_maxbits); 1681 } 1682 } 1683 if (act->ipa_apply.ipp_km_proto != 0) 1684 EMIT(SPD_ATTR_KM_PROTO, act->ipa_apply.ipp_km_proto); 1685 if (act->ipa_apply.ipp_km_cookie != 0) 1686 EMIT(SPD_ATTR_KM_PROTO, act->ipa_apply.ipp_km_cookie); 1687 if (act->ipa_apply.ipp_replay_depth != 0) 1688 EMIT(SPD_ATTR_REPLAY_DEPTH, 1689 act->ipa_apply.ipp_replay_depth); 1690 /* Add more here */ 1691 break; 1692 } 1693 1694 return (offset); 1695 } 1696 1697 static uint_t 1698 spdsock_encode_action_list(uint8_t *base, uint_t offset, 1699 const ipsec_action_t *ap) 1700 { 1701 struct spd_ext_actions *act; 1702 uint_t nact = 0; 1703 uint_t start = offset; 1704 1705 ASSERT(ALIGNED64(offset)); 1706 1707 if (base != NULL) { 1708 act = (struct spd_ext_actions *)(base + offset); 1709 act->spd_actions_len = 0; 1710 act->spd_actions_exttype = SPD_EXT_ACTION; 1711 act->spd_actions_count = 0; 1712 act->spd_actions_reserved = 0; 1713 } 1714 1715 offset += sizeof (*act); 1716 1717 ASSERT(ALIGNED64(offset)); 1718 1719 while (ap != NULL) { 1720 offset = spdsock_encode_action(base, offset, ap); 1721 ap = ap->ipa_next; 1722 nact++; 1723 if (ap != NULL) { 1724 EMIT(SPD_ATTR_NEXT, 0); 1725 } 1726 } 1727 EMIT(SPD_ATTR_END, 0); 1728 1729 ASSERT(ALIGNED64(offset)); 1730 1731 if (base != NULL) { 1732 act->spd_actions_count = nact; 1733 act->spd_actions_len = SPD_8TO64(offset - start); 1734 } 1735 1736 return (offset); 1737 } 1738 1739 #undef EMIT 1740 1741 /* ARGSUSED */ 1742 static uint_t 1743 spdsock_rule_flags(uint_t dir, uint_t af) 1744 { 1745 uint_t flags = 0; 1746 1747 if (dir == IPSEC_TYPE_INBOUND) 1748 flags |= SPD_RULE_FLAG_INBOUND; 1749 if (dir == IPSEC_TYPE_OUTBOUND) 1750 flags |= SPD_RULE_FLAG_OUTBOUND; 1751 1752 return (flags); 1753 } 1754 1755 1756 static uint_t 1757 spdsock_encode_rule_head(uint8_t *base, uint_t offset, spd_msg_t *req, 1758 const ipsec_policy_t *rule, uint_t dir, uint_t af, char *name, 1759 boolean_t tunnel) 1760 { 1761 struct spd_msg *spmsg; 1762 struct spd_rule *spr; 1763 spd_if_t *sid; 1764 1765 uint_t start = offset; 1766 1767 ASSERT(ALIGNED64(offset)); 1768 1769 if (base != NULL) { 1770 spmsg = (struct spd_msg *)(base + offset); 1771 bzero(spmsg, sizeof (*spmsg)); 1772 spmsg->spd_msg_version = PF_POLICY_V1; 1773 spmsg->spd_msg_type = SPD_DUMP; 1774 spmsg->spd_msg_seq = req->spd_msg_seq; 1775 spmsg->spd_msg_pid = req->spd_msg_pid; 1776 } 1777 offset += sizeof (struct spd_msg); 1778 1779 ASSERT(ALIGNED64(offset)); 1780 1781 if (base != NULL) { 1782 spr = (struct spd_rule *)(base + offset); 1783 spr->spd_rule_type = SPD_EXT_RULE; 1784 spr->spd_rule_priority = rule->ipsp_prio; 1785 spr->spd_rule_flags = spdsock_rule_flags(dir, af); 1786 if (tunnel) 1787 spr->spd_rule_flags |= SPD_RULE_FLAG_TUNNEL; 1788 spr->spd_rule_unused = 0; 1789 spr->spd_rule_len = SPD_8TO64(sizeof (*spr)); 1790 spr->spd_rule_index = rule->ipsp_index; 1791 } 1792 offset += sizeof (struct spd_rule); 1793 1794 /* 1795 * If we have an interface name (i.e. if this policy head came from 1796 * a tunnel), add the SPD_EXT_TUN_NAME extension. 1797 */ 1798 if (name != NULL) { 1799 1800 ASSERT(ALIGNED64(offset)); 1801 1802 if (base != NULL) { 1803 sid = (spd_if_t *)(base + offset); 1804 sid->spd_if_exttype = SPD_EXT_TUN_NAME; 1805 sid->spd_if_len = SPD_8TO64(sizeof (spd_if_t) + 1806 roundup((strlen(name) - 4), 8)); 1807 (void) strlcpy((char *)sid->spd_if_name, name, 1808 LIFNAMSIZ); 1809 } 1810 1811 offset += sizeof (spd_if_t) + roundup((strlen(name) - 4), 8); 1812 } 1813 1814 offset = spdsock_encode_sel(base, offset, rule->ipsp_sel); 1815 offset = spdsock_encode_action_list(base, offset, rule->ipsp_act); 1816 1817 ASSERT(ALIGNED64(offset)); 1818 1819 if (base != NULL) { 1820 spmsg->spd_msg_len = SPD_8TO64(offset - start); 1821 } 1822 return (offset); 1823 } 1824 1825 /* ARGSUSED */ 1826 static mblk_t * 1827 spdsock_encode_rule(mblk_t *req, const ipsec_policy_t *rule, 1828 uint_t dir, uint_t af, char *name, boolean_t tunnel) 1829 { 1830 mblk_t *m; 1831 uint_t len; 1832 spd_msg_t *mreq = (spd_msg_t *)req->b_rptr; 1833 1834 /* 1835 * Figure out how much space we'll need. 1836 */ 1837 len = spdsock_encode_rule_head(NULL, 0, mreq, rule, dir, af, name, 1838 tunnel); 1839 1840 /* 1841 * Allocate mblk. 1842 */ 1843 m = allocb(len, BPRI_HI); 1844 if (m == NULL) 1845 return (NULL); 1846 1847 /* 1848 * Fill it in.. 1849 */ 1850 m->b_wptr = m->b_rptr + len; 1851 bzero(m->b_rptr, len); 1852 (void) spdsock_encode_rule_head(m->b_rptr, 0, mreq, rule, dir, af, 1853 name, tunnel); 1854 return (m); 1855 } 1856 1857 static ipsec_policy_t * 1858 spdsock_dump_next_in_chain(spdsock_t *ss, ipsec_policy_head_t *iph, 1859 ipsec_policy_t *cur) 1860 { 1861 ASSERT(RW_READ_HELD(&iph->iph_lock)); 1862 1863 ss->spdsock_dump_count++; 1864 ss->spdsock_dump_cur_rule = cur->ipsp_hash.hash_next; 1865 return (cur); 1866 } 1867 1868 static ipsec_policy_t * 1869 spdsock_dump_next_rule(spdsock_t *ss, ipsec_policy_head_t *iph) 1870 { 1871 ipsec_policy_t *cur; 1872 ipsec_policy_root_t *ipr; 1873 int chain, nchains, type, af; 1874 1875 ASSERT(RW_READ_HELD(&iph->iph_lock)); 1876 1877 cur = ss->spdsock_dump_cur_rule; 1878 1879 if (cur != NULL) 1880 return (spdsock_dump_next_in_chain(ss, iph, cur)); 1881 1882 type = ss->spdsock_dump_cur_type; 1883 1884 next: 1885 chain = ss->spdsock_dump_cur_chain; 1886 ipr = &iph->iph_root[type]; 1887 nchains = ipr->ipr_nchains; 1888 1889 while (chain < nchains) { 1890 cur = ipr->ipr_hash[chain].hash_head; 1891 chain++; 1892 if (cur != NULL) { 1893 ss->spdsock_dump_cur_chain = chain; 1894 return (spdsock_dump_next_in_chain(ss, iph, cur)); 1895 } 1896 } 1897 ss->spdsock_dump_cur_chain = nchains; 1898 1899 af = ss->spdsock_dump_cur_af; 1900 while (af < IPSEC_NAF) { 1901 cur = ipr->ipr_nonhash[af]; 1902 af++; 1903 if (cur != NULL) { 1904 ss->spdsock_dump_cur_af = af; 1905 return (spdsock_dump_next_in_chain(ss, iph, cur)); 1906 } 1907 } 1908 1909 type++; 1910 if (type >= IPSEC_NTYPES) 1911 return (NULL); 1912 1913 ss->spdsock_dump_cur_chain = 0; 1914 ss->spdsock_dump_cur_type = type; 1915 ss->spdsock_dump_cur_af = IPSEC_AF_V4; 1916 goto next; 1917 1918 } 1919 1920 /* 1921 * If we're done with one policy head, but have more to go, we iterate through 1922 * another IPsec tunnel policy head (itp). Return NULL if it is an error 1923 * worthy of returning EAGAIN via PF_POLICY. 1924 */ 1925 static ipsec_tun_pol_t * 1926 spdsock_dump_iterate_next_tunnel(spdsock_t *ss, ipsec_stack_t *ipss) 1927 { 1928 ipsec_tun_pol_t *itp; 1929 1930 ASSERT(RW_READ_HELD(&ipss->ipsec_tunnel_policy_lock)); 1931 if (ipss->ipsec_tunnel_policy_gen > ss->spdsock_dump_tun_gen) { 1932 /* Oops, state of the tunnel polheads changed. */ 1933 itp = NULL; 1934 } else if (ss->spdsock_itp == NULL) { 1935 /* Just finished global, find first node. */ 1936 itp = avl_first(&ipss->ipsec_tunnel_policies); 1937 } else { 1938 /* We just finished current polhead, find the next one. */ 1939 itp = AVL_NEXT(&ipss->ipsec_tunnel_policies, ss->spdsock_itp); 1940 } 1941 if (itp != NULL) { 1942 ITP_REFHOLD(itp); 1943 } 1944 if (ss->spdsock_itp != NULL) { 1945 ITP_REFRELE(ss->spdsock_itp, ipss->ipsec_netstack); 1946 } 1947 ss->spdsock_itp = itp; 1948 return (itp); 1949 } 1950 1951 static mblk_t * 1952 spdsock_dump_next_record(spdsock_t *ss) 1953 { 1954 ipsec_policy_head_t *iph; 1955 ipsec_policy_t *rule; 1956 mblk_t *m; 1957 ipsec_tun_pol_t *itp; 1958 netstack_t *ns = ss->spdsock_spds->spds_netstack; 1959 ipsec_stack_t *ipss = ns->netstack_ipsec; 1960 1961 iph = ss->spdsock_dump_head; 1962 1963 ASSERT(iph != NULL); 1964 1965 rw_enter(&iph->iph_lock, RW_READER); 1966 1967 if (iph->iph_gen != ss->spdsock_dump_gen) { 1968 rw_exit(&iph->iph_lock); 1969 return (spdsock_dump_finish(ss, EAGAIN)); 1970 } 1971 1972 while ((rule = spdsock_dump_next_rule(ss, iph)) == NULL) { 1973 rw_exit(&iph->iph_lock); 1974 if (--(ss->spdsock_dump_remaining_polheads) == 0) 1975 return (spdsock_dump_finish(ss, 0)); 1976 1977 1978 /* 1979 * If we reach here, we have more policy heads (tunnel 1980 * entries) to dump. Let's reset to a new policy head 1981 * and get some more rules. 1982 * 1983 * An empty policy head will have spdsock_dump_next_rule() 1984 * return NULL, and we loop (while dropping the number of 1985 * remaining polheads). If we loop to 0, we finish. We 1986 * keep looping until we hit 0 or until we have a rule to 1987 * encode. 1988 * 1989 * NOTE: No need for ITP_REF*() macros here as we're only 1990 * going after and refholding the policy head itself. 1991 */ 1992 rw_enter(&ipss->ipsec_tunnel_policy_lock, RW_READER); 1993 itp = spdsock_dump_iterate_next_tunnel(ss, ipss); 1994 if (itp == NULL) { 1995 rw_exit(&ipss->ipsec_tunnel_policy_lock); 1996 return (spdsock_dump_finish(ss, EAGAIN)); 1997 } 1998 1999 /* Reset other spdsock_dump thingies. */ 2000 IPPH_REFRELE(ss->spdsock_dump_head, ns); 2001 if (ss->spdsock_dump_active) { 2002 ss->spdsock_dump_tunnel = 2003 itp->itp_flags & ITPF_P_TUNNEL; 2004 iph = itp->itp_policy; 2005 } else { 2006 ss->spdsock_dump_tunnel = 2007 itp->itp_flags & ITPF_I_TUNNEL; 2008 iph = itp->itp_inactive; 2009 } 2010 IPPH_REFHOLD(iph); 2011 rw_exit(&ipss->ipsec_tunnel_policy_lock); 2012 2013 rw_enter(&iph->iph_lock, RW_READER); 2014 RESET_SPDSOCK_DUMP_POLHEAD(ss, iph); 2015 } 2016 2017 m = spdsock_encode_rule(ss->spdsock_dump_req, rule, 2018 ss->spdsock_dump_cur_type, ss->spdsock_dump_cur_af, 2019 (ss->spdsock_itp == NULL) ? NULL : ss->spdsock_itp->itp_name, 2020 ss->spdsock_dump_tunnel); 2021 rw_exit(&iph->iph_lock); 2022 2023 if (m == NULL) 2024 return (spdsock_dump_finish(ss, ENOMEM)); 2025 return (m); 2026 } 2027 2028 /* 2029 * Dump records until we run into flow-control back-pressure. 2030 */ 2031 static void 2032 spdsock_dump_some(queue_t *q, spdsock_t *ss) 2033 { 2034 mblk_t *m, *dataind; 2035 2036 while ((ss->spdsock_dump_req != NULL) && canputnext(q)) { 2037 m = spdsock_dump_next_record(ss); 2038 if (m == NULL) 2039 return; 2040 dataind = allocb(sizeof (struct T_data_req), BPRI_HI); 2041 if (dataind == NULL) { 2042 freemsg(m); 2043 return; 2044 } 2045 dataind->b_cont = m; 2046 dataind->b_wptr += sizeof (struct T_data_req); 2047 ((struct T_data_ind *)dataind->b_rptr)->PRIM_type = T_DATA_IND; 2048 ((struct T_data_ind *)dataind->b_rptr)->MORE_flag = 0; 2049 dataind->b_datap->db_type = M_PROTO; 2050 putnext(q, dataind); 2051 } 2052 } 2053 2054 /* 2055 * Start dumping. 2056 * Format a start-of-dump record, and set up the stream and kick the rsrv 2057 * procedure to continue the job.. 2058 */ 2059 /* ARGSUSED */ 2060 static void 2061 spdsock_dump(queue_t *q, ipsec_policy_head_t *iph, mblk_t *mp) 2062 { 2063 spdsock_t *ss = (spdsock_t *)q->q_ptr; 2064 netstack_t *ns = ss->spdsock_spds->spds_netstack; 2065 ipsec_stack_t *ipss = ns->netstack_ipsec; 2066 mblk_t *mr; 2067 2068 /* spdsock_open() already set spdsock_itp to NULL. */ 2069 if (iph == ALL_ACTIVE_POLHEADS || iph == ALL_INACTIVE_POLHEADS) { 2070 rw_enter(&ipss->ipsec_tunnel_policy_lock, RW_READER); 2071 ss->spdsock_dump_remaining_polheads = 1 + 2072 avl_numnodes(&ipss->ipsec_tunnel_policies); 2073 ss->spdsock_dump_tun_gen = ipss->ipsec_tunnel_policy_gen; 2074 rw_exit(&ipss->ipsec_tunnel_policy_lock); 2075 if (iph == ALL_ACTIVE_POLHEADS) { 2076 iph = ipsec_system_policy(ns); 2077 ss->spdsock_dump_active = B_TRUE; 2078 } else { 2079 iph = ipsec_inactive_policy(ns); 2080 ss->spdsock_dump_active = B_FALSE; 2081 } 2082 ASSERT(ss->spdsock_itp == NULL); 2083 } else { 2084 ss->spdsock_dump_remaining_polheads = 1; 2085 } 2086 2087 rw_enter(&iph->iph_lock, RW_READER); 2088 2089 mr = spdsock_dump_ruleset(mp, iph, 0, 0); 2090 2091 if (!mr) { 2092 rw_exit(&iph->iph_lock); 2093 spdsock_error(q, mp, ENOMEM, 0); 2094 return; 2095 } 2096 2097 ss->spdsock_dump_req = mp; 2098 RESET_SPDSOCK_DUMP_POLHEAD(ss, iph); 2099 2100 rw_exit(&iph->iph_lock); 2101 2102 qreply(q, mr); 2103 qenable(OTHERQ(q)); 2104 } 2105 2106 /* Do NOT consume a reference to ITP. */ 2107 void 2108 spdsock_clone_node(ipsec_tun_pol_t *itp, void *ep, netstack_t *ns) 2109 { 2110 int *errptr = (int *)ep; 2111 2112 if (*errptr != 0) 2113 return; /* We've failed already for some reason. */ 2114 mutex_enter(&itp->itp_lock); 2115 ITPF_CLONE(itp->itp_flags); 2116 *errptr = ipsec_copy_polhead(itp->itp_policy, itp->itp_inactive, ns); 2117 mutex_exit(&itp->itp_lock); 2118 } 2119 2120 void 2121 spdsock_clone(queue_t *q, mblk_t *mp, spd_if_t *tunname) 2122 { 2123 int error; 2124 char *tname; 2125 ipsec_tun_pol_t *itp; 2126 spdsock_t *ss = (spdsock_t *)q->q_ptr; 2127 netstack_t *ns = ss->spdsock_spds->spds_netstack; 2128 uint32_t auditing = AU_AUDITING(); 2129 2130 if (tunname != NULL) { 2131 tname = (char *)tunname->spd_if_name; 2132 if (*tname == '\0') { 2133 error = ipsec_clone_system_policy(ns); 2134 if (auditing) { 2135 boolean_t active; 2136 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 2137 cred_t *cr; 2138 pid_t cpid; 2139 2140 cr = msg_getcred(mp, &cpid); 2141 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 2142 audit_pf_policy(SPD_CLONE, cr, ns, 2143 NULL, active, error, cpid); 2144 } 2145 if (error == 0) { 2146 itp_walk(spdsock_clone_node, &error, ns); 2147 if (auditing) { 2148 boolean_t active; 2149 spd_msg_t *spmsg = 2150 (spd_msg_t *)mp->b_rptr; 2151 cred_t *cr; 2152 pid_t cpid; 2153 2154 cr = msg_getcred(mp, &cpid); 2155 active = (spmsg->spd_msg_spdid == 2156 SPD_ACTIVE); 2157 audit_pf_policy(SPD_CLONE, cr, 2158 ns, "all tunnels", active, 0, 2159 cpid); 2160 } 2161 } 2162 } else { 2163 itp = get_tunnel_policy(tname, ns); 2164 if (itp == NULL) { 2165 spdsock_error(q, mp, ENOENT, 0); 2166 if (auditing) { 2167 boolean_t active; 2168 spd_msg_t *spmsg = 2169 (spd_msg_t *)mp->b_rptr; 2170 cred_t *cr; 2171 pid_t cpid; 2172 2173 cr = msg_getcred(mp, &cpid); 2174 active = (spmsg->spd_msg_spdid == 2175 SPD_ACTIVE); 2176 audit_pf_policy(SPD_CLONE, cr, 2177 ns, NULL, active, ENOENT, cpid); 2178 } 2179 return; 2180 } 2181 spdsock_clone_node(itp, &error, NULL); 2182 if (auditing) { 2183 boolean_t active; 2184 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 2185 cred_t *cr; 2186 pid_t cpid; 2187 2188 cr = msg_getcred(mp, &cpid); 2189 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 2190 audit_pf_policy(SPD_CLONE, cr, ns, 2191 ITP_NAME(itp), active, error, cpid); 2192 } 2193 ITP_REFRELE(itp, ns); 2194 } 2195 } else { 2196 error = ipsec_clone_system_policy(ns); 2197 if (auditing) { 2198 boolean_t active; 2199 spd_msg_t *spmsg = (spd_msg_t *)mp->b_rptr; 2200 cred_t *cr; 2201 pid_t cpid; 2202 2203 cr = msg_getcred(mp, &cpid); 2204 active = (spmsg->spd_msg_spdid == SPD_ACTIVE); 2205 audit_pf_policy(SPD_CLONE, cr, ns, NULL, 2206 active, error, cpid); 2207 } 2208 } 2209 2210 if (error != 0) 2211 spdsock_error(q, mp, error, 0); 2212 else 2213 spd_echo(q, mp); 2214 } 2215 2216 /* 2217 * Process a SPD_ALGLIST request. The caller expects separate alg entries 2218 * for AH authentication, ESP authentication, and ESP encryption. 2219 * The same distinction is then used when setting the min and max key 2220 * sizes when defining policies. 2221 */ 2222 2223 #define SPDSOCK_AH_AUTH 0 2224 #define SPDSOCK_ESP_AUTH 1 2225 #define SPDSOCK_ESP_ENCR 2 2226 #define SPDSOCK_NTYPES 3 2227 2228 static const uint_t algattr[SPDSOCK_NTYPES] = { 2229 SPD_ATTR_AH_AUTH, 2230 SPD_ATTR_ESP_AUTH, 2231 SPD_ATTR_ESP_ENCR 2232 }; 2233 static const uint_t minbitsattr[SPDSOCK_NTYPES] = { 2234 SPD_ATTR_AH_MINBITS, 2235 SPD_ATTR_ESPA_MINBITS, 2236 SPD_ATTR_ENCR_MINBITS 2237 }; 2238 static const uint_t maxbitsattr[SPDSOCK_NTYPES] = { 2239 SPD_ATTR_AH_MAXBITS, 2240 SPD_ATTR_ESPA_MAXBITS, 2241 SPD_ATTR_ENCR_MAXBITS 2242 }; 2243 static const uint_t defbitsattr[SPDSOCK_NTYPES] = { 2244 SPD_ATTR_AH_DEFBITS, 2245 SPD_ATTR_ESPA_DEFBITS, 2246 SPD_ATTR_ENCR_DEFBITS 2247 }; 2248 static const uint_t incrbitsattr[SPDSOCK_NTYPES] = { 2249 SPD_ATTR_AH_INCRBITS, 2250 SPD_ATTR_ESPA_INCRBITS, 2251 SPD_ATTR_ENCR_INCRBITS 2252 }; 2253 2254 #define ATTRPERALG 6 /* fixed attributes per algs */ 2255 2256 void 2257 spdsock_alglist(queue_t *q, mblk_t *mp) 2258 { 2259 uint_t algtype; 2260 uint_t algidx; 2261 uint_t algcount; 2262 uint_t size; 2263 mblk_t *m; 2264 uint8_t *cur; 2265 spd_msg_t *msg; 2266 struct spd_ext_actions *act; 2267 struct spd_attribute *attr; 2268 spdsock_t *ss = (spdsock_t *)q->q_ptr; 2269 ipsec_stack_t *ipss = ss->spdsock_spds->spds_netstack->netstack_ipsec; 2270 2271 mutex_enter(&ipss->ipsec_alg_lock); 2272 /* 2273 * The SPD client expects to receive separate entries for 2274 * AH authentication and ESP authentication supported algorithms. 2275 * 2276 * Don't return the "any" algorithms, if defined, as no 2277 * kernel policies can be set for these algorithms. 2278 */ 2279 algcount = 2 * ipss->ipsec_nalgs[IPSEC_ALG_AUTH] + 2280 ipss->ipsec_nalgs[IPSEC_ALG_ENCR]; 2281 2282 if (ipss->ipsec_alglists[IPSEC_ALG_AUTH][SADB_AALG_NONE] != NULL) 2283 algcount--; 2284 if (ipss->ipsec_alglists[IPSEC_ALG_ENCR][SADB_EALG_NONE] != NULL) 2285 algcount--; 2286 2287 /* 2288 * For each algorithm, we encode: 2289 * ALG / MINBITS / MAXBITS / DEFBITS / INCRBITS / {END, NEXT} 2290 */ 2291 2292 size = sizeof (spd_msg_t) + sizeof (struct spd_ext_actions) + 2293 ATTRPERALG * sizeof (struct spd_attribute) * algcount; 2294 2295 ASSERT(ALIGNED64(size)); 2296 2297 m = allocb(size, BPRI_HI); 2298 if (m == NULL) { 2299 mutex_exit(&ipss->ipsec_alg_lock); 2300 spdsock_error(q, mp, ENOMEM, 0); 2301 return; 2302 } 2303 2304 m->b_wptr = m->b_rptr + size; 2305 cur = m->b_rptr; 2306 2307 msg = (spd_msg_t *)cur; 2308 bcopy(mp->b_rptr, cur, sizeof (*msg)); 2309 2310 msg->spd_msg_len = SPD_8TO64(size); 2311 msg->spd_msg_errno = 0; 2312 msg->spd_msg_diagnostic = 0; 2313 2314 cur += sizeof (*msg); 2315 2316 act = (struct spd_ext_actions *)cur; 2317 cur += sizeof (*act); 2318 2319 act->spd_actions_len = SPD_8TO64(size - sizeof (spd_msg_t)); 2320 act->spd_actions_exttype = SPD_EXT_ACTION; 2321 act->spd_actions_count = algcount; 2322 act->spd_actions_reserved = 0; 2323 2324 attr = (struct spd_attribute *)cur; 2325 2326 #define EMIT(tag, value) { \ 2327 attr->spd_attr_tag = (tag); \ 2328 attr->spd_attr_value = (value); \ 2329 attr++; \ 2330 } 2331 2332 /* 2333 * If you change the number of EMIT's here, change 2334 * ATTRPERALG above to match 2335 */ 2336 #define EMITALGATTRS(_type) { \ 2337 EMIT(algattr[_type], algid); /* 1 */ \ 2338 EMIT(minbitsattr[_type], minbits); /* 2 */ \ 2339 EMIT(maxbitsattr[_type], maxbits); /* 3 */ \ 2340 EMIT(defbitsattr[_type], defbits); /* 4 */ \ 2341 EMIT(incrbitsattr[_type], incr); /* 5 */ \ 2342 EMIT(SPD_ATTR_NEXT, 0); /* 6 */ \ 2343 } 2344 2345 for (algtype = 0; algtype < IPSEC_NALGTYPES; algtype++) { 2346 for (algidx = 0; algidx < ipss->ipsec_nalgs[algtype]; 2347 algidx++) { 2348 int algid = ipss->ipsec_sortlist[algtype][algidx]; 2349 ipsec_alginfo_t *alg = 2350 ipss->ipsec_alglists[algtype][algid]; 2351 uint_t minbits = alg->alg_minbits; 2352 uint_t maxbits = alg->alg_maxbits; 2353 uint_t defbits = alg->alg_default_bits; 2354 uint_t incr = alg->alg_increment; 2355 2356 if (algtype == IPSEC_ALG_AUTH) { 2357 if (algid == SADB_AALG_NONE) 2358 continue; 2359 EMITALGATTRS(SPDSOCK_AH_AUTH); 2360 EMITALGATTRS(SPDSOCK_ESP_AUTH); 2361 } else { 2362 if (algid == SADB_EALG_NONE) 2363 continue; 2364 ASSERT(algtype == IPSEC_ALG_ENCR); 2365 EMITALGATTRS(SPDSOCK_ESP_ENCR); 2366 } 2367 } 2368 } 2369 2370 mutex_exit(&ipss->ipsec_alg_lock); 2371 2372 #undef EMITALGATTRS 2373 #undef EMIT 2374 #undef ATTRPERALG 2375 2376 attr--; 2377 attr->spd_attr_tag = SPD_ATTR_END; 2378 2379 freemsg(mp); 2380 qreply(q, m); 2381 } 2382 2383 /* 2384 * Process a SPD_DUMPALGS request. 2385 */ 2386 2387 #define ATTRPERALG 9 /* fixed attributes per algs */ 2388 2389 void 2390 spdsock_dumpalgs(queue_t *q, mblk_t *mp) 2391 { 2392 uint_t algtype; 2393 uint_t algidx; 2394 uint_t size; 2395 mblk_t *m; 2396 uint8_t *cur; 2397 spd_msg_t *msg; 2398 struct spd_ext_actions *act; 2399 struct spd_attribute *attr; 2400 ipsec_alginfo_t *alg; 2401 uint_t algid; 2402 uint_t i; 2403 uint_t alg_size; 2404 spdsock_t *ss = (spdsock_t *)q->q_ptr; 2405 ipsec_stack_t *ipss = ss->spdsock_spds->spds_netstack->netstack_ipsec; 2406 2407 mutex_enter(&ipss->ipsec_alg_lock); 2408 2409 /* 2410 * For each algorithm, we encode: 2411 * ALG / MINBITS / MAXBITS / DEFBITS / INCRBITS / {END, NEXT} 2412 * 2413 * ALG_ID / ALG_PROTO / ALG_INCRBITS / ALG_NKEYSIZES / ALG_KEYSIZE* 2414 * ALG_NBLOCKSIZES / ALG_BLOCKSIZE* / ALG_NPARAMS / ALG_PARAMS* / 2415 * ALG_MECHNAME / ALG_FLAGS / {END, NEXT} 2416 */ 2417 2418 /* 2419 * Compute the size of the SPD message. 2420 */ 2421 size = sizeof (spd_msg_t) + sizeof (struct spd_ext_actions); 2422 2423 for (algtype = 0; algtype < IPSEC_NALGTYPES; algtype++) { 2424 for (algidx = 0; algidx < ipss->ipsec_nalgs[algtype]; 2425 algidx++) { 2426 algid = ipss->ipsec_sortlist[algtype][algidx]; 2427 alg = ipss->ipsec_alglists[algtype][algid]; 2428 alg_size = sizeof (struct spd_attribute) * 2429 (ATTRPERALG + alg->alg_nkey_sizes + 2430 alg->alg_nblock_sizes + alg->alg_nparams) + 2431 CRYPTO_MAX_MECH_NAME; 2432 size += alg_size; 2433 } 2434 } 2435 2436 ASSERT(ALIGNED64(size)); 2437 2438 m = allocb(size, BPRI_HI); 2439 if (m == NULL) { 2440 mutex_exit(&ipss->ipsec_alg_lock); 2441 spdsock_error(q, mp, ENOMEM, 0); 2442 return; 2443 } 2444 2445 m->b_wptr = m->b_rptr + size; 2446 cur = m->b_rptr; 2447 2448 msg = (spd_msg_t *)cur; 2449 bcopy(mp->b_rptr, cur, sizeof (*msg)); 2450 2451 msg->spd_msg_len = SPD_8TO64(size); 2452 msg->spd_msg_errno = 0; 2453 msg->spd_msg_type = SPD_ALGLIST; 2454 2455 msg->spd_msg_diagnostic = 0; 2456 2457 cur += sizeof (*msg); 2458 2459 act = (struct spd_ext_actions *)cur; 2460 cur += sizeof (*act); 2461 2462 act->spd_actions_len = SPD_8TO64(size - sizeof (spd_msg_t)); 2463 act->spd_actions_exttype = SPD_EXT_ACTION; 2464 act->spd_actions_count = ipss->ipsec_nalgs[IPSEC_ALG_AUTH] + 2465 ipss->ipsec_nalgs[IPSEC_ALG_ENCR]; 2466 act->spd_actions_reserved = 0; 2467 2468 /* 2469 * If there aren't any algorithms registered, return an empty message. 2470 * spdsock_get_ext() knows how to deal with this. 2471 */ 2472 if (act->spd_actions_count == 0) { 2473 act->spd_actions_len = 0; 2474 mutex_exit(&ipss->ipsec_alg_lock); 2475 goto error; 2476 } 2477 2478 attr = (struct spd_attribute *)cur; 2479 2480 #define EMIT(tag, value) { \ 2481 attr->spd_attr_tag = (tag); \ 2482 attr->spd_attr_value = (value); \ 2483 attr++; \ 2484 } 2485 2486 for (algtype = 0; algtype < IPSEC_NALGTYPES; algtype++) { 2487 for (algidx = 0; algidx < ipss->ipsec_nalgs[algtype]; 2488 algidx++) { 2489 2490 algid = ipss->ipsec_sortlist[algtype][algidx]; 2491 alg = ipss->ipsec_alglists[algtype][algid]; 2492 2493 /* 2494 * If you change the number of EMIT's here, change 2495 * ATTRPERALG above to match 2496 */ 2497 EMIT(SPD_ATTR_ALG_ID, algid); 2498 EMIT(SPD_ATTR_ALG_PROTO, algproto[algtype]); 2499 EMIT(SPD_ATTR_ALG_INCRBITS, alg->alg_increment); 2500 EMIT(SPD_ATTR_ALG_NKEYSIZES, alg->alg_nkey_sizes); 2501 for (i = 0; i < alg->alg_nkey_sizes; i++) 2502 EMIT(SPD_ATTR_ALG_KEYSIZE, 2503 alg->alg_key_sizes[i]); 2504 2505 EMIT(SPD_ATTR_ALG_NBLOCKSIZES, alg->alg_nblock_sizes); 2506 for (i = 0; i < alg->alg_nblock_sizes; i++) 2507 EMIT(SPD_ATTR_ALG_BLOCKSIZE, 2508 alg->alg_block_sizes[i]); 2509 2510 EMIT(SPD_ATTR_ALG_NPARAMS, alg->alg_nparams); 2511 for (i = 0; i < alg->alg_nparams; i++) 2512 EMIT(SPD_ATTR_ALG_PARAMS, 2513 alg->alg_params[i]); 2514 2515 EMIT(SPD_ATTR_ALG_FLAGS, alg->alg_flags); 2516 2517 EMIT(SPD_ATTR_ALG_MECHNAME, CRYPTO_MAX_MECH_NAME); 2518 bcopy(alg->alg_mech_name, attr, CRYPTO_MAX_MECH_NAME); 2519 attr = (struct spd_attribute *)((char *)attr + 2520 CRYPTO_MAX_MECH_NAME); 2521 2522 EMIT(SPD_ATTR_NEXT, 0); 2523 } 2524 } 2525 2526 mutex_exit(&ipss->ipsec_alg_lock); 2527 2528 #undef EMITALGATTRS 2529 #undef EMIT 2530 #undef ATTRPERALG 2531 2532 attr--; 2533 attr->spd_attr_tag = SPD_ATTR_END; 2534 2535 error: 2536 freemsg(mp); 2537 qreply(q, m); 2538 } 2539 2540 /* 2541 * Do the actual work of processing an SPD_UPDATEALGS request. Can 2542 * be invoked either once IPsec is loaded on a cached request, or 2543 * when a request is received while IPsec is loaded. 2544 */ 2545 static int 2546 spdsock_do_updatealg(spd_ext_t *extv[], spd_stack_t *spds) 2547 { 2548 struct spd_ext_actions *actp; 2549 struct spd_attribute *attr, *endattr; 2550 uint64_t *start, *end; 2551 ipsec_alginfo_t *alg = NULL; 2552 ipsec_algtype_t alg_type = 0; 2553 boolean_t skip_alg = B_TRUE, doing_proto = B_FALSE; 2554 uint_t i, cur_key, cur_block, algid; 2555 int diag = -1; 2556 2557 ASSERT(MUTEX_HELD(&spds->spds_alg_lock)); 2558 2559 /* parse the message, building the list of algorithms */ 2560 2561 actp = (struct spd_ext_actions *)extv[SPD_EXT_ACTION]; 2562 if (actp == NULL) 2563 return (SPD_DIAGNOSTIC_NO_ACTION_EXT); 2564 2565 start = (uint64_t *)actp; 2566 end = (start + actp->spd_actions_len); 2567 endattr = (struct spd_attribute *)end; 2568 attr = (struct spd_attribute *)&actp[1]; 2569 2570 bzero(spds->spds_algs, IPSEC_NALGTYPES * IPSEC_MAX_ALGS * 2571 sizeof (ipsec_alginfo_t *)); 2572 2573 alg = kmem_zalloc(sizeof (*alg), KM_SLEEP); 2574 2575 #define ALG_KEY_SIZES(a) (((a)->alg_nkey_sizes + 1) * sizeof (uint16_t)) 2576 #define ALG_BLOCK_SIZES(a) (((a)->alg_nblock_sizes + 1) * sizeof (uint16_t)) 2577 #define ALG_PARAM_SIZES(a) (((a)->alg_nparams + 1) * sizeof (uint16_t)) 2578 2579 while (attr < endattr) { 2580 switch (attr->spd_attr_tag) { 2581 case SPD_ATTR_NOP: 2582 case SPD_ATTR_EMPTY: 2583 break; 2584 case SPD_ATTR_END: 2585 attr = endattr; 2586 /* FALLTHRU */ 2587 case SPD_ATTR_NEXT: 2588 if (doing_proto) { 2589 doing_proto = B_FALSE; 2590 break; 2591 } 2592 if (skip_alg) { 2593 ipsec_alg_free(alg); 2594 } else { 2595 ipsec_alg_free( 2596 spds->spds_algs[alg_type][alg->alg_id]); 2597 spds->spds_algs[alg_type][alg->alg_id] = 2598 alg; 2599 } 2600 alg = kmem_zalloc(sizeof (*alg), KM_SLEEP); 2601 break; 2602 2603 case SPD_ATTR_ALG_ID: 2604 if (attr->spd_attr_value >= IPSEC_MAX_ALGS) { 2605 ss1dbg(spds, ("spdsock_do_updatealg: " 2606 "invalid alg id %d\n", 2607 attr->spd_attr_value)); 2608 diag = SPD_DIAGNOSTIC_ALG_ID_RANGE; 2609 goto bail; 2610 } 2611 alg->alg_id = attr->spd_attr_value; 2612 break; 2613 2614 case SPD_ATTR_ALG_PROTO: 2615 /* find the alg type */ 2616 for (i = 0; i < NALGPROTOS; i++) 2617 if (algproto[i] == attr->spd_attr_value) 2618 break; 2619 skip_alg = (i == NALGPROTOS); 2620 if (!skip_alg) 2621 alg_type = i; 2622 break; 2623 2624 case SPD_ATTR_ALG_INCRBITS: 2625 alg->alg_increment = attr->spd_attr_value; 2626 break; 2627 2628 case SPD_ATTR_ALG_NKEYSIZES: 2629 if (alg->alg_key_sizes != NULL) { 2630 kmem_free(alg->alg_key_sizes, 2631 ALG_KEY_SIZES(alg)); 2632 } 2633 alg->alg_nkey_sizes = attr->spd_attr_value; 2634 /* 2635 * Allocate room for the trailing zero key size 2636 * value as well. 2637 */ 2638 alg->alg_key_sizes = kmem_zalloc(ALG_KEY_SIZES(alg), 2639 KM_SLEEP); 2640 cur_key = 0; 2641 break; 2642 2643 case SPD_ATTR_ALG_KEYSIZE: 2644 if (alg->alg_key_sizes == NULL || 2645 cur_key >= alg->alg_nkey_sizes) { 2646 ss1dbg(spds, ("spdsock_do_updatealg: " 2647 "too many key sizes\n")); 2648 diag = SPD_DIAGNOSTIC_ALG_NUM_KEY_SIZES; 2649 goto bail; 2650 } 2651 alg->alg_key_sizes[cur_key++] = attr->spd_attr_value; 2652 break; 2653 2654 case SPD_ATTR_ALG_FLAGS: 2655 /* 2656 * Flags (bit mask). The alg_flags element of 2657 * ipsecalg_flags_t is only 8 bits wide. The 2658 * user can set the VALID bit, but we will ignore it 2659 * and make the decision is the algorithm is valid. 2660 */ 2661 alg->alg_flags |= (uint8_t)attr->spd_attr_value; 2662 break; 2663 2664 case SPD_ATTR_ALG_NBLOCKSIZES: 2665 if (alg->alg_block_sizes != NULL) { 2666 kmem_free(alg->alg_block_sizes, 2667 ALG_BLOCK_SIZES(alg)); 2668 } 2669 alg->alg_nblock_sizes = attr->spd_attr_value; 2670 /* 2671 * Allocate room for the trailing zero block size 2672 * value as well. 2673 */ 2674 alg->alg_block_sizes = kmem_zalloc(ALG_BLOCK_SIZES(alg), 2675 KM_SLEEP); 2676 cur_block = 0; 2677 break; 2678 2679 case SPD_ATTR_ALG_BLOCKSIZE: 2680 if (alg->alg_block_sizes == NULL || 2681 cur_block >= alg->alg_nblock_sizes) { 2682 ss1dbg(spds, ("spdsock_do_updatealg: " 2683 "too many block sizes\n")); 2684 diag = SPD_DIAGNOSTIC_ALG_NUM_BLOCK_SIZES; 2685 goto bail; 2686 } 2687 alg->alg_block_sizes[cur_block++] = 2688 attr->spd_attr_value; 2689 break; 2690 2691 case SPD_ATTR_ALG_NPARAMS: 2692 if (alg->alg_params != NULL) { 2693 kmem_free(alg->alg_params, 2694 ALG_PARAM_SIZES(alg)); 2695 } 2696 alg->alg_nparams = attr->spd_attr_value; 2697 /* 2698 * Allocate room for the trailing zero block size 2699 * value as well. 2700 */ 2701 alg->alg_params = kmem_zalloc(ALG_PARAM_SIZES(alg), 2702 KM_SLEEP); 2703 cur_block = 0; 2704 break; 2705 2706 case SPD_ATTR_ALG_PARAMS: 2707 if (alg->alg_params == NULL || 2708 cur_block >= alg->alg_nparams) { 2709 ss1dbg(spds, ("spdsock_do_updatealg: " 2710 "too many params\n")); 2711 diag = SPD_DIAGNOSTIC_ALG_NUM_BLOCK_SIZES; 2712 goto bail; 2713 } 2714 /* 2715 * Array contains: iv_len, icv_len, salt_len 2716 * Any additional parameters are currently ignored. 2717 */ 2718 alg->alg_params[cur_block++] = 2719 attr->spd_attr_value; 2720 break; 2721 2722 case SPD_ATTR_ALG_MECHNAME: { 2723 char *mech_name; 2724 2725 if (attr->spd_attr_value > CRYPTO_MAX_MECH_NAME) { 2726 ss1dbg(spds, ("spdsock_do_updatealg: " 2727 "mech name too long\n")); 2728 diag = SPD_DIAGNOSTIC_ALG_MECH_NAME_LEN; 2729 goto bail; 2730 } 2731 mech_name = (char *)(attr + 1); 2732 bcopy(mech_name, alg->alg_mech_name, 2733 attr->spd_attr_value); 2734 alg->alg_mech_name[CRYPTO_MAX_MECH_NAME-1] = '\0'; 2735 attr = (struct spd_attribute *)((char *)attr + 2736 attr->spd_attr_value); 2737 break; 2738 } 2739 2740 case SPD_ATTR_PROTO_ID: 2741 doing_proto = B_TRUE; 2742 for (i = 0; i < NALGPROTOS; i++) { 2743 if (algproto[i] == attr->spd_attr_value) { 2744 alg_type = i; 2745 break; 2746 } 2747 } 2748 break; 2749 2750 case SPD_ATTR_PROTO_EXEC_MODE: 2751 if (!doing_proto) 2752 break; 2753 for (i = 0; i < NEXECMODES; i++) { 2754 if (execmodes[i] == attr->spd_attr_value) { 2755 spds->spds_algs_exec_mode[alg_type] = i; 2756 break; 2757 } 2758 } 2759 break; 2760 } 2761 attr++; 2762 } 2763 2764 #undef ALG_KEY_SIZES 2765 #undef ALG_BLOCK_SIZES 2766 #undef ALG_PARAM_SIZES 2767 2768 /* update the algorithm tables */ 2769 spdsock_merge_algs(spds); 2770 bail: 2771 /* cleanup */ 2772 ipsec_alg_free(alg); 2773 for (alg_type = 0; alg_type < IPSEC_NALGTYPES; alg_type++) 2774 for (algid = 0; algid < IPSEC_MAX_ALGS; algid++) 2775 if (spds->spds_algs[alg_type][algid] != NULL) 2776 ipsec_alg_free(spds->spds_algs[alg_type][algid]); 2777 return (diag); 2778 } 2779 2780 /* 2781 * Process an SPD_UPDATEALGS request. If IPsec is not loaded, queue 2782 * the request until IPsec loads. If IPsec is loaded, act on it 2783 * immediately. 2784 */ 2785 2786 static void 2787 spdsock_updatealg(queue_t *q, mblk_t *mp, spd_ext_t *extv[]) 2788 { 2789 spdsock_t *ss = (spdsock_t *)q->q_ptr; 2790 spd_stack_t *spds = ss->spdsock_spds; 2791 ipsec_stack_t *ipss = spds->spds_netstack->netstack_ipsec; 2792 uint32_t auditing = AU_AUDITING(); 2793 2794 if (!ipsec_loaded(ipss)) { 2795 /* 2796 * IPsec is not loaded, save request and return nicely, 2797 * the message will be processed once IPsec loads. 2798 */ 2799 mblk_t *new_mp; 2800 2801 /* last update message wins */ 2802 if ((new_mp = copymsg(mp)) == NULL) { 2803 spdsock_error(q, mp, ENOMEM, 0); 2804 return; 2805 } 2806 mutex_enter(&spds->spds_alg_lock); 2807 bcopy(extv, spds->spds_extv_algs, 2808 sizeof (spd_ext_t *) * (SPD_EXT_MAX + 1)); 2809 if (spds->spds_mp_algs != NULL) 2810 freemsg(spds->spds_mp_algs); 2811 spds->spds_mp_algs = mp; 2812 mutex_exit(&spds->spds_alg_lock); 2813 if (auditing) { 2814 cred_t *cr; 2815 pid_t cpid; 2816 2817 cr = msg_getcred(mp, &cpid); 2818 audit_pf_policy(SPD_UPDATEALGS, cr, 2819 spds->spds_netstack, NULL, B_TRUE, EAGAIN, 2820 cpid); 2821 } 2822 spd_echo(q, new_mp); 2823 } else { 2824 /* 2825 * IPsec is loaded, act on the message immediately. 2826 */ 2827 int diag; 2828 2829 mutex_enter(&spds->spds_alg_lock); 2830 diag = spdsock_do_updatealg(extv, spds); 2831 if (diag == -1) { 2832 /* Keep the lock held while we walk the SA tables. */ 2833 sadb_alg_update(IPSEC_ALG_ALL, 0, 0, 2834 spds->spds_netstack); 2835 mutex_exit(&spds->spds_alg_lock); 2836 spd_echo(q, mp); 2837 if (auditing) { 2838 cred_t *cr; 2839 pid_t cpid; 2840 2841 cr = msg_getcred(mp, &cpid); 2842 audit_pf_policy(SPD_UPDATEALGS, cr, 2843 spds->spds_netstack, NULL, B_TRUE, 0, 2844 cpid); 2845 } 2846 } else { 2847 mutex_exit(&spds->spds_alg_lock); 2848 spdsock_diag(q, mp, diag); 2849 if (auditing) { 2850 cred_t *cr; 2851 pid_t cpid; 2852 2853 cr = msg_getcred(mp, &cpid); 2854 audit_pf_policy(SPD_UPDATEALGS, cr, 2855 spds->spds_netstack, NULL, B_TRUE, diag, 2856 cpid); 2857 } 2858 } 2859 } 2860 } 2861 2862 /* 2863 * Find a tunnel instance (using the name to link ID mapping), and 2864 * update it after an IPsec change. We need to do this always in case 2865 * we add policy AFTER plumbing a tunnel. We also need to do this 2866 * because, as a side-effect, the tunnel's MTU is updated to reflect 2867 * any IPsec overhead in the itp's policy. 2868 */ 2869 static void 2870 update_iptun_policy(ipsec_tun_pol_t *itp) 2871 { 2872 datalink_id_t linkid; 2873 2874 if (dls_mgmt_get_linkid(itp->itp_name, &linkid) == 0) 2875 iptun_set_policy(linkid, itp); 2876 } 2877 2878 /* 2879 * Sort through the mess of polhead options to retrieve an appropriate one. 2880 * Returns NULL if we send an spdsock error. Returns a valid pointer if we 2881 * found a valid polhead. Returns ALL_ACTIVE_POLHEADS (aka. -1) or 2882 * ALL_INACTIVE_POLHEADS (aka. -2) if the operation calls for the operation to 2883 * act on ALL policy heads. 2884 */ 2885 static ipsec_policy_head_t * 2886 get_appropriate_polhead(queue_t *q, mblk_t *mp, spd_if_t *tunname, int spdid, 2887 int msgtype, ipsec_tun_pol_t **itpp) 2888 { 2889 ipsec_tun_pol_t *itp; 2890 ipsec_policy_head_t *iph; 2891 int errno; 2892 char *tname; 2893 boolean_t active; 2894 spdsock_t *ss = (spdsock_t *)q->q_ptr; 2895 netstack_t *ns = ss->spdsock_spds->spds_netstack; 2896 uint64_t gen; /* Placeholder */ 2897 2898 active = (spdid == SPD_ACTIVE); 2899 *itpp = NULL; 2900 if (!active && spdid != SPD_STANDBY) { 2901 spdsock_diag(q, mp, SPD_DIAGNOSTIC_BAD_SPDID); 2902 return (NULL); 2903 } 2904 2905 if (tunname != NULL) { 2906 /* Acting on a tunnel's SPD. */ 2907 tname = (char *)tunname->spd_if_name; 2908 if (*tname == '\0') { 2909 /* Handle all-polhead cases here. */ 2910 if (msgtype != SPD_FLUSH && msgtype != SPD_DUMP) { 2911 spdsock_diag(q, mp, 2912 SPD_DIAGNOSTIC_NOT_GLOBAL_OP); 2913 return (NULL); 2914 } 2915 return (active ? ALL_ACTIVE_POLHEADS : 2916 ALL_INACTIVE_POLHEADS); 2917 } 2918 2919 itp = get_tunnel_policy(tname, ns); 2920 if (itp == NULL) { 2921 if (msgtype != SPD_ADDRULE) { 2922 /* "Tunnel not found" */ 2923 spdsock_error(q, mp, ENOENT, 0); 2924 return (NULL); 2925 } 2926 2927 errno = 0; 2928 itp = create_tunnel_policy(tname, &errno, &gen, ns); 2929 if (itp == NULL) { 2930 /* 2931 * Something very bad happened, most likely 2932 * ENOMEM. Return an indicator. 2933 */ 2934 spdsock_error(q, mp, errno, 0); 2935 return (NULL); 2936 } 2937 } 2938 2939 /* Match up the itp to an iptun instance. */ 2940 update_iptun_policy(itp); 2941 2942 *itpp = itp; 2943 /* For spdsock dump state, set the polhead's name. */ 2944 if (msgtype == SPD_DUMP) { 2945 ITP_REFHOLD(itp); 2946 ss->spdsock_itp = itp; 2947 ss->spdsock_dump_tunnel = itp->itp_flags & 2948 (active ? ITPF_P_TUNNEL : ITPF_I_TUNNEL); 2949 } 2950 } else { 2951 itp = NULL; 2952 /* For spdsock dump state, indicate it's global policy. */ 2953 if (msgtype == SPD_DUMP) 2954 ss->spdsock_itp = NULL; 2955 } 2956 2957 if (active) 2958 iph = (itp == NULL) ? ipsec_system_policy(ns) : itp->itp_policy; 2959 else 2960 iph = (itp == NULL) ? ipsec_inactive_policy(ns) : 2961 itp->itp_inactive; 2962 2963 ASSERT(iph != NULL); 2964 if (itp != NULL) { 2965 IPPH_REFHOLD(iph); 2966 } 2967 2968 return (iph); 2969 } 2970 2971 static void 2972 spdsock_parse(queue_t *q, mblk_t *mp) 2973 { 2974 spd_msg_t *spmsg; 2975 spd_ext_t *extv[SPD_EXT_MAX + 1]; 2976 uint_t msgsize; 2977 ipsec_policy_head_t *iph; 2978 ipsec_tun_pol_t *itp; 2979 spd_if_t *tunname; 2980 spdsock_t *ss = (spdsock_t *)q->q_ptr; 2981 spd_stack_t *spds = ss->spdsock_spds; 2982 netstack_t *ns = spds->spds_netstack; 2983 ipsec_stack_t *ipss = ns->netstack_ipsec; 2984 2985 /* Make sure nothing's below me. */ 2986 ASSERT(WR(q)->q_next == NULL); 2987 2988 spmsg = (spd_msg_t *)mp->b_rptr; 2989 2990 msgsize = SPD_64TO8(spmsg->spd_msg_len); 2991 2992 if (msgdsize(mp) != msgsize) { 2993 /* 2994 * Message len incorrect w.r.t. actual size. Send an error 2995 * (EMSGSIZE). It may be necessary to massage things a 2996 * bit. For example, if the spd_msg_type is hosed, 2997 * I need to set it to SPD_RESERVED to get delivery to 2998 * do the right thing. Then again, maybe just letting 2999 * the error delivery do the right thing. 3000 */ 3001 ss2dbg(spds, 3002 ("mblk (%lu) and base (%d) message sizes don't jibe.\n", 3003 msgdsize(mp), msgsize)); 3004 spdsock_error(q, mp, EMSGSIZE, SPD_DIAGNOSTIC_NONE); 3005 return; 3006 } 3007 3008 if (msgsize > (uint_t)(mp->b_wptr - mp->b_rptr)) { 3009 /* Get all message into one mblk. */ 3010 if (pullupmsg(mp, -1) == 0) { 3011 /* 3012 * Something screwy happened. 3013 */ 3014 ss3dbg(spds, ("spdsock_parse: pullupmsg() failed.\n")); 3015 return; 3016 } else { 3017 spmsg = (spd_msg_t *)mp->b_rptr; 3018 } 3019 } 3020 3021 switch (spdsock_get_ext(extv, spmsg, msgsize)) { 3022 case KGE_DUP: 3023 /* Handle duplicate extension. */ 3024 ss1dbg(spds, ("Got duplicate extension of type %d.\n", 3025 extv[0]->spd_ext_type)); 3026 spdsock_diag(q, mp, dup_ext_diag[extv[0]->spd_ext_type]); 3027 return; 3028 case KGE_UNK: 3029 /* Handle unknown extension. */ 3030 ss1dbg(spds, ("Got unknown extension of type %d.\n", 3031 extv[0]->spd_ext_type)); 3032 spdsock_diag(q, mp, SPD_DIAGNOSTIC_UNKNOWN_EXT); 3033 return; 3034 case KGE_LEN: 3035 /* Length error. */ 3036 ss1dbg(spds, ("Length %d on extension type %d overrun or 0.\n", 3037 extv[0]->spd_ext_len, extv[0]->spd_ext_type)); 3038 spdsock_diag(q, mp, SPD_DIAGNOSTIC_BAD_EXTLEN); 3039 return; 3040 case KGE_CHK: 3041 /* Reality check failed. */ 3042 ss1dbg(spds, ("Reality check failed on extension type %d.\n", 3043 extv[0]->spd_ext_type)); 3044 spdsock_diag(q, mp, bad_ext_diag[extv[0]->spd_ext_type]); 3045 return; 3046 default: 3047 /* Default case is no errors. */ 3048 break; 3049 } 3050 3051 /* 3052 * Special-case SPD_UPDATEALGS so as not to load IPsec. 3053 */ 3054 if (!ipsec_loaded(ipss) && spmsg->spd_msg_type != SPD_UPDATEALGS) { 3055 spdsock_t *ss = (spdsock_t *)q->q_ptr; 3056 3057 ASSERT(ss != NULL); 3058 ipsec_loader_loadnow(ipss); 3059 ss->spdsock_timeout_arg = mp; 3060 ss->spdsock_timeout = qtimeout(q, spdsock_loadcheck, 3061 q, LOADCHECK_INTERVAL); 3062 return; 3063 } 3064 3065 /* First check for messages that need no polheads at all. */ 3066 switch (spmsg->spd_msg_type) { 3067 case SPD_UPDATEALGS: 3068 spdsock_updatealg(q, mp, extv); 3069 return; 3070 case SPD_ALGLIST: 3071 spdsock_alglist(q, mp); 3072 return; 3073 case SPD_DUMPALGS: 3074 spdsock_dumpalgs(q, mp); 3075 return; 3076 } 3077 3078 /* 3079 * Then check for ones that need both primary/secondary polheads, 3080 * finding the appropriate tunnel policy if need be. 3081 */ 3082 tunname = (spd_if_t *)extv[SPD_EXT_TUN_NAME]; 3083 switch (spmsg->spd_msg_type) { 3084 case SPD_FLIP: 3085 spdsock_flip(q, mp, tunname); 3086 return; 3087 case SPD_CLONE: 3088 spdsock_clone(q, mp, tunname); 3089 return; 3090 } 3091 3092 /* 3093 * Finally, find ones that operate on exactly one polhead, or 3094 * "all polheads" of a given type (active/inactive). 3095 */ 3096 iph = get_appropriate_polhead(q, mp, tunname, spmsg->spd_msg_spdid, 3097 spmsg->spd_msg_type, &itp); 3098 if (iph == NULL) 3099 return; 3100 3101 /* All-polheads-ready operations. */ 3102 switch (spmsg->spd_msg_type) { 3103 case SPD_FLUSH: 3104 if (itp != NULL) { 3105 mutex_enter(&itp->itp_lock); 3106 if (spmsg->spd_msg_spdid == SPD_ACTIVE) 3107 itp->itp_flags &= ~ITPF_PFLAGS; 3108 else 3109 itp->itp_flags &= ~ITPF_IFLAGS; 3110 mutex_exit(&itp->itp_lock); 3111 } 3112 3113 spdsock_flush(q, iph, itp, mp); 3114 3115 if (itp != NULL) { 3116 /* SPD_FLUSH is worth a tunnel MTU check. */ 3117 update_iptun_policy(itp); 3118 ITP_REFRELE(itp, ns); 3119 } 3120 return; 3121 case SPD_DUMP: 3122 if (itp != NULL) 3123 ITP_REFRELE(itp, ns); 3124 spdsock_dump(q, iph, mp); 3125 return; 3126 } 3127 3128 if (iph == ALL_ACTIVE_POLHEADS || iph == ALL_INACTIVE_POLHEADS) { 3129 spdsock_diag(q, mp, SPD_DIAGNOSTIC_NOT_GLOBAL_OP); 3130 return; 3131 } 3132 3133 /* Single-polhead-only operations. */ 3134 switch (spmsg->spd_msg_type) { 3135 case SPD_ADDRULE: 3136 spdsock_addrule(q, iph, mp, extv, itp); 3137 break; 3138 case SPD_DELETERULE: 3139 spdsock_deleterule(q, iph, mp, extv, itp); 3140 break; 3141 case SPD_LOOKUP: 3142 spdsock_lookup(q, iph, mp, extv, itp); 3143 break; 3144 default: 3145 spdsock_diag(q, mp, SPD_DIAGNOSTIC_BAD_MSG_TYPE); 3146 break; 3147 } 3148 3149 IPPH_REFRELE(iph, ns); 3150 if (itp != NULL) { 3151 /* SPD_{ADD,DELETE}RULE are worth a tunnel MTU check. */ 3152 if (spmsg->spd_msg_type == SPD_ADDRULE || 3153 spmsg->spd_msg_type == SPD_DELETERULE) 3154 update_iptun_policy(itp); 3155 ITP_REFRELE(itp, ns); 3156 } 3157 } 3158 3159 /* 3160 * If an algorithm mapping was received before IPsec was loaded, process it. 3161 * Called from the IPsec loader. 3162 */ 3163 void 3164 spdsock_update_pending_algs(netstack_t *ns) 3165 { 3166 spd_stack_t *spds = ns->netstack_spdsock; 3167 3168 mutex_enter(&spds->spds_alg_lock); 3169 if (spds->spds_mp_algs != NULL) { 3170 (void) spdsock_do_updatealg(spds->spds_extv_algs, spds); 3171 freemsg(spds->spds_mp_algs); 3172 spds->spds_mp_algs = NULL; 3173 } 3174 mutex_exit(&spds->spds_alg_lock); 3175 } 3176 3177 static void 3178 spdsock_loadcheck(void *arg) 3179 { 3180 queue_t *q = (queue_t *)arg; 3181 spdsock_t *ss = (spdsock_t *)q->q_ptr; 3182 mblk_t *mp; 3183 ipsec_stack_t *ipss = ss->spdsock_spds->spds_netstack->netstack_ipsec; 3184 3185 ASSERT(ss != NULL); 3186 3187 ss->spdsock_timeout = 0; 3188 mp = ss->spdsock_timeout_arg; 3189 ASSERT(mp != NULL); 3190 ss->spdsock_timeout_arg = NULL; 3191 if (ipsec_failed(ipss)) 3192 spdsock_error(q, mp, EPROTONOSUPPORT, 0); 3193 else 3194 spdsock_parse(q, mp); 3195 } 3196 3197 /* 3198 * Copy relevant state bits. 3199 */ 3200 static void 3201 spdsock_copy_info(struct T_info_ack *tap, spdsock_t *ss) 3202 { 3203 *tap = spdsock_g_t_info_ack; 3204 tap->CURRENT_state = ss->spdsock_state; 3205 tap->OPT_size = spdsock_max_optsize; 3206 } 3207 3208 /* 3209 * This routine responds to T_CAPABILITY_REQ messages. It is called by 3210 * spdsock_wput. Much of the T_CAPABILITY_ACK information is copied from 3211 * spdsock_g_t_info_ack. The current state of the stream is copied from 3212 * spdsock_state. 3213 */ 3214 static void 3215 spdsock_capability_req(queue_t *q, mblk_t *mp) 3216 { 3217 spdsock_t *ss = (spdsock_t *)q->q_ptr; 3218 t_uscalar_t cap_bits1; 3219 struct T_capability_ack *tcap; 3220 3221 cap_bits1 = ((struct T_capability_req *)mp->b_rptr)->CAP_bits1; 3222 3223 mp = tpi_ack_alloc(mp, sizeof (struct T_capability_ack), 3224 mp->b_datap->db_type, T_CAPABILITY_ACK); 3225 if (mp == NULL) 3226 return; 3227 3228 tcap = (struct T_capability_ack *)mp->b_rptr; 3229 tcap->CAP_bits1 = 0; 3230 3231 if (cap_bits1 & TC1_INFO) { 3232 spdsock_copy_info(&tcap->INFO_ack, ss); 3233 tcap->CAP_bits1 |= TC1_INFO; 3234 } 3235 3236 qreply(q, mp); 3237 } 3238 3239 /* 3240 * This routine responds to T_INFO_REQ messages. It is called by 3241 * spdsock_wput_other. 3242 * Most of the T_INFO_ACK information is copied from spdsock_g_t_info_ack. 3243 * The current state of the stream is copied from spdsock_state. 3244 */ 3245 static void 3246 spdsock_info_req(q, mp) 3247 queue_t *q; 3248 mblk_t *mp; 3249 { 3250 mp = tpi_ack_alloc(mp, sizeof (struct T_info_ack), M_PCPROTO, 3251 T_INFO_ACK); 3252 if (mp == NULL) 3253 return; 3254 spdsock_copy_info((struct T_info_ack *)mp->b_rptr, 3255 (spdsock_t *)q->q_ptr); 3256 qreply(q, mp); 3257 } 3258 3259 /* 3260 * spdsock_err_ack. This routine creates a 3261 * T_ERROR_ACK message and passes it 3262 * upstream. 3263 */ 3264 static void 3265 spdsock_err_ack(q, mp, t_error, sys_error) 3266 queue_t *q; 3267 mblk_t *mp; 3268 int t_error; 3269 int sys_error; 3270 { 3271 if ((mp = mi_tpi_err_ack_alloc(mp, t_error, sys_error)) != NULL) 3272 qreply(q, mp); 3273 } 3274 3275 /* 3276 * This routine retrieves the current status of socket options. 3277 * It returns the size of the option retrieved. 3278 */ 3279 /* ARGSUSED */ 3280 int 3281 spdsock_opt_get(queue_t *q, int level, int name, uchar_t *ptr) 3282 { 3283 int *i1 = (int *)ptr; 3284 3285 switch (level) { 3286 case SOL_SOCKET: 3287 switch (name) { 3288 case SO_TYPE: 3289 *i1 = SOCK_RAW; 3290 break; 3291 /* 3292 * The following two items can be manipulated, 3293 * but changing them should do nothing. 3294 */ 3295 case SO_SNDBUF: 3296 *i1 = (int)q->q_hiwat; 3297 break; 3298 case SO_RCVBUF: 3299 *i1 = (int)(RD(q)->q_hiwat); 3300 break; 3301 } 3302 break; 3303 default: 3304 return (0); 3305 } 3306 return (sizeof (int)); 3307 } 3308 3309 /* 3310 * This routine sets socket options. 3311 */ 3312 /* ARGSUSED */ 3313 int 3314 spdsock_opt_set(queue_t *q, uint_t mgmt_flags, int level, int name, 3315 uint_t inlen, uchar_t *invalp, uint_t *outlenp, uchar_t *outvalp, 3316 void *thisdg_attrs, cred_t *cr) 3317 { 3318 int *i1 = (int *)invalp; 3319 spdsock_t *ss = (spdsock_t *)q->q_ptr; 3320 spd_stack_t *spds = ss->spdsock_spds; 3321 3322 switch (level) { 3323 case SOL_SOCKET: 3324 switch (name) { 3325 case SO_SNDBUF: 3326 if (*i1 > spds->spds_max_buf) 3327 return (ENOBUFS); 3328 q->q_hiwat = *i1; 3329 break; 3330 case SO_RCVBUF: 3331 if (*i1 > spds->spds_max_buf) 3332 return (ENOBUFS); 3333 RD(q)->q_hiwat = *i1; 3334 (void) proto_set_rx_hiwat(RD(q), NULL, *i1); 3335 break; 3336 } 3337 break; 3338 } 3339 return (0); 3340 } 3341 3342 3343 /* 3344 * Handle STREAMS messages. 3345 */ 3346 static void 3347 spdsock_wput_other(queue_t *q, mblk_t *mp) 3348 { 3349 struct iocblk *iocp; 3350 int error; 3351 spdsock_t *ss = (spdsock_t *)q->q_ptr; 3352 spd_stack_t *spds = ss->spdsock_spds; 3353 cred_t *cr; 3354 3355 switch (mp->b_datap->db_type) { 3356 case M_PROTO: 3357 case M_PCPROTO: 3358 if ((mp->b_wptr - mp->b_rptr) < sizeof (long)) { 3359 ss3dbg(spds, ( 3360 "spdsock_wput_other: Not big enough M_PROTO\n")); 3361 freemsg(mp); 3362 return; 3363 } 3364 switch (((union T_primitives *)mp->b_rptr)->type) { 3365 case T_CAPABILITY_REQ: 3366 spdsock_capability_req(q, mp); 3367 break; 3368 case T_INFO_REQ: 3369 spdsock_info_req(q, mp); 3370 break; 3371 case T_SVR4_OPTMGMT_REQ: 3372 case T_OPTMGMT_REQ: 3373 /* 3374 * All Solaris components should pass a db_credp 3375 * for this TPI message, hence we ASSERT. 3376 * But in case there is some other M_PROTO that looks 3377 * like a TPI message sent by some other kernel 3378 * component, we check and return an error. 3379 */ 3380 cr = msg_getcred(mp, NULL); 3381 ASSERT(cr != NULL); 3382 if (cr == NULL) { 3383 spdsock_err_ack(q, mp, TSYSERR, EINVAL); 3384 return; 3385 } 3386 if (((union T_primitives *)mp->b_rptr)->type == 3387 T_SVR4_OPTMGMT_REQ) { 3388 svr4_optcom_req(q, mp, cr, &spdsock_opt_obj); 3389 } else { 3390 tpi_optcom_req(q, mp, cr, &spdsock_opt_obj); 3391 } 3392 break; 3393 case T_DATA_REQ: 3394 case T_EXDATA_REQ: 3395 case T_ORDREL_REQ: 3396 /* Illegal for spdsock. */ 3397 freemsg(mp); 3398 (void) putnextctl1(RD(q), M_ERROR, EPROTO); 3399 break; 3400 default: 3401 /* Not supported by spdsock. */ 3402 spdsock_err_ack(q, mp, TNOTSUPPORT, 0); 3403 break; 3404 } 3405 return; 3406 case M_IOCTL: 3407 iocp = (struct iocblk *)mp->b_rptr; 3408 error = EINVAL; 3409 3410 switch (iocp->ioc_cmd) { 3411 case ND_SET: 3412 case ND_GET: 3413 if (nd_getset(q, spds->spds_g_nd, mp)) { 3414 qreply(q, mp); 3415 return; 3416 } else 3417 error = ENOENT; 3418 /* FALLTHRU */ 3419 default: 3420 miocnak(q, mp, 0, error); 3421 return; 3422 } 3423 case M_FLUSH: 3424 if (*mp->b_rptr & FLUSHW) { 3425 flushq(q, FLUSHALL); 3426 *mp->b_rptr &= ~FLUSHW; 3427 } 3428 if (*mp->b_rptr & FLUSHR) { 3429 qreply(q, mp); 3430 return; 3431 } 3432 /* Else FALLTHRU */ 3433 } 3434 3435 /* If fell through, just black-hole the message. */ 3436 freemsg(mp); 3437 } 3438 3439 static void 3440 spdsock_wput(queue_t *q, mblk_t *mp) 3441 { 3442 uint8_t *rptr = mp->b_rptr; 3443 mblk_t *mp1; 3444 spdsock_t *ss = (spdsock_t *)q->q_ptr; 3445 spd_stack_t *spds = ss->spdsock_spds; 3446 3447 /* 3448 * If we're dumping, defer processing other messages until the 3449 * dump completes. 3450 */ 3451 if (ss->spdsock_dump_req != NULL) { 3452 if (!putq(q, mp)) 3453 freemsg(mp); 3454 return; 3455 } 3456 3457 switch (mp->b_datap->db_type) { 3458 case M_DATA: 3459 /* 3460 * Silently discard. 3461 */ 3462 ss2dbg(spds, ("raw M_DATA in spdsock.\n")); 3463 freemsg(mp); 3464 return; 3465 case M_PROTO: 3466 case M_PCPROTO: 3467 if ((mp->b_wptr - rptr) >= sizeof (struct T_data_req)) { 3468 if (((union T_primitives *)rptr)->type == T_DATA_REQ) { 3469 if ((mp1 = mp->b_cont) == NULL) { 3470 /* No data after T_DATA_REQ. */ 3471 ss2dbg(spds, 3472 ("No data after DATA_REQ.\n")); 3473 freemsg(mp); 3474 return; 3475 } 3476 freeb(mp); 3477 mp = mp1; 3478 ss2dbg(spds, ("T_DATA_REQ\n")); 3479 break; /* Out of switch. */ 3480 } 3481 } 3482 /* FALLTHRU */ 3483 default: 3484 ss3dbg(spds, ("In default wput case (%d %d).\n", 3485 mp->b_datap->db_type, ((union T_primitives *)rptr)->type)); 3486 spdsock_wput_other(q, mp); 3487 return; 3488 } 3489 3490 /* I now have a PF_POLICY message in an M_DATA block. */ 3491 spdsock_parse(q, mp); 3492 } 3493 3494 /* 3495 * Device open procedure, called when new queue pair created. 3496 * We are passed the read-side queue. 3497 */ 3498 /* ARGSUSED */ 3499 static int 3500 spdsock_open(queue_t *q, dev_t *devp, int flag, int sflag, cred_t *credp) 3501 { 3502 spdsock_t *ss; 3503 queue_t *oq = OTHERQ(q); 3504 minor_t ssminor; 3505 netstack_t *ns; 3506 spd_stack_t *spds; 3507 3508 if (secpolicy_ip_config(credp, B_FALSE) != 0) 3509 return (EPERM); 3510 3511 if (q->q_ptr != NULL) 3512 return (0); /* Re-open of an already open instance. */ 3513 3514 if (sflag & MODOPEN) 3515 return (EINVAL); 3516 3517 ns = netstack_find_by_cred(credp); 3518 ASSERT(ns != NULL); 3519 spds = ns->netstack_spdsock; 3520 ASSERT(spds != NULL); 3521 3522 ss2dbg(spds, ("Made it into PF_POLICY socket open.\n")); 3523 3524 ssminor = (minor_t)(uintptr_t)vmem_alloc(spdsock_vmem, 1, VM_NOSLEEP); 3525 if (ssminor == 0) { 3526 netstack_rele(spds->spds_netstack); 3527 return (ENOMEM); 3528 } 3529 ss = kmem_zalloc(sizeof (spdsock_t), KM_NOSLEEP); 3530 if (ss == NULL) { 3531 vmem_free(spdsock_vmem, (void *)(uintptr_t)ssminor, 1); 3532 netstack_rele(spds->spds_netstack); 3533 return (ENOMEM); 3534 } 3535 3536 ss->spdsock_minor = ssminor; 3537 ss->spdsock_state = TS_UNBND; 3538 ss->spdsock_dump_req = NULL; 3539 3540 ss->spdsock_spds = spds; 3541 3542 q->q_ptr = ss; 3543 oq->q_ptr = ss; 3544 3545 q->q_hiwat = spds->spds_recv_hiwat; 3546 3547 oq->q_hiwat = spds->spds_xmit_hiwat; 3548 oq->q_lowat = spds->spds_xmit_lowat; 3549 3550 qprocson(q); 3551 (void) proto_set_rx_hiwat(q, NULL, spds->spds_recv_hiwat); 3552 3553 *devp = makedevice(getmajor(*devp), ss->spdsock_minor); 3554 return (0); 3555 } 3556 3557 /* 3558 * Read-side service procedure, invoked when we get back-enabled 3559 * when buffer space becomes available. 3560 * 3561 * Dump another chunk if we were dumping before; when we finish, kick 3562 * the write-side queue in case it's waiting for read queue space. 3563 */ 3564 void 3565 spdsock_rsrv(queue_t *q) 3566 { 3567 spdsock_t *ss = q->q_ptr; 3568 3569 if (ss->spdsock_dump_req != NULL) 3570 spdsock_dump_some(q, ss); 3571 3572 if (ss->spdsock_dump_req == NULL) 3573 qenable(OTHERQ(q)); 3574 } 3575 3576 /* 3577 * Write-side service procedure, invoked when we defer processing 3578 * if another message is received while a dump is in progress. 3579 */ 3580 void 3581 spdsock_wsrv(queue_t *q) 3582 { 3583 spdsock_t *ss = q->q_ptr; 3584 mblk_t *mp; 3585 ipsec_stack_t *ipss = ss->spdsock_spds->spds_netstack->netstack_ipsec; 3586 3587 if (ss->spdsock_dump_req != NULL) { 3588 qenable(OTHERQ(q)); 3589 return; 3590 } 3591 3592 while ((mp = getq(q)) != NULL) { 3593 if (ipsec_loaded(ipss)) { 3594 spdsock_wput(q, mp); 3595 if (ss->spdsock_dump_req != NULL) 3596 return; 3597 } else if (!ipsec_failed(ipss)) { 3598 (void) putq(q, mp); 3599 } else { 3600 spdsock_error(q, mp, EPFNOSUPPORT, 0); 3601 } 3602 } 3603 } 3604 3605 static int 3606 spdsock_close(queue_t *q) 3607 { 3608 spdsock_t *ss = q->q_ptr; 3609 spd_stack_t *spds = ss->spdsock_spds; 3610 3611 qprocsoff(q); 3612 3613 /* Safe assumption. */ 3614 ASSERT(ss != NULL); 3615 3616 if (ss->spdsock_timeout != 0) 3617 (void) quntimeout(q, ss->spdsock_timeout); 3618 3619 ss3dbg(spds, ("Driver close, PF_POLICY socket is going away.\n")); 3620 3621 vmem_free(spdsock_vmem, (void *)(uintptr_t)ss->spdsock_minor, 1); 3622 netstack_rele(ss->spdsock_spds->spds_netstack); 3623 3624 kmem_free(ss, sizeof (spdsock_t)); 3625 return (0); 3626 } 3627 3628 /* 3629 * Merge the IPsec algorithms tables with the received algorithm information. 3630 */ 3631 void 3632 spdsock_merge_algs(spd_stack_t *spds) 3633 { 3634 ipsec_alginfo_t *alg, *oalg; 3635 ipsec_algtype_t algtype; 3636 uint_t algidx, algid, nalgs; 3637 crypto_mech_name_t *mechs; 3638 uint_t mech_count, mech_idx; 3639 netstack_t *ns = spds->spds_netstack; 3640 ipsec_stack_t *ipss = ns->netstack_ipsec; 3641 3642 ASSERT(MUTEX_HELD(&spds->spds_alg_lock)); 3643 3644 /* 3645 * Get the list of supported mechanisms from the crypto framework. 3646 * If a mechanism is supported by KCF, resolve its mechanism 3647 * id and mark it as being valid. This operation must be done 3648 * without holding alg_lock, since it can cause a provider 3649 * module to be loaded and the provider notification callback to 3650 * be invoked. 3651 */ 3652 mechs = crypto_get_mech_list(&mech_count, KM_SLEEP); 3653 for (algtype = 0; algtype < IPSEC_NALGTYPES; algtype++) { 3654 for (algid = 0; algid < IPSEC_MAX_ALGS; algid++) { 3655 int algflags = 0; 3656 crypto_mech_type_t mt = CRYPTO_MECHANISM_INVALID; 3657 3658 alg = spds->spds_algs[algtype][algid]; 3659 if (alg == NULL) 3660 continue; 3661 3662 /* 3663 * The NULL encryption algorithm is a special 3664 * case because there are no mechanisms, yet 3665 * the algorithm is still valid. 3666 */ 3667 if (alg->alg_id == SADB_EALG_NULL) { 3668 alg->alg_mech_type = CRYPTO_MECHANISM_INVALID; 3669 alg->alg_flags |= ALG_FLAG_VALID; 3670 continue; 3671 } 3672 3673 for (mech_idx = 0; mech_idx < mech_count; mech_idx++) { 3674 if (strncmp(alg->alg_mech_name, mechs[mech_idx], 3675 CRYPTO_MAX_MECH_NAME) == 0) { 3676 mt = crypto_mech2id(alg->alg_mech_name); 3677 ASSERT(mt != CRYPTO_MECHANISM_INVALID); 3678 algflags = ALG_FLAG_VALID; 3679 break; 3680 } 3681 } 3682 alg->alg_mech_type = mt; 3683 alg->alg_flags |= algflags; 3684 } 3685 } 3686 3687 mutex_enter(&ipss->ipsec_alg_lock); 3688 3689 /* 3690 * For each algorithm currently defined, check if it is 3691 * present in the new tables created from the SPD_UPDATEALGS 3692 * message received from user-space. 3693 * Delete the algorithm entries that are currently defined 3694 * but not part of the new tables. 3695 */ 3696 for (algtype = 0; algtype < IPSEC_NALGTYPES; algtype++) { 3697 nalgs = ipss->ipsec_nalgs[algtype]; 3698 for (algidx = 0; algidx < nalgs; algidx++) { 3699 algid = ipss->ipsec_sortlist[algtype][algidx]; 3700 if (spds->spds_algs[algtype][algid] == NULL) 3701 ipsec_alg_unreg(algtype, algid, ns); 3702 } 3703 } 3704 3705 /* 3706 * For each algorithm we just received, check if it is 3707 * present in the currently defined tables. If it is, swap 3708 * the entry with the one we just allocated. 3709 * If the new algorithm is not in the current tables, 3710 * add it. 3711 */ 3712 for (algtype = 0; algtype < IPSEC_NALGTYPES; algtype++) { 3713 for (algid = 0; algid < IPSEC_MAX_ALGS; algid++) { 3714 alg = spds->spds_algs[algtype][algid]; 3715 if (alg == NULL) 3716 continue; 3717 3718 if ((oalg = ipss->ipsec_alglists[algtype][algid]) == 3719 NULL) { 3720 /* 3721 * New algorithm, add it to the algorithm 3722 * table. 3723 */ 3724 ipsec_alg_reg(algtype, alg, ns); 3725 } else { 3726 /* 3727 * Algorithm is already in the table. Swap 3728 * the existing entry with the new one. 3729 */ 3730 ipsec_alg_fix_min_max(alg, algtype, ns); 3731 ipss->ipsec_alglists[algtype][algid] = alg; 3732 ipsec_alg_free(oalg); 3733 } 3734 spds->spds_algs[algtype][algid] = NULL; 3735 } 3736 } 3737 3738 for (algtype = 0; algtype < IPSEC_NALGTYPES; algtype++) { 3739 ipss->ipsec_algs_exec_mode[algtype] = 3740 spds->spds_algs_exec_mode[algtype]; 3741 } 3742 3743 mutex_exit(&ipss->ipsec_alg_lock); 3744 3745 crypto_free_mech_list(mechs, mech_count); 3746 3747 ipsecah_algs_changed(ns); 3748 ipsecesp_algs_changed(ns); 3749 } 3750