1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2009 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 #include <sys/types.h> 28 #include <sys/conf.h> 29 #include <sys/time.h> 30 #include <sys/taskq.h> 31 #include <sys/cmn_err.h> 32 #include <sys/sdt.h> 33 #include <sys/atomic.h> 34 #include <netinet/in.h> 35 #include <inet/ip.h> 36 #include <inet/ip6.h> 37 #include <inet/tcp.h> 38 #include <inet/udp_impl.h> 39 #include <inet/ilb.h> 40 41 #include "ilb_stack.h" 42 #include "ilb_impl.h" 43 #include "ilb_conn.h" 44 #include "ilb_nat.h" 45 46 /* 47 * Timer struct for ilb_conn_t and ilb_sticky_t garbage collection 48 * 49 * start: starting index into the hash table to do gc 50 * end: ending index into the hash table to do gc 51 * ilbs: pointer to the ilb_stack_t of the IP stack 52 * tid_lock: mutex to protect the timer id. 53 * tid: timer id of the timer 54 */ 55 typedef struct ilb_timer_s { 56 uint32_t start; 57 uint32_t end; 58 ilb_stack_t *ilbs; 59 kmutex_t tid_lock; 60 timeout_id_t tid; 61 } ilb_timer_t; 62 63 /* Hash macro for finding the index to the conn hash table */ 64 #define ILB_CONN_HASH(saddr, sport, daddr, dport, hash_size) \ 65 (((*((saddr) + 3) ^ *((daddr) + 3)) * 50653 + \ 66 (*((saddr) + 2) ^ *((daddr) + 2)) * 1369 + \ 67 (*((saddr) + 1) ^ *((daddr) + 1)) * 37 + \ 68 (*(saddr) ^ *(daddr)) + (sport) * 37 + (dport)) & \ 69 ((hash_size) - 1)) 70 71 /* Kmem cache for the conn hash entry */ 72 static struct kmem_cache *ilb_conn_cache = NULL; 73 74 /* 75 * There are 60 timers running to do conn cache garbage collection. Each 76 * gc thread is responsible for 1/60 of the conn hash table. 77 */ 78 static int ilb_conn_timer_size = 60; 79 80 /* Each of the above gc timers wake up every 15s to do the gc. */ 81 static int ilb_conn_cache_timeout = 15; 82 83 #define ILB_STICKY_HASH(saddr, rule, hash_size) \ 84 (((*((saddr) + 3) ^ ((rule) >> 24)) * 29791 + \ 85 (*((saddr) + 2) ^ ((rule) >> 16)) * 961 + \ 86 (*((saddr) + 1) ^ ((rule) >> 8)) * 31 + \ 87 (*(saddr) ^ (rule))) & ((hash_size) - 1)) 88 89 static struct kmem_cache *ilb_sticky_cache = NULL; 90 91 /* 92 * There are 60 timers running to do sticky cache garbage collection. Each 93 * gc thread is responsible for 1/60 of the sticky hash table. 94 */ 95 static int ilb_sticky_timer_size = 60; 96 97 /* Each of the above gc timers wake up every 15s to do the gc. */ 98 static int ilb_sticky_timeout = 15; 99 100 #define ILB_STICKY_REFRELE(s) \ 101 { \ 102 mutex_enter(&(s)->hash->sticky_lock); \ 103 (s)->refcnt--; \ 104 (s)->atime = ddi_get_lbolt64(); \ 105 mutex_exit(&s->hash->sticky_lock); \ 106 } 107 108 109 static void 110 ilb_conn_cache_init(void) 111 { 112 ilb_conn_cache = kmem_cache_create("ilb_conn_cache", 113 sizeof (ilb_conn_t), 0, NULL, NULL, NULL, NULL, NULL, 114 ilb_kmem_flags); 115 } 116 117 void 118 ilb_conn_cache_fini(void) 119 { 120 if (ilb_conn_cache != NULL) { 121 kmem_cache_destroy(ilb_conn_cache); 122 ilb_conn_cache = NULL; 123 } 124 } 125 126 static void 127 ilb_conn_remove_common(ilb_conn_t *connp, boolean_t c2s) 128 { 129 ilb_conn_hash_t *hash; 130 ilb_conn_t **next, **prev; 131 ilb_conn_t **next_prev, **prev_next; 132 133 if (c2s) { 134 hash = connp->conn_c2s_hash; 135 ASSERT(MUTEX_HELD(&hash->ilb_conn_hash_lock)); 136 next = &connp->conn_c2s_next; 137 prev = &connp->conn_c2s_prev; 138 if (*next != NULL) 139 next_prev = &(*next)->conn_c2s_prev; 140 if (*prev != NULL) 141 prev_next = &(*prev)->conn_c2s_next; 142 } else { 143 hash = connp->conn_s2c_hash; 144 ASSERT(MUTEX_HELD(&hash->ilb_conn_hash_lock)); 145 next = &connp->conn_s2c_next; 146 prev = &connp->conn_s2c_prev; 147 if (*next != NULL) 148 next_prev = &(*next)->conn_s2c_prev; 149 if (*prev != NULL) 150 prev_next = &(*prev)->conn_s2c_next; 151 } 152 153 if (hash->ilb_connp == connp) { 154 hash->ilb_connp = *next; 155 if (*next != NULL) 156 *next_prev = NULL; 157 } else { 158 if (*prev != NULL) 159 *prev_next = *next; 160 if (*next != NULL) 161 *next_prev = *prev; 162 } 163 ASSERT(hash->ilb_conn_cnt > 0); 164 hash->ilb_conn_cnt--; 165 166 *next = NULL; 167 *prev = NULL; 168 } 169 170 static void 171 ilb_conn_remove(ilb_conn_t *connp) 172 { 173 ASSERT(MUTEX_HELD(&connp->conn_c2s_hash->ilb_conn_hash_lock)); 174 ilb_conn_remove_common(connp, B_TRUE); 175 ASSERT(MUTEX_HELD(&connp->conn_s2c_hash->ilb_conn_hash_lock)); 176 ilb_conn_remove_common(connp, B_FALSE); 177 178 if (connp->conn_rule_cache.topo == ILB_TOPO_IMPL_NAT) { 179 in_port_t port; 180 181 port = ntohs(connp->conn_rule_cache.info.nat_sport); 182 vmem_free(connp->conn_rule_cache.info.src_ent->nse_port_arena, 183 (void *)(uintptr_t)port, 1); 184 } 185 186 if (connp->conn_sticky != NULL) 187 ILB_STICKY_REFRELE(connp->conn_sticky); 188 ILB_SERVER_REFRELE(connp->conn_server); 189 kmem_cache_free(ilb_conn_cache, connp); 190 } 191 192 /* 193 * Routine to do periodic garbage collection of conn hash entries. When 194 * a conn hash timer fires, it dispatches a taskq to call this function 195 * to do the gc. Note that each taskq is responisble for a portion of 196 * the table. The portion is stored in timer->start, timer->end. 197 */ 198 static void 199 ilb_conn_cleanup(void *arg) 200 { 201 ilb_timer_t *timer = (ilb_timer_t *)arg; 202 uint32_t i; 203 ilb_stack_t *ilbs; 204 ilb_conn_hash_t *c2s_hash, *s2c_hash; 205 ilb_conn_t *connp, *nxt_connp; 206 int64_t now; 207 int64_t expiry; 208 boolean_t die_now; 209 210 ilbs = timer->ilbs; 211 c2s_hash = ilbs->ilbs_c2s_conn_hash; 212 ASSERT(c2s_hash != NULL); 213 214 now = ddi_get_lbolt64(); 215 for (i = timer->start; i < timer->end; i++) { 216 mutex_enter(&c2s_hash[i].ilb_conn_hash_lock); 217 if ((connp = c2s_hash[i].ilb_connp) == NULL) { 218 ASSERT(c2s_hash[i].ilb_conn_cnt == 0); 219 mutex_exit(&c2s_hash[i].ilb_conn_hash_lock); 220 continue; 221 } 222 do { 223 ASSERT(c2s_hash[i].ilb_conn_cnt > 0); 224 ASSERT(connp->conn_c2s_hash == &c2s_hash[i]); 225 nxt_connp = connp->conn_c2s_next; 226 expiry = now - SEC_TO_TICK(connp->conn_expiry); 227 if (connp->conn_server->iser_die_time != 0 && 228 connp->conn_server->iser_die_time < now) 229 die_now = B_TRUE; 230 else 231 die_now = B_FALSE; 232 s2c_hash = connp->conn_s2c_hash; 233 mutex_enter(&s2c_hash->ilb_conn_hash_lock); 234 235 if (connp->conn_gc || die_now || 236 (connp->conn_c2s_atime < expiry && 237 connp->conn_s2c_atime < expiry)) { 238 /* Need to update the nat list cur_connp */ 239 if (connp == ilbs->ilbs_conn_list_connp) { 240 ilbs->ilbs_conn_list_connp = 241 connp->conn_c2s_next; 242 } 243 ilb_conn_remove(connp); 244 goto nxt_connp; 245 } 246 247 if (connp->conn_l4 != IPPROTO_TCP) 248 goto nxt_connp; 249 250 /* Update and check TCP related conn info */ 251 if (connp->conn_c2s_tcp_fin_sent && 252 SEQ_GT(connp->conn_s2c_tcp_ack, 253 connp->conn_c2s_tcp_fss)) { 254 connp->conn_c2s_tcp_fin_acked = B_TRUE; 255 } 256 if (connp->conn_s2c_tcp_fin_sent && 257 SEQ_GT(connp->conn_c2s_tcp_ack, 258 connp->conn_s2c_tcp_fss)) { 259 connp->conn_s2c_tcp_fin_acked = B_TRUE; 260 } 261 if (connp->conn_c2s_tcp_fin_acked && 262 connp->conn_s2c_tcp_fin_acked) { 263 ilb_conn_remove(connp); 264 } 265 nxt_connp: 266 mutex_exit(&s2c_hash->ilb_conn_hash_lock); 267 connp = nxt_connp; 268 } while (connp != NULL); 269 mutex_exit(&c2s_hash[i].ilb_conn_hash_lock); 270 } 271 } 272 273 /* Conn hash timer routine. It dispatches a taskq and restart the timer */ 274 static void 275 ilb_conn_timer(void *arg) 276 { 277 ilb_timer_t *timer = (ilb_timer_t *)arg; 278 279 (void) taskq_dispatch(timer->ilbs->ilbs_conn_taskq, ilb_conn_cleanup, 280 arg, TQ_SLEEP); 281 mutex_enter(&timer->tid_lock); 282 if (timer->tid == 0) { 283 mutex_exit(&timer->tid_lock); 284 } else { 285 timer->tid = timeout(ilb_conn_timer, arg, 286 SEC_TO_TICK(ilb_conn_cache_timeout)); 287 mutex_exit(&timer->tid_lock); 288 } 289 } 290 291 void 292 ilb_conn_hash_init(ilb_stack_t *ilbs) 293 { 294 extern pri_t minclsyspri; 295 int i, part; 296 ilb_timer_t *tm; 297 char tq_name[TASKQ_NAMELEN]; 298 299 /* 300 * If ilbs->ilbs_conn_hash_size is not a power of 2, bump it up to 301 * the next power of 2. 302 */ 303 if (ilbs->ilbs_conn_hash_size & (ilbs->ilbs_conn_hash_size - 1)) { 304 for (i = 0; i < 31; i++) { 305 if (ilbs->ilbs_conn_hash_size < (1 << i)) 306 break; 307 } 308 ilbs->ilbs_conn_hash_size = 1 << i; 309 } 310 311 /* 312 * Can sleep since this should be called when a rule is being added, 313 * hence we are not in interrupt context. 314 */ 315 ilbs->ilbs_c2s_conn_hash = kmem_zalloc(sizeof (ilb_conn_hash_t) * 316 ilbs->ilbs_conn_hash_size, KM_SLEEP); 317 ilbs->ilbs_s2c_conn_hash = kmem_zalloc(sizeof (ilb_conn_hash_t) * 318 ilbs->ilbs_conn_hash_size, KM_SLEEP); 319 320 for (i = 0; i < ilbs->ilbs_conn_hash_size; i++) { 321 mutex_init(&ilbs->ilbs_c2s_conn_hash[i].ilb_conn_hash_lock, 322 NULL, MUTEX_DEFAULT, NULL); 323 } 324 for (i = 0; i < ilbs->ilbs_conn_hash_size; i++) { 325 mutex_init(&ilbs->ilbs_s2c_conn_hash[i].ilb_conn_hash_lock, 326 NULL, MUTEX_DEFAULT, NULL); 327 } 328 329 if (ilb_conn_cache == NULL) 330 ilb_conn_cache_init(); 331 332 (void) snprintf(tq_name, sizeof (tq_name), "ilb_conn_taskq_%p", 333 (void *)ilbs->ilbs_netstack); 334 ASSERT(ilbs->ilbs_conn_taskq == NULL); 335 ilbs->ilbs_conn_taskq = taskq_create(tq_name, 336 ilb_conn_timer_size * 2, minclsyspri, ilb_conn_timer_size, 337 ilb_conn_timer_size * 2, TASKQ_PREPOPULATE|TASKQ_DYNAMIC); 338 339 ASSERT(ilbs->ilbs_conn_timer_list == NULL); 340 ilbs->ilbs_conn_timer_list = kmem_zalloc(sizeof (ilb_timer_t) * 341 ilb_conn_timer_size, KM_SLEEP); 342 343 /* 344 * The hash table is divided in equal partition for those timers 345 * to do garbage collection. 346 */ 347 part = ilbs->ilbs_conn_hash_size / ilb_conn_timer_size + 1; 348 for (i = 0; i < ilb_conn_timer_size; i++) { 349 tm = ilbs->ilbs_conn_timer_list + i; 350 tm->start = i * part; 351 tm->end = i * part + part; 352 if (tm->end > ilbs->ilbs_conn_hash_size) 353 tm->end = ilbs->ilbs_conn_hash_size; 354 tm->ilbs = ilbs; 355 mutex_init(&tm->tid_lock, NULL, MUTEX_DEFAULT, NULL); 356 /* Spread out the starting execution time of all the timers. */ 357 tm->tid = timeout(ilb_conn_timer, tm, 358 SEC_TO_TICK(ilb_conn_cache_timeout + i)); 359 } 360 } 361 362 void 363 ilb_conn_hash_fini(ilb_stack_t *ilbs) 364 { 365 uint32_t i; 366 ilb_conn_t *connp; 367 368 if (ilbs->ilbs_c2s_conn_hash == NULL) { 369 ASSERT(ilbs->ilbs_s2c_conn_hash == NULL); 370 return; 371 } 372 373 /* Stop all the timers first. */ 374 for (i = 0; i < ilb_conn_timer_size; i++) { 375 timeout_id_t tid; 376 377 /* Setting tid to 0 tells the timer handler not to restart. */ 378 mutex_enter(&ilbs->ilbs_conn_timer_list[i].tid_lock); 379 tid = ilbs->ilbs_conn_timer_list[i].tid; 380 ilbs->ilbs_conn_timer_list[i].tid = 0; 381 mutex_exit(&ilbs->ilbs_conn_timer_list[i].tid_lock); 382 (void) untimeout(tid); 383 } 384 kmem_free(ilbs->ilbs_conn_timer_list, sizeof (ilb_timer_t) * 385 ilb_conn_timer_size); 386 taskq_destroy(ilbs->ilbs_conn_taskq); 387 ilbs->ilbs_conn_taskq = NULL; 388 389 /* Then remove all the conns. */ 390 for (i = 0; i < ilbs->ilbs_conn_hash_size; i++) { 391 while ((connp = ilbs->ilbs_s2c_conn_hash->ilb_connp) != NULL) { 392 ilbs->ilbs_s2c_conn_hash->ilb_connp = 393 connp->conn_s2c_next; 394 ILB_SERVER_REFRELE(connp->conn_server); 395 if (connp->conn_rule_cache.topo == ILB_TOPO_IMPL_NAT) { 396 ilb_nat_src_entry_t *ent; 397 in_port_t port; 398 399 /* 400 * src_ent will be freed in ilb_nat_src_fini(). 401 */ 402 port = ntohs( 403 connp->conn_rule_cache.info.nat_sport); 404 ent = connp->conn_rule_cache.info.src_ent; 405 vmem_free(ent->nse_port_arena, 406 (void *)(uintptr_t)port, 1); 407 } 408 kmem_cache_free(ilb_conn_cache, connp); 409 } 410 } 411 kmem_free(ilbs->ilbs_c2s_conn_hash, sizeof (ilb_conn_hash_t) * 412 ilbs->ilbs_conn_hash_size); 413 kmem_free(ilbs->ilbs_s2c_conn_hash, sizeof (ilb_conn_hash_t) * 414 ilbs->ilbs_conn_hash_size); 415 } 416 417 /* 418 * Internet checksum adjustment calculation routines. We pre-calculate 419 * checksum adjustment so that we don't need to compute the checksum on 420 * the whole packet when we change address/port in the packet. 421 */ 422 423 static void 424 hnat_cksum_v4(uint16_t *oaddr, uint16_t *naddr, in_port_t old_port, 425 in_port_t new_port, uint32_t *adj_sum) 426 { 427 uint32_t sum; 428 429 sum = *oaddr + *(oaddr + 1) + old_port; 430 while ((sum >> 16) != 0) 431 sum = (sum & 0xffff) + (sum >> 16); 432 *adj_sum = (uint16_t)~sum + *naddr + *(naddr + 1) + new_port; 433 } 434 435 static void 436 hnat_cksum_v6(uint16_t *oaddr, uint16_t *naddr, in_port_t old_port, 437 in_port_t new_port, uint32_t *adj_sum) 438 { 439 uint32_t sum = 0; 440 441 sum = *oaddr + *(oaddr + 1) + *(oaddr + 2) + *(oaddr + 3) + 442 *(oaddr + 4) + *(oaddr + 5) + *(oaddr + 6) + *(oaddr + 7) + 443 old_port; 444 while ((sum >> 16) != 0) 445 sum = (sum & 0xffff) + (sum >> 16); 446 *adj_sum = (uint16_t)~sum + *naddr + *(naddr + 1) + 447 *(naddr + 2) + *(naddr + 3) + *(naddr + 4) + *(naddr + 5) + 448 *(naddr + 6) + *(naddr + 7) + new_port; 449 } 450 451 static void 452 fnat_cksum_v4(uint16_t *oaddr1, uint16_t *oaddr2, uint16_t *naddr1, 453 uint16_t *naddr2, in_port_t old_port1, in_port_t old_port2, 454 in_port_t new_port1, in_port_t new_port2, uint32_t *adj_sum) 455 { 456 uint32_t sum; 457 458 sum = *oaddr1 + *(oaddr1 + 1) + old_port1 + *oaddr2 + *(oaddr2 + 1) + 459 old_port2; 460 while ((sum >> 16) != 0) 461 sum = (sum & 0xffff) + (sum >> 16); 462 *adj_sum = (uint16_t)~sum + *naddr1 + *(naddr1 + 1) + new_port1 + 463 *naddr2 + *(naddr2 + 1) + new_port2; 464 } 465 466 static void 467 fnat_cksum_v6(uint16_t *oaddr1, uint16_t *oaddr2, uint16_t *naddr1, 468 uint16_t *naddr2, in_port_t old_port1, in_port_t old_port2, 469 in_port_t new_port1, in_port_t new_port2, uint32_t *adj_sum) 470 { 471 uint32_t sum = 0; 472 473 sum = *oaddr1 + *(oaddr1 + 1) + *(oaddr1 + 2) + *(oaddr1 + 3) + 474 *(oaddr1 + 4) + *(oaddr1 + 5) + *(oaddr1 + 6) + *(oaddr1 + 7) + 475 old_port1; 476 sum += *oaddr2 + *(oaddr2 + 1) + *(oaddr2 + 2) + *(oaddr2 + 3) + 477 *(oaddr2 + 4) + *(oaddr2 + 5) + *(oaddr2 + 6) + *(oaddr2 + 7) + 478 old_port2; 479 while ((sum >> 16) != 0) 480 sum = (sum & 0xffff) + (sum >> 16); 481 sum = (uint16_t)~sum + *naddr1 + *(naddr1 + 1) + *(naddr1 + 2) + 482 *(naddr1 + 3) + *(naddr1 + 4) + *(naddr1 + 5) + *(naddr1 + 6) + 483 *(naddr1 + 7) + new_port1; 484 *adj_sum = sum + *naddr2 + *(naddr2 + 1) + *(naddr2 + 2) + 485 *(naddr2 + 3) + *(naddr2 + 4) + *(naddr2 + 5) + *(naddr2 + 6) + 486 *(naddr2 + 7) + new_port2; 487 } 488 489 /* 490 * Add a conn hash entry to the tables. Note that a conn hash entry 491 * (ilb_conn_t) contains info on both directions. And there are two hash 492 * tables, one for client to server and the other for server to client. 493 * So the same entry is added to both tables and can be ccessed by two 494 * thread simultaneously. But each thread will only access data on one 495 * direction, so there is no conflict. 496 */ 497 int 498 ilb_conn_add(ilb_stack_t *ilbs, ilb_rule_t *rule, ilb_server_t *server, 499 in6_addr_t *src, in_port_t sport, in6_addr_t *dst, in_port_t dport, 500 ilb_nat_info_t *info, uint32_t *ip_sum, uint32_t *tp_sum, ilb_sticky_t *s) 501 { 502 ilb_conn_t *connp; 503 ilb_conn_hash_t *hash; 504 int i; 505 506 connp = kmem_cache_alloc(ilb_conn_cache, KM_NOSLEEP); 507 if (connp == NULL) { 508 if (s != NULL) { 509 if (rule->ir_topo == ILB_TOPO_IMPL_NAT) { 510 ilb_nat_src_entry_t **entry; 511 512 entry = s->server->iser_nat_src->src_list; 513 vmem_free(entry[s->nat_src_idx]->nse_port_arena, 514 (void *)(uintptr_t)ntohs(info->nat_sport), 515 1); 516 } 517 ILB_STICKY_REFRELE(s); 518 } 519 return (ENOMEM); 520 } 521 522 connp->conn_l4 = rule->ir_proto; 523 524 connp->conn_server = server; 525 ILB_SERVER_REFHOLD(server); 526 connp->conn_sticky = s; 527 528 connp->conn_rule_cache.topo = rule->ir_topo; 529 connp->conn_rule_cache.info = *info; 530 531 connp->conn_gc = B_FALSE; 532 533 connp->conn_expiry = rule->ir_nat_expiry; 534 connp->conn_cr_time = ddi_get_lbolt64(); 535 536 /* Client to server info. */ 537 connp->conn_c2s_saddr = *src; 538 connp->conn_c2s_sport = sport; 539 connp->conn_c2s_daddr = *dst; 540 connp->conn_c2s_dport = dport; 541 542 connp->conn_c2s_atime = ddi_get_lbolt64(); 543 /* The packet ths triggers this creation should be counted */ 544 connp->conn_c2s_pkt_cnt = 1; 545 connp->conn_c2s_tcp_fin_sent = B_FALSE; 546 connp->conn_c2s_tcp_fin_acked = B_FALSE; 547 548 /* Server to client info, before NAT */ 549 switch (rule->ir_topo) { 550 case ILB_TOPO_IMPL_HALF_NAT: 551 connp->conn_s2c_saddr = info->nat_dst; 552 connp->conn_s2c_sport = info->nat_dport; 553 connp->conn_s2c_daddr = *src; 554 connp->conn_s2c_dport = sport; 555 556 /* Pre-calculate checksum changes for both directions */ 557 if (rule->ir_ipver == IPPROTO_IP) { 558 hnat_cksum_v4((uint16_t *)&dst->s6_addr32[3], 559 (uint16_t *)&info->nat_dst.s6_addr32[3], 0, 0, 560 &connp->conn_c2s_ip_sum); 561 hnat_cksum_v4((uint16_t *)&dst->s6_addr32[3], 562 (uint16_t *)&info->nat_dst.s6_addr32[3], dport, 563 info->nat_dport, &connp->conn_c2s_tp_sum); 564 *ip_sum = connp->conn_c2s_ip_sum; 565 *tp_sum = connp->conn_c2s_tp_sum; 566 567 hnat_cksum_v4( 568 (uint16_t *)&info->nat_dst.s6_addr32[3], 569 (uint16_t *)&dst->s6_addr32[3], 0, 0, 570 &connp->conn_s2c_ip_sum); 571 hnat_cksum_v4( 572 (uint16_t *)&info->nat_dst.s6_addr32[3], 573 (uint16_t *)&dst->s6_addr32[3], 574 info->nat_dport, dport, 575 &connp->conn_s2c_tp_sum); 576 } else { 577 connp->conn_c2s_ip_sum = 0; 578 hnat_cksum_v6((uint16_t *)dst, 579 (uint16_t *)&info->nat_dst, dport, 580 info->nat_dport, &connp->conn_c2s_tp_sum); 581 *ip_sum = 0; 582 *tp_sum = connp->conn_c2s_tp_sum; 583 584 connp->conn_s2c_ip_sum = 0; 585 hnat_cksum_v6((uint16_t *)&info->nat_dst, 586 (uint16_t *)dst, info->nat_dport, dport, 587 &connp->conn_s2c_tp_sum); 588 } 589 break; 590 case ILB_TOPO_IMPL_NAT: 591 connp->conn_s2c_saddr = info->nat_dst; 592 connp->conn_s2c_sport = info->nat_dport; 593 connp->conn_s2c_daddr = info->nat_src; 594 connp->conn_s2c_dport = info->nat_sport; 595 596 if (rule->ir_ipver == IPPROTO_IP) { 597 fnat_cksum_v4((uint16_t *)&src->s6_addr32[3], 598 (uint16_t *)&dst->s6_addr32[3], 599 (uint16_t *)&info->nat_src.s6_addr32[3], 600 (uint16_t *)&info->nat_dst.s6_addr32[3], 601 0, 0, 0, 0, &connp->conn_c2s_ip_sum); 602 fnat_cksum_v4((uint16_t *)&src->s6_addr32[3], 603 (uint16_t *)&dst->s6_addr32[3], 604 (uint16_t *)&info->nat_src.s6_addr32[3], 605 (uint16_t *)&info->nat_dst.s6_addr32[3], 606 sport, dport, info->nat_sport, 607 info->nat_dport, &connp->conn_c2s_tp_sum); 608 *ip_sum = connp->conn_c2s_ip_sum; 609 *tp_sum = connp->conn_c2s_tp_sum; 610 611 fnat_cksum_v4( 612 (uint16_t *)&info->nat_src.s6_addr32[3], 613 (uint16_t *)&info->nat_dst.s6_addr32[3], 614 (uint16_t *)&src->s6_addr32[3], 615 (uint16_t *)&dst->s6_addr32[3], 616 0, 0, 0, 0, &connp->conn_s2c_ip_sum); 617 fnat_cksum_v4( 618 (uint16_t *)&info->nat_src.s6_addr32[3], 619 (uint16_t *)&info->nat_dst.s6_addr32[3], 620 (uint16_t *)&src->s6_addr32[3], 621 (uint16_t *)&dst->s6_addr32[3], 622 info->nat_sport, info->nat_dport, 623 sport, dport, &connp->conn_s2c_tp_sum); 624 } else { 625 fnat_cksum_v6((uint16_t *)src, (uint16_t *)dst, 626 (uint16_t *)&info->nat_src, 627 (uint16_t *)&info->nat_dst, 628 sport, dport, info->nat_sport, 629 info->nat_dport, &connp->conn_c2s_tp_sum); 630 connp->conn_c2s_ip_sum = 0; 631 *ip_sum = 0; 632 *tp_sum = connp->conn_c2s_tp_sum; 633 634 fnat_cksum_v6((uint16_t *)&info->nat_src, 635 (uint16_t *)&info->nat_dst, (uint16_t *)src, 636 (uint16_t *)dst, info->nat_sport, 637 info->nat_dport, sport, dport, 638 &connp->conn_s2c_tp_sum); 639 connp->conn_s2c_ip_sum = 0; 640 } 641 break; 642 } 643 644 connp->conn_s2c_atime = ddi_get_lbolt64(); 645 connp->conn_s2c_pkt_cnt = 1; 646 connp->conn_s2c_tcp_fin_sent = B_FALSE; 647 connp->conn_s2c_tcp_fin_acked = B_FALSE; 648 649 /* Add it to the s2c hash table. */ 650 hash = ilbs->ilbs_s2c_conn_hash; 651 i = ILB_CONN_HASH((uint8_t *)&connp->conn_s2c_saddr.s6_addr32[3], 652 ntohs(connp->conn_s2c_sport), 653 (uint8_t *)&connp->conn_s2c_daddr.s6_addr32[3], 654 ntohs(connp->conn_s2c_dport), ilbs->ilbs_conn_hash_size); 655 connp->conn_s2c_hash = &hash[i]; 656 DTRACE_PROBE2(ilb__conn__hash__add__s2c, ilb_conn_t *, connp, int, i); 657 658 mutex_enter(&hash[i].ilb_conn_hash_lock); 659 hash[i].ilb_conn_cnt++; 660 connp->conn_s2c_next = hash[i].ilb_connp; 661 if (hash[i].ilb_connp != NULL) 662 hash[i].ilb_connp->conn_s2c_prev = connp; 663 connp->conn_s2c_prev = NULL; 664 hash[i].ilb_connp = connp; 665 mutex_exit(&hash[i].ilb_conn_hash_lock); 666 667 /* Add it to the c2s hash table. */ 668 hash = ilbs->ilbs_c2s_conn_hash; 669 i = ILB_CONN_HASH((uint8_t *)&src->s6_addr32[3], ntohs(sport), 670 (uint8_t *)&dst->s6_addr32[3], ntohs(dport), 671 ilbs->ilbs_conn_hash_size); 672 connp->conn_c2s_hash = &hash[i]; 673 DTRACE_PROBE2(ilb__conn__hash__add__c2s, ilb_conn_t *, connp, int, i); 674 675 mutex_enter(&hash[i].ilb_conn_hash_lock); 676 hash[i].ilb_conn_cnt++; 677 connp->conn_c2s_next = hash[i].ilb_connp; 678 if (hash[i].ilb_connp != NULL) 679 hash[i].ilb_connp->conn_c2s_prev = connp; 680 connp->conn_c2s_prev = NULL; 681 hash[i].ilb_connp = connp; 682 mutex_exit(&hash[i].ilb_conn_hash_lock); 683 684 return (0); 685 } 686 687 /* 688 * If a connection is using TCP, we keep track of simple TCP state transition 689 * so that we know when to clean up an entry. 690 */ 691 static boolean_t 692 update_conn_tcp(ilb_conn_t *connp, void *iph, tcpha_t *tcpha, int32_t pkt_len, 693 boolean_t c2s) 694 { 695 uint32_t ack, seq; 696 int32_t seg_len; 697 698 if (tcpha->tha_flags & TH_RST) 699 return (B_FALSE); 700 701 seg_len = pkt_len - ((uint8_t *)tcpha - (uint8_t *)iph) - 702 TCP_HDR_LENGTH((tcph_t *)tcpha); 703 704 if (tcpha->tha_flags & TH_ACK) 705 ack = ntohl(tcpha->tha_ack); 706 seq = ntohl(tcpha->tha_seq); 707 if (c2s) { 708 ASSERT(MUTEX_HELD(&connp->conn_c2s_hash->ilb_conn_hash_lock)); 709 if (tcpha->tha_flags & TH_FIN) { 710 connp->conn_c2s_tcp_fss = seq + seg_len; 711 connp->conn_c2s_tcp_fin_sent = B_TRUE; 712 } 713 connp->conn_c2s_tcp_ack = ack; 714 715 /* Port reuse by the client, restart the conn. */ 716 if (connp->conn_c2s_tcp_fin_sent && 717 SEQ_GT(seq, connp->conn_c2s_tcp_fss + 1)) { 718 connp->conn_c2s_tcp_fin_sent = B_FALSE; 719 connp->conn_c2s_tcp_fin_acked = B_FALSE; 720 } 721 } else { 722 ASSERT(MUTEX_HELD(&connp->conn_s2c_hash->ilb_conn_hash_lock)); 723 if (tcpha->tha_flags & TH_FIN) { 724 connp->conn_s2c_tcp_fss = seq + seg_len; 725 connp->conn_s2c_tcp_fin_sent = B_TRUE; 726 } 727 connp->conn_s2c_tcp_ack = ack; 728 729 /* Port reuse by the client, restart the conn. */ 730 if (connp->conn_s2c_tcp_fin_sent && 731 SEQ_GT(seq, connp->conn_s2c_tcp_fss + 1)) { 732 connp->conn_s2c_tcp_fin_sent = B_FALSE; 733 connp->conn_s2c_tcp_fin_acked = B_FALSE; 734 } 735 } 736 737 return (B_TRUE); 738 } 739 740 /* 741 * Helper routint to find conn hash entry given some packet information and 742 * the traffic direction (c2s, client to server?) 743 */ 744 static boolean_t 745 ilb_find_conn(ilb_stack_t *ilbs, void *iph, void *tph, int l4, in6_addr_t *src, 746 in_port_t sport, in6_addr_t *dst, in_port_t dport, 747 ilb_rule_info_t *rule_cache, uint32_t *ip_sum, uint32_t *tp_sum, 748 int32_t pkt_len, boolean_t c2s) 749 { 750 ilb_conn_hash_t *hash; 751 uint_t i; 752 ilb_conn_t *connp; 753 boolean_t tcp_alive; 754 boolean_t ret = B_FALSE; 755 756 i = ILB_CONN_HASH((uint8_t *)&src->s6_addr32[3], ntohs(sport), 757 (uint8_t *)&dst->s6_addr32[3], ntohs(dport), 758 ilbs->ilbs_conn_hash_size); 759 if (c2s) { 760 hash = ilbs->ilbs_c2s_conn_hash; 761 mutex_enter(&hash[i].ilb_conn_hash_lock); 762 for (connp = hash[i].ilb_connp; connp != NULL; 763 connp = connp->conn_c2s_next) { 764 if (connp->conn_l4 == l4 && 765 connp->conn_c2s_dport == dport && 766 connp->conn_c2s_sport == sport && 767 IN6_ARE_ADDR_EQUAL(src, &connp->conn_c2s_saddr) && 768 IN6_ARE_ADDR_EQUAL(dst, &connp->conn_c2s_daddr)) { 769 connp->conn_c2s_atime = ddi_get_lbolt64(); 770 connp->conn_c2s_pkt_cnt++; 771 *rule_cache = connp->conn_rule_cache; 772 *ip_sum = connp->conn_c2s_ip_sum; 773 *tp_sum = connp->conn_c2s_tp_sum; 774 ret = B_TRUE; 775 break; 776 } 777 } 778 } else { 779 hash = ilbs->ilbs_s2c_conn_hash; 780 mutex_enter(&hash[i].ilb_conn_hash_lock); 781 for (connp = hash[i].ilb_connp; connp != NULL; 782 connp = connp->conn_s2c_next) { 783 if (connp->conn_l4 == l4 && 784 connp->conn_s2c_dport == dport && 785 connp->conn_s2c_sport == sport && 786 IN6_ARE_ADDR_EQUAL(src, &connp->conn_s2c_saddr) && 787 IN6_ARE_ADDR_EQUAL(dst, &connp->conn_s2c_daddr)) { 788 connp->conn_s2c_atime = ddi_get_lbolt64(); 789 connp->conn_s2c_pkt_cnt++; 790 *rule_cache = connp->conn_rule_cache; 791 *ip_sum = connp->conn_s2c_ip_sum; 792 *tp_sum = connp->conn_s2c_tp_sum; 793 ret = B_TRUE; 794 break; 795 } 796 } 797 } 798 if (ret) { 799 ILB_S_KSTAT(connp->conn_server, pkt_processed); 800 ILB_S_KSTAT_UPDATE(connp->conn_server, bytes_processed, 801 pkt_len); 802 803 switch (l4) { 804 case (IPPROTO_TCP): 805 tcp_alive = update_conn_tcp(connp, iph, tph, pkt_len, 806 c2s); 807 if (!tcp_alive) { 808 connp->conn_gc = B_TRUE; 809 } 810 break; 811 default: 812 break; 813 } 814 } 815 mutex_exit(&hash[i].ilb_conn_hash_lock); 816 817 return (ret); 818 } 819 820 /* 821 * To check if a give packet matches an existing conn hash entry. If it 822 * does, return the information about this entry so that the caller can 823 * do the proper NAT. 824 */ 825 boolean_t 826 ilb_check_conn(ilb_stack_t *ilbs, int l3, void *iph, int l4, void *tph, 827 in6_addr_t *src, in6_addr_t *dst, in_port_t sport, in_port_t dport, 828 uint32_t pkt_len, in6_addr_t *lb_dst) 829 { 830 ilb_rule_info_t rule_cache; 831 uint32_t adj_ip_sum, adj_tp_sum; 832 boolean_t ret; 833 834 /* Check the incoming hash table. */ 835 if (ilb_find_conn(ilbs, iph, tph, l4, src, sport, dst, dport, 836 &rule_cache, &adj_ip_sum, &adj_tp_sum, pkt_len, B_TRUE)) { 837 switch (rule_cache.topo) { 838 case ILB_TOPO_IMPL_NAT: 839 *lb_dst = rule_cache.info.nat_dst; 840 ilb_full_nat(l3, iph, l4, tph, &rule_cache.info, 841 adj_ip_sum, adj_tp_sum, B_TRUE); 842 ret = B_TRUE; 843 break; 844 case ILB_TOPO_IMPL_HALF_NAT: 845 *lb_dst = rule_cache.info.nat_dst; 846 ilb_half_nat(l3, iph, l4, tph, &rule_cache.info, 847 adj_ip_sum, adj_tp_sum, B_TRUE); 848 ret = B_TRUE; 849 break; 850 default: 851 ret = B_FALSE; 852 break; 853 } 854 return (ret); 855 } 856 if (ilb_find_conn(ilbs, iph, tph, l4, src, sport, dst, dport, 857 &rule_cache, &adj_ip_sum, &adj_tp_sum, pkt_len, B_FALSE)) { 858 switch (rule_cache.topo) { 859 case ILB_TOPO_IMPL_NAT: 860 *lb_dst = rule_cache.info.src; 861 ilb_full_nat(l3, iph, l4, tph, &rule_cache.info, 862 adj_ip_sum, adj_tp_sum, B_FALSE); 863 ret = B_TRUE; 864 break; 865 case ILB_TOPO_IMPL_HALF_NAT: 866 *lb_dst = *dst; 867 ilb_half_nat(l3, iph, l4, tph, &rule_cache.info, 868 adj_ip_sum, adj_tp_sum, B_FALSE); 869 ret = B_TRUE; 870 break; 871 default: 872 ret = B_FALSE; 873 break; 874 } 875 return (ret); 876 } 877 878 return (B_FALSE); 879 } 880 881 /* 882 * To check if an ICMP packet belongs to a connection in one of the conn 883 * hash entries. 884 */ 885 boolean_t 886 ilb_check_icmp_conn(ilb_stack_t *ilbs, mblk_t *mp, int l3, void *out_iph, 887 void *icmph, in6_addr_t *lb_dst) 888 { 889 ilb_conn_hash_t *hash; 890 ipha_t *in_iph4; 891 ip6_t *in_iph6; 892 icmph_t *icmph4; 893 icmp6_t *icmph6; 894 in6_addr_t *in_src_p, *in_dst_p; 895 in_port_t *sport, *dport; 896 int l4; 897 uint_t i; 898 ilb_conn_t *connp; 899 ilb_rule_info_t rule_cache; 900 uint32_t adj_ip_sum; 901 boolean_t full_nat; 902 903 if (l3 == IPPROTO_IP) { 904 in6_addr_t in_src, in_dst; 905 906 icmph4 = (icmph_t *)icmph; 907 in_iph4 = (ipha_t *)&icmph4[1]; 908 909 if ((uint8_t *)in_iph4 + IPH_HDR_LENGTH(in_iph4) + 910 ICMP_MIN_TP_HDR_LEN > mp->b_wptr) { 911 return (B_FALSE); 912 } 913 914 IN6_IPADDR_TO_V4MAPPED(in_iph4->ipha_src, &in_src); 915 in_src_p = &in_src; 916 IN6_IPADDR_TO_V4MAPPED(in_iph4->ipha_dst, &in_dst); 917 in_dst_p = &in_dst; 918 919 l4 = in_iph4->ipha_protocol; 920 if (l4 != IPPROTO_TCP && l4 != IPPROTO_UDP) 921 return (B_FALSE); 922 923 sport = (in_port_t *)((char *)in_iph4 + 924 IPH_HDR_LENGTH(in_iph4)); 925 dport = sport + 1; 926 927 DTRACE_PROBE4(ilb__chk__icmp__conn__v4, uint32_t, 928 in_iph4->ipha_src, uint32_t, in_iph4->ipha_dst, uint16_t, 929 ntohs(*sport), uint16_t, ntohs(*dport)); 930 } else { 931 ASSERT(l3 == IPPROTO_IPV6); 932 933 icmph6 = (icmp6_t *)icmph; 934 in_iph6 = (ip6_t *)&icmph6[1]; 935 in_src_p = &in_iph6->ip6_src; 936 in_dst_p = &in_iph6->ip6_dst; 937 938 if ((uint8_t *)in_iph6 + sizeof (ip6_t) + 939 ICMP_MIN_TP_HDR_LEN > mp->b_wptr) { 940 return (B_FALSE); 941 } 942 943 l4 = in_iph6->ip6_nxt; 944 /* We don't go deep inside an IPv6 packet yet. */ 945 if (l4 != IPPROTO_TCP && l4 != IPPROTO_UDP) 946 return (B_FALSE); 947 948 sport = (in_port_t *)&in_iph6[1]; 949 dport = sport + 1; 950 951 DTRACE_PROBE4(ilb__chk__icmp__conn__v6, in6_addr_t *, 952 &in_iph6->ip6_src, in6_addr_t *, &in_iph6->ip6_dst, 953 uint16_t, ntohs(*sport), uint16_t, ntohs(*dport)); 954 } 955 956 i = ILB_CONN_HASH((uint8_t *)&in_dst_p->s6_addr32[3], ntohs(*dport), 957 (uint8_t *)&in_src_p->s6_addr32[3], ntohs(*sport), 958 ilbs->ilbs_conn_hash_size); 959 hash = ilbs->ilbs_c2s_conn_hash; 960 961 mutex_enter(&hash[i].ilb_conn_hash_lock); 962 for (connp = hash[i].ilb_connp; connp != NULL; 963 connp = connp->conn_c2s_next) { 964 if (connp->conn_l4 == l4 && 965 connp->conn_c2s_dport == *sport && 966 connp->conn_c2s_sport == *dport && 967 IN6_ARE_ADDR_EQUAL(in_dst_p, &connp->conn_c2s_saddr) && 968 IN6_ARE_ADDR_EQUAL(in_src_p, &connp->conn_c2s_daddr)) { 969 connp->conn_c2s_atime = ddi_get_lbolt64(); 970 connp->conn_c2s_pkt_cnt++; 971 rule_cache = connp->conn_rule_cache; 972 adj_ip_sum = connp->conn_c2s_ip_sum; 973 break; 974 } 975 } 976 mutex_exit(&hash[i].ilb_conn_hash_lock); 977 978 if (connp == NULL) { 979 DTRACE_PROBE(ilb__chk__icmp__conn__failed); 980 return (B_FALSE); 981 } 982 983 switch (rule_cache.topo) { 984 case ILB_TOPO_IMPL_NAT: 985 full_nat = B_TRUE; 986 break; 987 case ILB_TOPO_IMPL_HALF_NAT: 988 full_nat = B_FALSE; 989 break; 990 default: 991 return (B_FALSE); 992 } 993 994 *lb_dst = rule_cache.info.nat_dst; 995 if (l3 == IPPROTO_IP) { 996 ilb_nat_icmpv4(mp, out_iph, icmph4, in_iph4, sport, dport, 997 &rule_cache.info, adj_ip_sum, full_nat); 998 } else { 999 ilb_nat_icmpv6(mp, out_iph, icmph6, in_iph6, sport, dport, 1000 &rule_cache.info, full_nat); 1001 } 1002 return (B_TRUE); 1003 } 1004 1005 /* 1006 * This routine sends up the conn hash table to user land. Note that the 1007 * request is an ioctl, hence we cannot really differentiate requests 1008 * from different clients. There is no context shared between different 1009 * ioctls. Here we make the assumption that the user land ilbd will 1010 * only allow one client to show the conn hash table at any time. 1011 * Otherwise, the results will be "very" inconsistent. 1012 * 1013 * In each ioctl, a flag (ILB_LIST_BEGIN) indicates whether the client wants 1014 * to read from the beginning of the able. After a certain entries 1015 * are reported, the kernel remembers the position of the last returned 1016 * entry. When the next ioctl comes in with the ILB_LIST_BEGIN flag, 1017 * it will return entries starting from where it was left off. When 1018 * the end of table is reached, a flag (ILB_LIST_END) is set to tell 1019 * the client that there is no more entry. 1020 * 1021 * It is assumed that the caller has checked the size of nat so that it 1022 * can hold num entries. 1023 */ 1024 /* ARGSUSED */ 1025 int 1026 ilb_list_nat(ilb_stack_t *ilbs, zoneid_t zoneid, ilb_nat_entry_t *nat, 1027 uint32_t *num, uint32_t *flags) 1028 { 1029 ilb_conn_hash_t *hash; 1030 ilb_conn_t *cur_connp; 1031 uint32_t i, j; 1032 int ret = 0; 1033 1034 mutex_enter(&ilbs->ilbs_conn_list_lock); 1035 while (ilbs->ilbs_conn_list_busy) { 1036 if (cv_wait_sig(&ilbs->ilbs_conn_list_cv, 1037 &ilbs->ilbs_conn_list_lock) == 0) { 1038 mutex_exit(&ilbs->ilbs_conn_list_lock); 1039 return (EINTR); 1040 } 1041 } 1042 if ((hash = ilbs->ilbs_c2s_conn_hash) == NULL) { 1043 ASSERT(ilbs->ilbs_s2c_conn_hash == NULL); 1044 mutex_exit(&ilbs->ilbs_conn_list_lock); 1045 *num = 0; 1046 *flags |= ILB_LIST_END; 1047 return (0); 1048 } 1049 ilbs->ilbs_conn_list_busy = B_TRUE; 1050 mutex_exit(&ilbs->ilbs_conn_list_lock); 1051 1052 if (*flags & ILB_LIST_BEGIN) { 1053 i = 0; 1054 mutex_enter(&hash[0].ilb_conn_hash_lock); 1055 cur_connp = hash[0].ilb_connp; 1056 } else if (*flags & ILB_LIST_CONT) { 1057 if (ilbs->ilbs_conn_list_cur == ilbs->ilbs_conn_hash_size) { 1058 *num = 0; 1059 *flags |= ILB_LIST_END; 1060 goto done; 1061 } 1062 i = ilbs->ilbs_conn_list_cur; 1063 mutex_enter(&hash[i].ilb_conn_hash_lock); 1064 cur_connp = ilbs->ilbs_conn_list_connp; 1065 } else { 1066 ret = EINVAL; 1067 goto done; 1068 } 1069 1070 j = 0; 1071 while (j < *num) { 1072 if (cur_connp == NULL) { 1073 mutex_exit(&hash[i].ilb_conn_hash_lock); 1074 if (++i == ilbs->ilbs_conn_hash_size) { 1075 *flags |= ILB_LIST_END; 1076 break; 1077 } 1078 mutex_enter(&hash[i].ilb_conn_hash_lock); 1079 cur_connp = hash[i].ilb_connp; 1080 continue; 1081 } 1082 nat[j].proto = cur_connp->conn_l4; 1083 1084 nat[j].in_global = cur_connp->conn_c2s_daddr; 1085 nat[j].in_global_port = cur_connp->conn_c2s_dport; 1086 nat[j].out_global = cur_connp->conn_c2s_saddr; 1087 nat[j].out_global_port = cur_connp->conn_c2s_sport; 1088 1089 nat[j].in_local = cur_connp->conn_s2c_saddr; 1090 nat[j].in_local_port = cur_connp->conn_s2c_sport; 1091 nat[j].out_local = cur_connp->conn_s2c_daddr; 1092 nat[j].out_local_port = cur_connp->conn_s2c_dport; 1093 1094 nat[j].create_time = TICK_TO_MSEC(cur_connp->conn_cr_time); 1095 nat[j].last_access_time = 1096 TICK_TO_MSEC(cur_connp->conn_c2s_atime); 1097 1098 /* 1099 * The conn_s2c_pkt_cnt may not be accurate since we are not 1100 * holding the s2c hash lock. 1101 */ 1102 nat[j].pkt_cnt = cur_connp->conn_c2s_pkt_cnt + 1103 cur_connp->conn_s2c_pkt_cnt; 1104 j++; 1105 1106 cur_connp = cur_connp->conn_c2s_next; 1107 } 1108 ilbs->ilbs_conn_list_connp = cur_connp; 1109 if (j == *num) 1110 mutex_exit(&hash[i].ilb_conn_hash_lock); 1111 1112 ilbs->ilbs_conn_list_cur = i; 1113 1114 *num = j; 1115 done: 1116 mutex_enter(&ilbs->ilbs_conn_list_lock); 1117 ilbs->ilbs_conn_list_busy = B_FALSE; 1118 cv_signal(&ilbs->ilbs_conn_list_cv); 1119 mutex_exit(&ilbs->ilbs_conn_list_lock); 1120 1121 return (ret); 1122 } 1123 1124 1125 /* 1126 * Stickiness (persistence) handling routines. 1127 */ 1128 1129 1130 static void 1131 ilb_sticky_cache_init(void) 1132 { 1133 ilb_sticky_cache = kmem_cache_create("ilb_sticky_cache", 1134 sizeof (ilb_sticky_t), 0, NULL, NULL, NULL, NULL, NULL, 1135 ilb_kmem_flags); 1136 } 1137 1138 void 1139 ilb_sticky_cache_fini(void) 1140 { 1141 if (ilb_sticky_cache != NULL) { 1142 kmem_cache_destroy(ilb_sticky_cache); 1143 ilb_sticky_cache = NULL; 1144 } 1145 } 1146 1147 void 1148 ilb_sticky_refrele(ilb_sticky_t *s) 1149 { 1150 ILB_STICKY_REFRELE(s); 1151 } 1152 1153 static ilb_sticky_t * 1154 ilb_sticky_lookup(ilb_sticky_hash_t *hash, ilb_rule_t *rule, in6_addr_t *src) 1155 { 1156 ilb_sticky_t *s; 1157 1158 ASSERT(mutex_owned(&hash->sticky_lock)); 1159 1160 for (s = list_head(&hash->sticky_head); s != NULL; 1161 s = list_next(&hash->sticky_head, s)) { 1162 if (s->rule_instance == rule->ir_ks_instance) { 1163 if (IN6_ARE_ADDR_EQUAL(src, &s->src)) 1164 return (s); 1165 } 1166 } 1167 return (NULL); 1168 } 1169 1170 static ilb_sticky_t * 1171 ilb_sticky_add(ilb_sticky_hash_t *hash, ilb_rule_t *rule, ilb_server_t *server, 1172 in6_addr_t *src) 1173 { 1174 ilb_sticky_t *s; 1175 1176 ASSERT(mutex_owned(&hash->sticky_lock)); 1177 1178 if ((s = kmem_cache_alloc(ilb_sticky_cache, KM_NOSLEEP)) == NULL) 1179 return (NULL); 1180 1181 /* 1182 * The rule instance is for handling the scenario when the same 1183 * client talks to different rules at the same time. Stickiness 1184 * is per rule so we can use the rule instance to differentiate 1185 * the client's request. 1186 */ 1187 s->rule_instance = rule->ir_ks_instance; 1188 /* 1189 * Copy the rule name for listing all sticky cache entry. ir_name 1190 * is guaranteed to be NULL terminated. 1191 */ 1192 (void) strcpy(s->rule_name, rule->ir_name); 1193 s->server = server; 1194 1195 /* 1196 * Grab a ref cnt on the server so that it won't go away while 1197 * it is still in the sticky table. 1198 */ 1199 ILB_SERVER_REFHOLD(server); 1200 s->src = *src; 1201 s->expiry = rule->ir_sticky_expiry; 1202 s->refcnt = 1; 1203 s->hash = hash; 1204 1205 /* 1206 * There is no need to set atime here since the refcnt is not 1207 * zero. A sticky entry is removed only when the refcnt is 1208 * zero. But just set it here for debugging purpose. The 1209 * atime is set when a refrele is done on a sticky entry. 1210 */ 1211 s->atime = ddi_get_lbolt64(); 1212 1213 list_insert_head(&hash->sticky_head, s); 1214 hash->sticky_cnt++; 1215 return (s); 1216 } 1217 1218 /* 1219 * This routine checks if there is an existing sticky entry which matches 1220 * a given packet. If there is one, return it. If there is not, create 1221 * a sticky entry using the packet's info. 1222 */ 1223 ilb_server_t * 1224 ilb_sticky_find_add(ilb_stack_t *ilbs, ilb_rule_t *rule, in6_addr_t *src, 1225 ilb_server_t *server, ilb_sticky_t **res, uint16_t *src_ent_idx) 1226 { 1227 int i; 1228 ilb_sticky_hash_t *hash; 1229 ilb_sticky_t *s; 1230 1231 ASSERT(server != NULL); 1232 1233 *res = NULL; 1234 1235 i = ILB_STICKY_HASH((uint8_t *)&src->s6_addr32[3], 1236 (uint32_t)(uintptr_t)rule, ilbs->ilbs_sticky_hash_size); 1237 hash = &ilbs->ilbs_sticky_hash[i]; 1238 1239 /* First check if there is already an entry. */ 1240 mutex_enter(&hash->sticky_lock); 1241 s = ilb_sticky_lookup(hash, rule, src); 1242 1243 /* No sticky entry, add one. */ 1244 if (s == NULL) { 1245 add_new_entry: 1246 s = ilb_sticky_add(hash, rule, server, src); 1247 if (s == NULL) { 1248 mutex_exit(&hash->sticky_lock); 1249 return (NULL); 1250 } 1251 /* 1252 * Find a source for this server. All subseqent requests from 1253 * the same client matching this sticky entry will use this 1254 * source address in doing NAT. The current algorithm is 1255 * simple, rotate the source address. Note that the 1256 * source address array does not change after it's created, so 1257 * it is OK to just increment the cur index. 1258 */ 1259 if (server->iser_nat_src != NULL) { 1260 /* It is a hint, does not need to be atomic. */ 1261 *src_ent_idx = (server->iser_nat_src->cur++ % 1262 server->iser_nat_src->num_src); 1263 s->nat_src_idx = *src_ent_idx; 1264 } 1265 mutex_exit(&hash->sticky_lock); 1266 *res = s; 1267 return (server); 1268 } 1269 1270 /* 1271 * We don't hold any lock accessing iser_enabled. Refer to the 1272 * comment in ilb_server_add() about iser_lock. 1273 */ 1274 if (!s->server->iser_enabled) { 1275 /* 1276 * s->server == server can only happen if there is a race in 1277 * toggling the iser_enabled flag (we don't hold a lock doing 1278 * that) so that the load balance algorithm still returns a 1279 * disabled server. In this case, just drop the packet... 1280 */ 1281 if (s->server == server) { 1282 mutex_exit(&hash->sticky_lock); 1283 return (NULL); 1284 } 1285 1286 /* 1287 * The old server is disabled and there is a new server, use 1288 * the new one to create a sticky entry. Since we will 1289 * add the entry at the beginning, subsequent lookup will 1290 * find this new entry instead of the old one. 1291 */ 1292 goto add_new_entry; 1293 } 1294 1295 s->refcnt++; 1296 *res = s; 1297 mutex_exit(&hash->sticky_lock); 1298 if (server->iser_nat_src != NULL) 1299 *src_ent_idx = s->nat_src_idx; 1300 return (s->server); 1301 } 1302 1303 static void 1304 ilb_sticky_cleanup(void *arg) 1305 { 1306 ilb_timer_t *timer = (ilb_timer_t *)arg; 1307 uint32_t i; 1308 ilb_stack_t *ilbs; 1309 ilb_sticky_hash_t *hash; 1310 ilb_sticky_t *s, *nxt_s; 1311 int64_t now, expiry; 1312 1313 ilbs = timer->ilbs; 1314 hash = ilbs->ilbs_sticky_hash; 1315 ASSERT(hash != NULL); 1316 1317 now = ddi_get_lbolt64(); 1318 for (i = timer->start; i < timer->end; i++) { 1319 mutex_enter(&hash[i].sticky_lock); 1320 for (s = list_head(&hash[i].sticky_head); s != NULL; 1321 s = nxt_s) { 1322 nxt_s = list_next(&hash[i].sticky_head, s); 1323 if (s->refcnt != 0) 1324 continue; 1325 expiry = now - SEC_TO_TICK(s->expiry); 1326 if (s->atime < expiry) { 1327 ILB_SERVER_REFRELE(s->server); 1328 list_remove(&hash[i].sticky_head, s); 1329 kmem_cache_free(ilb_sticky_cache, s); 1330 hash[i].sticky_cnt--; 1331 } 1332 } 1333 mutex_exit(&hash[i].sticky_lock); 1334 } 1335 } 1336 1337 static void 1338 ilb_sticky_timer(void *arg) 1339 { 1340 ilb_timer_t *timer = (ilb_timer_t *)arg; 1341 1342 (void) taskq_dispatch(timer->ilbs->ilbs_sticky_taskq, 1343 ilb_sticky_cleanup, arg, TQ_SLEEP); 1344 mutex_enter(&timer->tid_lock); 1345 if (timer->tid == 0) { 1346 mutex_exit(&timer->tid_lock); 1347 } else { 1348 timer->tid = timeout(ilb_sticky_timer, arg, 1349 SEC_TO_TICK(ilb_sticky_timeout)); 1350 mutex_exit(&timer->tid_lock); 1351 } 1352 } 1353 1354 void 1355 ilb_sticky_hash_init(ilb_stack_t *ilbs) 1356 { 1357 extern pri_t minclsyspri; 1358 int i, part; 1359 char tq_name[TASKQ_NAMELEN]; 1360 ilb_timer_t *tm; 1361 1362 if (ilbs->ilbs_sticky_hash_size & (ilbs->ilbs_sticky_hash_size - 1)) { 1363 for (i = 0; i < 31; i++) { 1364 if (ilbs->ilbs_sticky_hash_size < (1 << i)) 1365 break; 1366 } 1367 ilbs->ilbs_sticky_hash_size = 1 << i; 1368 } 1369 1370 ilbs->ilbs_sticky_hash = kmem_zalloc(sizeof (ilb_sticky_hash_t) * 1371 ilbs->ilbs_sticky_hash_size, KM_SLEEP); 1372 for (i = 0; i < ilbs->ilbs_sticky_hash_size; i++) { 1373 mutex_init(&ilbs->ilbs_sticky_hash[i].sticky_lock, NULL, 1374 MUTEX_DEFAULT, NULL); 1375 list_create(&ilbs->ilbs_sticky_hash[i].sticky_head, 1376 sizeof (ilb_sticky_t), 1377 offsetof(ilb_sticky_t, list)); 1378 } 1379 1380 if (ilb_sticky_cache == NULL) 1381 ilb_sticky_cache_init(); 1382 1383 (void) snprintf(tq_name, sizeof (tq_name), "ilb_sticky_taskq_%p", 1384 (void *)ilbs->ilbs_netstack); 1385 ASSERT(ilbs->ilbs_sticky_taskq == NULL); 1386 ilbs->ilbs_sticky_taskq = taskq_create(tq_name, 1387 ilb_sticky_timer_size * 2, minclsyspri, ilb_sticky_timer_size, 1388 ilb_sticky_timer_size * 2, TASKQ_PREPOPULATE|TASKQ_DYNAMIC); 1389 1390 ASSERT(ilbs->ilbs_sticky_timer_list == NULL); 1391 ilbs->ilbs_sticky_timer_list = kmem_zalloc(sizeof (ilb_timer_t) * 1392 ilb_sticky_timer_size, KM_SLEEP); 1393 part = ilbs->ilbs_sticky_hash_size / ilb_sticky_timer_size + 1; 1394 for (i = 0; i < ilb_sticky_timer_size; i++) { 1395 tm = ilbs->ilbs_sticky_timer_list + i; 1396 tm->start = i * part; 1397 tm->end = i * part + part; 1398 if (tm->end > ilbs->ilbs_sticky_hash_size) 1399 tm->end = ilbs->ilbs_sticky_hash_size; 1400 tm->ilbs = ilbs; 1401 mutex_init(&tm->tid_lock, NULL, MUTEX_DEFAULT, NULL); 1402 /* Spread out the starting execution time of all the timers. */ 1403 tm->tid = timeout(ilb_sticky_timer, tm, 1404 SEC_TO_TICK(ilb_sticky_timeout + i)); 1405 } 1406 } 1407 1408 void 1409 ilb_sticky_hash_fini(ilb_stack_t *ilbs) 1410 { 1411 int i; 1412 ilb_sticky_t *s; 1413 1414 if (ilbs->ilbs_sticky_hash == NULL) 1415 return; 1416 1417 /* Stop all the timers first. */ 1418 for (i = 0; i < ilb_sticky_timer_size; i++) { 1419 timeout_id_t tid; 1420 1421 /* Setting tid to 0 tells the timer handler not to restart. */ 1422 mutex_enter(&ilbs->ilbs_sticky_timer_list[i].tid_lock); 1423 tid = ilbs->ilbs_sticky_timer_list[i].tid; 1424 ilbs->ilbs_sticky_timer_list[i].tid = 0; 1425 mutex_exit(&ilbs->ilbs_sticky_timer_list[i].tid_lock); 1426 (void) untimeout(tid); 1427 } 1428 kmem_free(ilbs->ilbs_sticky_timer_list, sizeof (ilb_timer_t) * 1429 ilb_sticky_timer_size); 1430 taskq_destroy(ilbs->ilbs_sticky_taskq); 1431 ilbs->ilbs_sticky_taskq = NULL; 1432 1433 for (i = 0; i < ilbs->ilbs_sticky_hash_size; i++) { 1434 while ((s = list_head(&ilbs->ilbs_sticky_hash[i].sticky_head)) 1435 != NULL) { 1436 list_remove(&ilbs->ilbs_sticky_hash[i].sticky_head, s); 1437 ILB_SERVER_REFRELE(s->server); 1438 kmem_free(s, sizeof (ilb_sticky_t)); 1439 } 1440 } 1441 kmem_free(ilbs->ilbs_sticky_hash, ilbs->ilbs_sticky_hash_size * 1442 sizeof (ilb_sticky_hash_t)); 1443 } 1444 1445 /* 1446 * This routine sends up the sticky hash table to user land. Refer to 1447 * the comments before ilb_list_nat(). Both routines assume similar 1448 * conditions. 1449 * 1450 * It is assumed that the caller has checked the size of st so that it 1451 * can hold num entries. 1452 */ 1453 /* ARGSUSED */ 1454 int 1455 ilb_list_sticky(ilb_stack_t *ilbs, zoneid_t zoneid, ilb_sticky_entry_t *st, 1456 uint32_t *num, uint32_t *flags) 1457 { 1458 ilb_sticky_hash_t *hash; 1459 ilb_sticky_t *curp; 1460 uint32_t i, j; 1461 int ret = 0; 1462 1463 mutex_enter(&ilbs->ilbs_sticky_list_lock); 1464 while (ilbs->ilbs_sticky_list_busy) { 1465 if (cv_wait_sig(&ilbs->ilbs_sticky_list_cv, 1466 &ilbs->ilbs_sticky_list_lock) == 0) { 1467 mutex_exit(&ilbs->ilbs_sticky_list_lock); 1468 return (EINTR); 1469 } 1470 } 1471 if ((hash = ilbs->ilbs_sticky_hash) == NULL) { 1472 mutex_exit(&ilbs->ilbs_sticky_list_lock); 1473 *num = 0; 1474 *flags |= ILB_LIST_END; 1475 return (0); 1476 } 1477 ilbs->ilbs_sticky_list_busy = B_TRUE; 1478 mutex_exit(&ilbs->ilbs_sticky_list_lock); 1479 1480 if (*flags & ILB_LIST_BEGIN) { 1481 i = 0; 1482 mutex_enter(&hash[0].sticky_lock); 1483 curp = list_head(&hash[0].sticky_head); 1484 } else if (*flags & ILB_LIST_CONT) { 1485 if (ilbs->ilbs_sticky_list_cur == ilbs->ilbs_sticky_hash_size) { 1486 *num = 0; 1487 *flags |= ILB_LIST_END; 1488 goto done; 1489 } 1490 i = ilbs->ilbs_sticky_list_cur; 1491 mutex_enter(&hash[i].sticky_lock); 1492 curp = ilbs->ilbs_sticky_list_curp; 1493 } else { 1494 ret = EINVAL; 1495 goto done; 1496 } 1497 1498 j = 0; 1499 while (j < *num) { 1500 if (curp == NULL) { 1501 mutex_exit(&hash[i].sticky_lock); 1502 if (++i == ilbs->ilbs_sticky_hash_size) { 1503 *flags |= ILB_LIST_END; 1504 break; 1505 } 1506 mutex_enter(&hash[i].sticky_lock); 1507 curp = list_head(&hash[i].sticky_head); 1508 continue; 1509 } 1510 (void) strcpy(st[j].rule_name, curp->rule_name); 1511 st[j].req_addr = curp->src; 1512 st[j].srv_addr = curp->server->iser_addr_v6; 1513 st[j].expiry_time = TICK_TO_MSEC(curp->expiry); 1514 j++; 1515 curp = list_next(&hash[i].sticky_head, curp); 1516 } 1517 ilbs->ilbs_sticky_list_curp = curp; 1518 if (j == *num) 1519 mutex_exit(&hash[i].sticky_lock); 1520 1521 ilbs->ilbs_sticky_list_cur = i; 1522 1523 *num = j; 1524 done: 1525 mutex_enter(&ilbs->ilbs_sticky_list_lock); 1526 ilbs->ilbs_sticky_list_busy = B_FALSE; 1527 cv_signal(&ilbs->ilbs_sticky_list_cv); 1528 mutex_exit(&ilbs->ilbs_sticky_list_lock); 1529 1530 return (ret); 1531 } 1532