1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #ifndef _INET_IP_IMPL_H 27 #define _INET_IP_IMPL_H 28 29 /* 30 * IP implementation private declarations. These interfaces are 31 * used to build the IP module and are not meant to be accessed 32 * by any modules except IP itself. They are undocumented and are 33 * subject to change without notice. 34 */ 35 36 #ifdef __cplusplus 37 extern "C" { 38 #endif 39 40 #ifdef _KERNEL 41 42 #include <sys/sdt.h> 43 #include <sys/dld.h> 44 45 #define IP_MOD_ID 5701 46 47 #ifdef _BIG_ENDIAN 48 #define IP_HDR_CSUM_TTL_ADJUST 256 49 #define IP_TCP_CSUM_COMP IPPROTO_TCP 50 #define IP_UDP_CSUM_COMP IPPROTO_UDP 51 #else 52 #define IP_HDR_CSUM_TTL_ADJUST 1 53 #define IP_TCP_CSUM_COMP (IPPROTO_TCP << 8) 54 #define IP_UDP_CSUM_COMP (IPPROTO_UDP << 8) 55 #endif 56 57 #define TCP_CHECKSUM_OFFSET 16 58 #define TCP_CHECKSUM_SIZE 2 59 60 #define UDP_CHECKSUM_OFFSET 6 61 #define UDP_CHECKSUM_SIZE 2 62 63 #define IPH_TCPH_CHECKSUMP(ipha, hlen) \ 64 ((uint16_t *)(((uchar_t *)(ipha)) + ((hlen) + TCP_CHECKSUM_OFFSET))) 65 66 #define IPH_UDPH_CHECKSUMP(ipha, hlen) \ 67 ((uint16_t *)(((uchar_t *)(ipha)) + ((hlen) + UDP_CHECKSUM_OFFSET))) 68 69 #define ILL_HCKSUM_CAPABLE(ill) \ 70 (((ill)->ill_capabilities & ILL_CAPAB_HCKSUM) != 0) 71 /* 72 * Macro that performs software checksum calculation on the IP header. 73 */ 74 #define IP_HDR_CKSUM(ipha, sum, v_hlen_tos_len, ttl_protocol) { \ 75 (sum) += (ttl_protocol) + (ipha)->ipha_ident + \ 76 ((v_hlen_tos_len) >> 16) + \ 77 ((v_hlen_tos_len) & 0xFFFF) + \ 78 (ipha)->ipha_fragment_offset_and_flags; \ 79 (sum) = (((sum) & 0xFFFF) + ((sum) >> 16)); \ 80 (sum) = ~((sum) + ((sum) >> 16)); \ 81 (ipha)->ipha_hdr_checksum = (uint16_t)(sum); \ 82 } 83 84 #define IS_IP_HDR_HWCKSUM(ipsec, mp, ill) \ 85 ((!ipsec) && (DB_CKSUMFLAGS(mp) & HCK_IPV4_HDRCKSUM) && \ 86 ILL_HCKSUM_CAPABLE(ill) && dohwcksum) 87 88 /* 89 * This macro acts as a wrapper around IP_CKSUM_XMIT_FAST, and it performs 90 * several checks on the IRE and ILL (among other things) in order to see 91 * whether or not hardware checksum offload is allowed for the outgoing 92 * packet. It assumes that the caller has held a reference to the IRE. 93 */ 94 #define IP_CKSUM_XMIT(ill, ire, mp, ihp, up, proto, start, end, \ 95 max_frag, ipsec_len, pseudo) { \ 96 uint32_t _hck_flags; \ 97 /* \ 98 * We offload checksum calculation to hardware when IPsec isn't \ 99 * present and if fragmentation isn't required. We also check \ 100 * if M_DATA fastpath is safe to be used on the corresponding \ 101 * IRE; this check is performed without grabbing ire_lock but \ 102 * instead by holding a reference to it. This is sufficient \ 103 * for IRE_CACHE; for IRE_BROADCAST on non-Ethernet links, the \ 104 * DL_NOTE_FASTPATH_FLUSH indication could come up from the \ 105 * driver and trigger the IRE (hence fp_mp) deletion. This is \ 106 * why only IRE_CACHE type is eligible for offload. \ 107 * \ 108 * The presense of IP options also forces the network stack to \ 109 * calculate the checksum in software. This is because: \ 110 * \ 111 * Wrap around: certain partial-checksum NICs (eri, ce) limit \ 112 * the size of "start offset" width to 6-bit. This effectively \ 113 * sets the largest value of the offset to 64-bytes, starting \ 114 * from the MAC header. When the cumulative MAC and IP headers \ 115 * exceed such limit, the offset will wrap around. This causes \ 116 * the checksum to be calculated at the wrong place. \ 117 * \ 118 * IPv4 source routing: none of the full-checksum capable NICs \ 119 * is capable of correctly handling the IPv4 source-routing \ 120 * option for purposes of calculating the pseudo-header; the \ 121 * actual destination is different from the destination in the \ 122 * header which is that of the next-hop. (This case may not be \ 123 * true for NICs which can parse IPv6 extension headers, but \ 124 * we choose to simplify the implementation by not offloading \ 125 * checksum when they are present.) \ 126 * \ 127 */ \ 128 if ((ill) != NULL && ILL_HCKSUM_CAPABLE(ill) && \ 129 !((ire)->ire_flags & RTF_MULTIRT) && \ 130 (!((ire)->ire_type & IRE_BROADCAST) || \ 131 (ill)->ill_type == IFT_ETHER) && \ 132 (ipsec_len) == 0 && \ 133 (((ire)->ire_ipversion == IPV4_VERSION && \ 134 (start) == IP_SIMPLE_HDR_LENGTH && \ 135 ((ire)->ire_nce != NULL && \ 136 (ire)->ire_nce->nce_fp_mp != NULL && \ 137 MBLKHEAD(mp) >= MBLKL((ire)->ire_nce->nce_fp_mp))) || \ 138 ((ire)->ire_ipversion == IPV6_VERSION && \ 139 (start) == IPV6_HDR_LEN && \ 140 (ire)->ire_nce->nce_fp_mp != NULL && \ 141 MBLKHEAD(mp) >= MBLKL((ire)->ire_nce->nce_fp_mp))) && \ 142 (max_frag) >= (uint_t)((end) + (ipsec_len)) && \ 143 dohwcksum) { \ 144 _hck_flags = (ill)->ill_hcksum_capab->ill_hcksum_txflags; \ 145 } else { \ 146 _hck_flags = 0; \ 147 } \ 148 IP_CKSUM_XMIT_FAST((ire)->ire_ipversion, _hck_flags, mp, ihp, \ 149 up, proto, start, end, pseudo); \ 150 } 151 152 /* 153 * Based on the device capabilities, this macro either marks an outgoing 154 * packet with hardware checksum offload information or calculate the 155 * checksum in software. If the latter is performed, the checksum field 156 * of the dblk is cleared; otherwise it will be non-zero and contain the 157 * necessary flag(s) for the driver. 158 */ 159 #define IP_CKSUM_XMIT_FAST(ipver, hck_flags, mp, ihp, up, proto, start, \ 160 end, pseudo) { \ 161 uint32_t _sum; \ 162 /* \ 163 * Underlying interface supports hardware checksum offload for \ 164 * the payload; leave the payload checksum for the hardware to \ 165 * calculate. N.B: We only need to set up checksum info on the \ 166 * first mblk. \ 167 */ \ 168 DB_CKSUMFLAGS(mp) = 0; \ 169 if (((ipver) == IPV4_VERSION && \ 170 ((hck_flags) & HCKSUM_INET_FULL_V4)) || \ 171 ((ipver) == IPV6_VERSION && \ 172 ((hck_flags) & HCKSUM_INET_FULL_V6))) { \ 173 /* \ 174 * Hardware calculates pseudo-header, header and the \ 175 * payload checksums, so clear the checksum field in \ 176 * the protocol header. \ 177 */ \ 178 *(up) = 0; \ 179 DB_CKSUMFLAGS(mp) |= HCK_FULLCKSUM; \ 180 } else if ((hck_flags) & HCKSUM_INET_PARTIAL) { \ 181 /* \ 182 * Partial checksum offload has been enabled. Fill \ 183 * the checksum field in the protocl header with the \ 184 * pseudo-header checksum value. \ 185 */ \ 186 _sum = ((proto) == IPPROTO_UDP) ? \ 187 IP_UDP_CSUM_COMP : IP_TCP_CSUM_COMP; \ 188 _sum += *(up) + (pseudo); \ 189 _sum = (_sum & 0xFFFF) + (_sum >> 16); \ 190 *(up) = (_sum & 0xFFFF) + (_sum >> 16); \ 191 /* \ 192 * Offsets are relative to beginning of IP header. \ 193 */ \ 194 DB_CKSUMSTART(mp) = (start); \ 195 DB_CKSUMSTUFF(mp) = ((proto) == IPPROTO_UDP) ? \ 196 (start) + UDP_CHECKSUM_OFFSET : \ 197 (start) + TCP_CHECKSUM_OFFSET; \ 198 DB_CKSUMEND(mp) = (end); \ 199 DB_CKSUMFLAGS(mp) |= HCK_PARTIALCKSUM; \ 200 } else { \ 201 /* \ 202 * Software checksumming. \ 203 */ \ 204 _sum = ((proto) == IPPROTO_UDP) ? \ 205 IP_UDP_CSUM_COMP : IP_TCP_CSUM_COMP; \ 206 _sum += (pseudo); \ 207 _sum = IP_CSUM(mp, start, _sum); \ 208 *(up) = (uint16_t)(((proto) == IPPROTO_UDP) ? \ 209 (_sum ? _sum : ~_sum) : _sum); \ 210 } \ 211 /* \ 212 * Hardware supports IP header checksum offload; clear the \ 213 * contents of IP header checksum field as expected by NIC. \ 214 * Do this only if we offloaded either full or partial sum. \ 215 */ \ 216 if ((ipver) == IPV4_VERSION && DB_CKSUMFLAGS(mp) != 0 && \ 217 ((hck_flags) & HCKSUM_IPHDRCKSUM)) { \ 218 DB_CKSUMFLAGS(mp) |= HCK_IPV4_HDRCKSUM; \ 219 ((ipha_t *)(ihp))->ipha_hdr_checksum = 0; \ 220 } \ 221 } 222 223 /* 224 * Macro to inspect the checksum of a fully-reassembled incoming datagram. 225 */ 226 #define IP_CKSUM_RECV_REASS(hck_flags, off, pseudo, sum, err) { \ 227 (err) = B_FALSE; \ 228 if ((hck_flags) & HCK_FULLCKSUM) { \ 229 /* \ 230 * The sum of all fragment checksums should \ 231 * result in -0 (0xFFFF) or otherwise invalid. \ 232 */ \ 233 if ((sum) != 0xFFFF) \ 234 (err) = B_TRUE; \ 235 } else if ((hck_flags) & HCK_PARTIALCKSUM) { \ 236 (sum) += (pseudo); \ 237 (sum) = ((sum) & 0xFFFF) + ((sum) >> 16); \ 238 (sum) = ((sum) & 0xFFFF) + ((sum) >> 16); \ 239 if (~(sum) & 0xFFFF) \ 240 (err) = B_TRUE; \ 241 } else if (((sum) = IP_CSUM(mp, off, pseudo)) != 0) { \ 242 (err) = B_TRUE; \ 243 } \ 244 } 245 246 /* 247 * This macro inspects an incoming packet to see if the checksum value 248 * contained in it is valid; if the hardware has provided the information, 249 * the value is verified, otherwise it performs software checksumming. 250 * The checksum value is returned to caller. 251 */ 252 #define IP_CKSUM_RECV(hck_flags, sum, cksum_start, ulph_off, mp, mp1, err) { \ 253 int32_t _len; \ 254 \ 255 (err) = B_FALSE; \ 256 if ((hck_flags) & HCK_FULLCKSUM) { \ 257 /* \ 258 * Full checksum has been computed by the hardware \ 259 * and has been attached. If the driver wants us to \ 260 * verify the correctness of the attached value, in \ 261 * order to protect against faulty hardware, compare \ 262 * it against -0 (0xFFFF) to see if it's valid. \ 263 */ \ 264 (sum) = DB_CKSUM16(mp); \ 265 if (!((hck_flags) & HCK_FULLCKSUM_OK) && (sum) != 0xFFFF) \ 266 (err) = B_TRUE; \ 267 } else if (((hck_flags) & HCK_PARTIALCKSUM) && \ 268 ((mp1) == NULL || (mp1)->b_cont == NULL) && \ 269 (ulph_off) >= DB_CKSUMSTART(mp) && \ 270 ((_len = (ulph_off) - DB_CKSUMSTART(mp)) & 1) == 0) { \ 271 uint32_t _adj; \ 272 /* \ 273 * Partial checksum has been calculated by hardware \ 274 * and attached to the packet; in addition, any \ 275 * prepended extraneous data is even byte aligned, \ 276 * and there are at most two mblks associated with \ 277 * the packet. If any such data exists, we adjust \ 278 * the checksum; also take care any postpended data. \ 279 */ \ 280 IP_ADJCKSUM_PARTIAL(cksum_start, mp, mp1, _len, _adj); \ 281 /* \ 282 * One's complement subtract extraneous checksum \ 283 */ \ 284 (sum) += DB_CKSUM16(mp); \ 285 if (_adj >= (sum)) \ 286 (sum) = ~(_adj - (sum)) & 0xFFFF; \ 287 else \ 288 (sum) -= _adj; \ 289 (sum) = ((sum) & 0xFFFF) + ((int)(sum) >> 16); \ 290 (sum) = ((sum) & 0xFFFF) + ((int)(sum) >> 16); \ 291 if (~(sum) & 0xFFFF) \ 292 (err) = B_TRUE; \ 293 } else if (((sum) = IP_CSUM(mp, ulph_off, sum)) != 0) { \ 294 (err) = B_TRUE; \ 295 } \ 296 } 297 298 /* 299 * Macro to adjust a given checksum value depending on any prepended 300 * or postpended data on the packet. It expects the start offset to 301 * begin at an even boundary and that the packet consists of at most 302 * two mblks. 303 */ 304 #define IP_ADJCKSUM_PARTIAL(cksum_start, mp, mp1, len, adj) { \ 305 /* \ 306 * Prepended extraneous data; adjust checksum. \ 307 */ \ 308 if ((len) > 0) \ 309 (adj) = IP_BCSUM_PARTIAL(cksum_start, len, 0); \ 310 else \ 311 (adj) = 0; \ 312 /* \ 313 * len is now the total length of mblk(s) \ 314 */ \ 315 (len) = MBLKL(mp); \ 316 if ((mp1) == NULL) \ 317 (mp1) = (mp); \ 318 else \ 319 (len) += MBLKL(mp1); \ 320 /* \ 321 * Postpended extraneous data; adjust checksum. \ 322 */ \ 323 if (((len) = (DB_CKSUMEND(mp) - len)) > 0) { \ 324 uint32_t _pad; \ 325 \ 326 _pad = IP_BCSUM_PARTIAL((mp1)->b_wptr, len, 0); \ 327 /* \ 328 * If the postpended extraneous data was odd \ 329 * byte aligned, swap resulting checksum bytes. \ 330 */ \ 331 if ((uintptr_t)(mp1)->b_wptr & 1) \ 332 (adj) += ((_pad << 8) & 0xFFFF) | (_pad >> 8); \ 333 else \ 334 (adj) += _pad; \ 335 (adj) = ((adj) & 0xFFFF) + ((int)(adj) >> 16); \ 336 } \ 337 } 338 339 #define ILL_MDT_CAPABLE(ill) \ 340 (((ill)->ill_capabilities & ILL_CAPAB_MDT) != 0) 341 342 /* 343 * ioctl identifier and structure for Multidata Transmit update 344 * private M_CTL communication from IP to ULP. 345 */ 346 #define MDT_IOC_INFO_UPDATE (('M' << 8) + 1020) 347 348 typedef struct ip_mdt_info_s { 349 uint_t mdt_info_id; /* MDT_IOC_INFO_UPDATE */ 350 ill_mdt_capab_t mdt_capab; /* ILL MDT capabilities */ 351 } ip_mdt_info_t; 352 353 /* 354 * Macro that determines whether or not a given ILL is allowed for MDT. 355 */ 356 #define ILL_MDT_USABLE(ill) \ 357 (ILL_MDT_CAPABLE(ill) && \ 358 ill->ill_mdt_capab != NULL && \ 359 ill->ill_mdt_capab->ill_mdt_version == MDT_VERSION_2 && \ 360 ill->ill_mdt_capab->ill_mdt_on != 0) 361 362 #define ILL_LSO_CAPABLE(ill) \ 363 (((ill)->ill_capabilities & ILL_CAPAB_DLD_LSO) != 0) 364 365 /* 366 * ioctl identifier and structure for Large Segment Offload 367 * private M_CTL communication from IP to ULP. 368 */ 369 #define LSO_IOC_INFO_UPDATE (('L' << 24) + ('S' << 16) + ('O' << 8)) 370 371 typedef struct ip_lso_info_s { 372 uint_t lso_info_id; /* LSO_IOC_INFO_UPDATE */ 373 ill_lso_capab_t lso_capab; /* ILL LSO capabilities */ 374 } ip_lso_info_t; 375 376 /* 377 * Macro that determines whether or not a given ILL is allowed for LSO. 378 */ 379 #define ILL_LSO_USABLE(ill) \ 380 (ILL_LSO_CAPABLE(ill) && \ 381 ill->ill_lso_capab != NULL && \ 382 ill->ill_lso_capab->ill_lso_on != 0) 383 384 #define ILL_LSO_TCP_USABLE(ill) \ 385 (ILL_LSO_USABLE(ill) && \ 386 ill->ill_lso_capab->ill_lso_flags & DLD_LSO_TX_BASIC_TCP_IPV4) 387 388 /* 389 * Macro that determines whether or not a given CONN may be considered 390 * for fast path prior to proceeding further with LSO or Multidata. 391 */ 392 #define CONN_IS_LSO_MD_FASTPATH(connp) \ 393 ((connp)->conn_dontroute == 0 && /* SO_DONTROUTE */ \ 394 !((connp)->conn_nexthop_set) && /* IP_NEXTHOP */ \ 395 (connp)->conn_nofailover_ill == NULL && /* IPIF_NOFAILOVER */ \ 396 (connp)->conn_outgoing_pill == NULL && /* IP{V6}_BOUND_PIF */ \ 397 (connp)->conn_outgoing_ill == NULL) /* IP{V6}_BOUND_IF */ 398 399 /* Definitons for fragmenting IP packets using MDT. */ 400 401 /* 402 * Smaller and private version of pdescinfo_t used specifically for IP, 403 * which allows for only a single payload span per packet. 404 */ 405 typedef struct ip_pdescinfo_s PDESCINFO_STRUCT(2) ip_pdescinfo_t; 406 407 /* 408 * Macro version of ip_can_frag_mdt() which avoids the function call if we 409 * only examine a single message block. 410 */ 411 #define IP_CAN_FRAG_MDT(mp, hdr_len, len) \ 412 (((mp)->b_cont == NULL) ? \ 413 (MBLKL(mp) >= ((hdr_len) + ip_wput_frag_mdt_min)) : \ 414 ip_can_frag_mdt((mp), (hdr_len), (len))) 415 416 /* 417 * Macro that determines whether or not a given IPC requires 418 * outbound IPSEC processing. 419 */ 420 #define CONN_IPSEC_OUT_ENCAPSULATED(connp) \ 421 ((connp)->conn_out_enforce_policy || \ 422 ((connp)->conn_latch != NULL && \ 423 (connp)->conn_latch->ipl_out_policy != NULL)) 424 425 /* 426 * These are used by the synchronous streams code in tcp and udp. 427 * When we set the flags for a wakeup from a synchronous stream we 428 * always set RSLEEP in sd_wakeq, even if we have a read thread waiting 429 * to do the io. This is in case the read thread gets interrupted 430 * before completing the io. The RSLEEP flag in sd_wakeq is used to 431 * indicate that there is data available at the synchronous barrier. 432 * The assumption is that subsequent functions calls through rwnext() 433 * will reset sd_wakeq appropriately. 434 */ 435 #define STR_WAKEUP_CLEAR(stp) { \ 436 mutex_enter(&stp->sd_lock); \ 437 stp->sd_wakeq &= ~RSLEEP; \ 438 mutex_exit(&stp->sd_lock); \ 439 } 440 441 #define STR_WAKEUP_SET(stp) { \ 442 mutex_enter(&stp->sd_lock); \ 443 if (stp->sd_flag & RSLEEP) { \ 444 stp->sd_flag &= ~RSLEEP; \ 445 cv_broadcast(&_RD(stp->sd_wrq)->q_wait); \ 446 } \ 447 stp->sd_wakeq |= RSLEEP; \ 448 mutex_exit(&stp->sd_lock); \ 449 } 450 451 /* 452 * Combined wakeup and sendsig to avoid dropping and reacquiring the 453 * sd_lock. The list of messages waiting at the synchronous barrier is 454 * supplied in order to determine whether a wakeup needs to occur. We 455 * only send a wakeup to the application when necessary, i.e. during 456 * the first enqueue when the received messages list will be NULL. 457 */ 458 #define STR_WAKEUP_SENDSIG(stp, rcv_list) { \ 459 int _events; \ 460 mutex_enter(&stp->sd_lock); \ 461 if (rcv_list == NULL) { \ 462 if (stp->sd_flag & RSLEEP) { \ 463 stp->sd_flag &= ~RSLEEP; \ 464 cv_broadcast(&_RD(stp->sd_wrq)->q_wait); \ 465 } \ 466 stp->sd_wakeq |= RSLEEP; \ 467 } \ 468 if ((_events = stp->sd_sigflags & (S_INPUT | S_RDNORM)) != 0) \ 469 strsendsig(stp->sd_siglist, _events, 0, 0); \ 470 if (stp->sd_rput_opt & SR_POLLIN) { \ 471 stp->sd_rput_opt &= ~SR_POLLIN; \ 472 mutex_exit(&stp->sd_lock); \ 473 pollwakeup(&stp->sd_pollist, POLLIN | POLLRDNORM); \ 474 } else { \ 475 mutex_exit(&stp->sd_lock); \ 476 } \ 477 } 478 479 #define CONN_UDP_SYNCSTR(connp) \ 480 (IPCL_IS_UDP(connp) && (connp)->conn_udp->udp_direct_sockfs) 481 482 /* 483 * Macro that checks whether or not a particular UDP conn is 484 * flow-controlling on the read-side. If udp module is directly 485 * above ip, check to see if the drain queue is full; note here 486 * that we check this without any lock protection because this 487 * is a coarse granularity inbound flow-control. If the module 488 * above ip is not udp, then use canputnext to determine the 489 * flow-control. 490 * 491 * Note that these checks are done after the conn is found in 492 * the UDP fanout table. 493 * FIXME? Might be faster to check both udp_drain_qfull and canputnext. 494 */ 495 #define CONN_UDP_FLOWCTLD(connp) \ 496 (CONN_UDP_SYNCSTR(connp) ? \ 497 (connp)->conn_udp->udp_drain_qfull : \ 498 !canputnext((connp)->conn_rq)) 499 500 /* Macro that follows definitions of flags for mac_tx() (see mac_client.h) */ 501 #define IP_DROP_ON_NO_DESC 0x01 /* Equivalent to MAC_DROP_ON_NO_DESC */ 502 503 #define ILL_DIRECT_CAPABLE(ill) \ 504 (((ill)->ill_capabilities & ILL_CAPAB_DLD_DIRECT) != 0) 505 506 #define ILL_SEND_TX(ill, ire, hint, mp, flag) { \ 507 if (ILL_DIRECT_CAPABLE(ill) && DB_TYPE(mp) == M_DATA) { \ 508 ill_dld_direct_t *idd; \ 509 \ 510 idd = &(ill)->ill_dld_capab->idc_direct; \ 511 /* \ 512 * Send the packet directly to DLD, where it \ 513 * may be queued depending on the availability \ 514 * of transmit resources at the media layer. \ 515 * Ignore the returned value for the time being \ 516 * In future, we may want to take this into \ 517 * account and flow control the TCP. \ 518 */ \ 519 (void) idd->idd_tx_df(idd->idd_tx_dh, mp, \ 520 (uintptr_t)(hint), flag); \ 521 } else { \ 522 putnext((ire)->ire_stq, mp); \ 523 } \ 524 } 525 526 #define MBLK_RX_FANOUT_SLOWPATH(mp, ipha) \ 527 (DB_TYPE(mp) != M_DATA || DB_REF(mp) != 1 || !OK_32PTR(ipha) || \ 528 (((uchar_t *)ipha + IP_SIMPLE_HDR_LENGTH) >= (mp)->b_wptr)) 529 530 /* 531 * In non-global zone exclusive IP stacks, data structures such as IRE 532 * entries pretend that they're in the global zone. The following 533 * macro evaluates to the real zoneid instead of a pretend 534 * GLOBAL_ZONEID. 535 */ 536 #define IP_REAL_ZONEID(zoneid, ipst) \ 537 (((zoneid) == GLOBAL_ZONEID) ? \ 538 netstackid_to_zoneid((ipst)->ips_netstack->netstack_stackid) : \ 539 (zoneid)) 540 541 extern int ip_wput_frag_mdt_min; 542 extern boolean_t ip_can_frag_mdt(mblk_t *, ssize_t, ssize_t); 543 extern mblk_t *ip_prepend_zoneid(mblk_t *, zoneid_t, ip_stack_t *); 544 extern void ill_flow_enable(void *, ip_mac_tx_cookie_t); 545 extern zoneid_t ip_get_zoneid_v4(ipaddr_t, mblk_t *, ip_stack_t *, zoneid_t); 546 extern zoneid_t ip_get_zoneid_v6(in6_addr_t *, mblk_t *, const ill_t *, 547 ip_stack_t *, zoneid_t); 548 549 #endif /* _KERNEL */ 550 551 #ifdef __cplusplus 552 } 553 #endif 554 555 #endif /* _INET_IP_IMPL_H */ 556