1 /*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21
22 /*
23 * Copyright (c) 2009, 2010, Oracle and/or its affiliates. All rights reserved.
24 *
25 * Copyright 2011 Nexenta Systems, Inc. All rights reserved.
26 * Copyright 2019 Joyent, Inc.
27 * Copyright 2024 Oxide Computer Company
28 */
29 /* Copyright (c) 1990 Mentat Inc. */
30
31 #include <sys/types.h>
32 #include <sys/stream.h>
33 #include <sys/dlpi.h>
34 #include <sys/stropts.h>
35 #include <sys/sysmacros.h>
36 #include <sys/strsubr.h>
37 #include <sys/strlog.h>
38 #include <sys/strsun.h>
39 #include <sys/zone.h>
40 #define _SUN_TPI_VERSION 2
41 #include <sys/tihdr.h>
42 #include <sys/xti_inet.h>
43 #include <sys/ddi.h>
44 #include <sys/sunddi.h>
45 #include <sys/cmn_err.h>
46 #include <sys/debug.h>
47 #include <sys/kobj.h>
48 #include <sys/modctl.h>
49 #include <sys/atomic.h>
50 #include <sys/policy.h>
51 #include <sys/priv.h>
52
53 #include <sys/systm.h>
54 #include <sys/param.h>
55 #include <sys/kmem.h>
56 #include <sys/sdt.h>
57 #include <sys/socket.h>
58 #include <sys/vtrace.h>
59 #include <sys/isa_defs.h>
60 #include <sys/mac.h>
61 #include <sys/mac_client.h>
62 #include <net/if.h>
63 #include <net/if_arp.h>
64 #include <net/route.h>
65 #include <sys/sockio.h>
66 #include <netinet/in.h>
67 #include <net/if_dl.h>
68
69 #include <inet/common.h>
70 #include <inet/mi.h>
71 #include <inet/mib2.h>
72 #include <inet/nd.h>
73 #include <inet/arp.h>
74 #include <inet/snmpcom.h>
75 #include <inet/kstatcom.h>
76
77 #include <netinet/igmp_var.h>
78 #include <netinet/ip6.h>
79 #include <netinet/icmp6.h>
80 #include <netinet/sctp.h>
81
82 #include <inet/ip.h>
83 #include <inet/ip_impl.h>
84 #include <inet/ip6.h>
85 #include <inet/ip6_asp.h>
86 #include <inet/optcom.h>
87 #include <inet/tcp.h>
88 #include <inet/tcp_impl.h>
89 #include <inet/ip_multi.h>
90 #include <inet/ip_if.h>
91 #include <inet/ip_ire.h>
92 #include <inet/ip_ftable.h>
93 #include <inet/ip_rts.h>
94 #include <inet/ip_ndp.h>
95 #include <inet/ip_listutils.h>
96 #include <netinet/igmp.h>
97 #include <netinet/ip_mroute.h>
98 #include <inet/ipp_common.h>
99
100 #include <net/pfkeyv2.h>
101 #include <inet/sadb.h>
102 #include <inet/ipsec_impl.h>
103 #include <inet/ipdrop.h>
104 #include <inet/ip_netinfo.h>
105 #include <inet/ilb_ip.h>
106 #include <sys/squeue_impl.h>
107 #include <sys/squeue.h>
108
109 #include <sys/ethernet.h>
110 #include <net/if_types.h>
111 #include <sys/cpuvar.h>
112
113 #include <ipp/ipp.h>
114 #include <ipp/ipp_impl.h>
115 #include <ipp/ipgpc/ipgpc.h>
116
117 #include <sys/pattr.h>
118 #include <inet/ipclassifier.h>
119 #include <inet/sctp_ip.h>
120 #include <inet/sctp/sctp_impl.h>
121 #include <inet/udp_impl.h>
122 #include <sys/sunddi.h>
123
124 #include <sys/tsol/label.h>
125 #include <sys/tsol/tnet.h>
126
127 #include <sys/clock_impl.h> /* For LBOLT_FASTPATH{,64} */
128
129 #ifdef DEBUG
130 extern boolean_t skip_sctp_cksum;
131 #endif
132
133 static void ip_input_local_v4(ire_t *, mblk_t *, ipha_t *,
134 ip_recv_attr_t *);
135
136 static void ip_input_broadcast_v4(ire_t *, mblk_t *, ipha_t *,
137 ip_recv_attr_t *);
138 static void ip_input_multicast_v4(ire_t *, mblk_t *, ipha_t *,
139 ip_recv_attr_t *);
140
141 #pragma inline(ip_input_common_v4, ip_input_local_v4, ip_forward_xmit_v4)
142
143 /*
144 * Direct read side procedure capable of dealing with chains. GLDv3 based
145 * drivers call this function directly with mblk chains while STREAMS
146 * read side procedure ip_rput() calls this for single packet with ip_ring
147 * set to NULL to process one packet at a time.
148 *
149 * The ill will always be valid if this function is called directly from
150 * the driver.
151 *
152 * If this chain is part of a VLAN stream, then the VLAN tag is
153 * stripped from the MAC header before being delivered to this
154 * function.
155 *
156 * If the IP header in packet is not 32-bit aligned, every message in the
157 * chain will be aligned before further operations. This is required on SPARC
158 * platform.
159 */
160 void
ip_input(ill_t * ill,ill_rx_ring_t * ip_ring,mblk_t * mp_chain,struct mac_header_info_s * mhip)161 ip_input(ill_t *ill, ill_rx_ring_t *ip_ring, mblk_t *mp_chain,
162 struct mac_header_info_s *mhip)
163 {
164 (void) ip_input_common_v4(ill, ip_ring, mp_chain, mhip, NULL, NULL,
165 NULL);
166 }
167
168 /*
169 * ip_accept_tcp() - This function is called by the squeue when it retrieves
170 * a chain of packets in the poll mode. The packets have gone through the
171 * data link processing but not IP processing. For performance and latency
172 * reasons, the squeue wants to process the chain in line instead of feeding
173 * it back via ip_input path.
174 *
175 * We set up the ip_recv_attr_t with IRAF_TARGET_SQP to that ip_fanout_v4
176 * will pass back any TCP packets matching the target sqp to
177 * ip_input_common_v4 using ira_target_sqp_mp. Other packets are handled by
178 * ip_input_v4 and ip_fanout_v4 as normal.
179 * The TCP packets that match the target squeue are returned to the caller
180 * as a b_next chain after each packet has been prepend with an mblk
181 * from ip_recv_attr_to_mblk.
182 */
183 mblk_t *
ip_accept_tcp(ill_t * ill,ill_rx_ring_t * ip_ring,squeue_t * target_sqp,mblk_t * mp_chain,mblk_t ** last,uint_t * cnt)184 ip_accept_tcp(ill_t *ill, ill_rx_ring_t *ip_ring, squeue_t *target_sqp,
185 mblk_t *mp_chain, mblk_t **last, uint_t *cnt)
186 {
187 return (ip_input_common_v4(ill, ip_ring, mp_chain, NULL, target_sqp,
188 last, cnt));
189 }
190
191 /*
192 * Used by ip_input and ip_accept_tcp
193 * The last three arguments are only used by ip_accept_tcp, and mhip is
194 * only used by ip_input.
195 */
196 mblk_t *
ip_input_common_v4(ill_t * ill,ill_rx_ring_t * ip_ring,mblk_t * mp_chain,struct mac_header_info_s * mhip,squeue_t * target_sqp,mblk_t ** last,uint_t * cnt)197 ip_input_common_v4(ill_t *ill, ill_rx_ring_t *ip_ring, mblk_t *mp_chain,
198 struct mac_header_info_s *mhip, squeue_t *target_sqp,
199 mblk_t **last, uint_t *cnt)
200 {
201 mblk_t *mp;
202 ipha_t *ipha;
203 ip_recv_attr_t iras; /* Receive attributes */
204 rtc_t rtc;
205 iaflags_t chain_flags = 0; /* Fixed for chain */
206 mblk_t *ahead = NULL; /* Accepted head */
207 mblk_t *atail = NULL; /* Accepted tail */
208 uint_t acnt = 0; /* Accepted count */
209
210 ASSERT(mp_chain != NULL);
211 ASSERT(ill != NULL);
212
213 /* These ones do not change as we loop over packets */
214 iras.ira_ill = iras.ira_rill = ill;
215 iras.ira_ruifindex = ill->ill_phyint->phyint_ifindex;
216 iras.ira_rifindex = iras.ira_ruifindex;
217 iras.ira_sqp = NULL;
218 iras.ira_ring = ip_ring;
219 /* For ECMP and outbound transmit ring selection */
220 iras.ira_xmit_hint = ILL_RING_TO_XMIT_HINT(ip_ring);
221
222 iras.ira_target_sqp = target_sqp;
223 iras.ira_target_sqp_mp = NULL;
224 if (target_sqp != NULL)
225 chain_flags |= IRAF_TARGET_SQP;
226
227 /*
228 * We try to have a mhip pointer when possible, but
229 * it might be NULL in some cases. In those cases we
230 * have to assume unicast.
231 */
232 iras.ira_mhip = mhip;
233 iras.ira_flags = 0;
234 if (mhip != NULL) {
235 switch (mhip->mhi_dsttype) {
236 case MAC_ADDRTYPE_MULTICAST :
237 chain_flags |= IRAF_L2DST_MULTICAST;
238 break;
239 case MAC_ADDRTYPE_BROADCAST :
240 chain_flags |= IRAF_L2DST_BROADCAST;
241 break;
242 }
243 }
244
245 /*
246 * Initialize the one-element route cache.
247 *
248 * We do ire caching from one iteration to
249 * another. In the event the packet chain contains
250 * all packets from the same dst, this caching saves
251 * an ire_route_recursive for each of the succeeding
252 * packets in a packet chain.
253 */
254 rtc.rtc_ire = NULL;
255 rtc.rtc_ipaddr = INADDR_ANY;
256
257 /* Loop over b_next */
258 for (mp = mp_chain; mp != NULL; mp = mp_chain) {
259 mp_chain = mp->b_next;
260 mp->b_next = NULL;
261
262 ASSERT(DB_TYPE(mp) == M_DATA);
263
264
265 /*
266 * if db_ref > 1 then copymsg and free original. Packet
267 * may be changed and we do not want the other entity
268 * who has a reference to this message to trip over the
269 * changes. This is a blind change because trying to
270 * catch all places that might change the packet is too
271 * difficult.
272 *
273 * This corresponds to the fast path case, where we have
274 * a chain of M_DATA mblks. We check the db_ref count
275 * of only the 1st data block in the mblk chain. There
276 * doesn't seem to be a reason why a device driver would
277 * send up data with varying db_ref counts in the mblk
278 * chain. In any case the Fast path is a private
279 * interface, and our drivers don't do such a thing.
280 * Given the above assumption, there is no need to walk
281 * down the entire mblk chain (which could have a
282 * potential performance problem)
283 *
284 * The "(DB_REF(mp) > 1)" check was moved from ip_rput()
285 * to here because of exclusive ip stacks and vnics.
286 * Packets transmitted from exclusive stack over vnic
287 * can have db_ref > 1 and when it gets looped back to
288 * another vnic in a different zone, you have ip_input()
289 * getting dblks with db_ref > 1. So if someone
290 * complains of TCP performance under this scenario,
291 * take a serious look here on the impact of copymsg().
292 */
293 if (DB_REF(mp) > 1) {
294 if ((mp = ip_fix_dbref(mp, &iras)) == NULL) {
295 /* mhip might point into 1st packet in chain */
296 iras.ira_mhip = NULL;
297 continue;
298 }
299 }
300
301 /*
302 * IP header ptr not aligned?
303 * OR IP header not complete in first mblk
304 */
305 ipha = (ipha_t *)mp->b_rptr;
306 if (!OK_32PTR(ipha) || MBLKL(mp) < IP_SIMPLE_HDR_LENGTH) {
307 mp = ip_check_and_align_header(mp, IP_SIMPLE_HDR_LENGTH,
308 &iras);
309 if (mp == NULL) {
310 /* mhip might point into 1st packet in chain */
311 iras.ira_mhip = NULL;
312 continue;
313 }
314 ipha = (ipha_t *)mp->b_rptr;
315 }
316
317 /* Protect against a mix of Ethertypes and IP versions */
318 if (IPH_HDR_VERSION(ipha) != IPV4_VERSION) {
319 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInHdrErrors);
320 ip_drop_input("ipIfStatsInHdrErrors", mp, ill);
321 freemsg(mp);
322 /* mhip might point into 1st packet in the chain. */
323 iras.ira_mhip = NULL;
324 continue;
325 }
326
327 /*
328 * Check for Martian addrs; we have to explicitly
329 * test for for zero dst since this is also used as
330 * an indication that the rtc is not used.
331 */
332 if (ipha->ipha_dst == INADDR_ANY) {
333 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInAddrErrors);
334 ip_drop_input("ipIfStatsInAddrErrors", mp, ill);
335 freemsg(mp);
336 /* mhip might point into 1st packet in the chain. */
337 iras.ira_mhip = NULL;
338 continue;
339 }
340
341 /*
342 * Keep L2SRC from a previous packet in chain since mhip
343 * might point into an earlier packet in the chain.
344 * Keep IRAF_VERIFIED_SRC to avoid redoing broadcast
345 * source check in forwarding path.
346 */
347 chain_flags |= (iras.ira_flags &
348 (IRAF_L2SRC_SET|IRAF_VERIFIED_SRC));
349
350 iras.ira_flags = IRAF_IS_IPV4 | IRAF_VERIFY_IP_CKSUM |
351 IRAF_VERIFY_ULP_CKSUM | chain_flags;
352 iras.ira_free_flags = 0;
353 iras.ira_cred = NULL;
354 iras.ira_cpid = NOPID;
355 iras.ira_tsl = NULL;
356 iras.ira_zoneid = ALL_ZONES; /* Default for forwarding */
357
358 /*
359 * We must count all incoming packets, even if they end
360 * up being dropped later on. Defer counting bytes until
361 * we have the whole IP header in first mblk.
362 */
363 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInReceives);
364
365 iras.ira_pktlen = ntohs(ipha->ipha_length);
366 UPDATE_MIB(ill->ill_ip_mib, ipIfStatsHCInOctets,
367 iras.ira_pktlen);
368 iras.ira_ttl = ipha->ipha_ttl;
369
370 /*
371 * Call one of:
372 * ill_input_full_v4
373 * ill_input_short_v4
374 * The former is used in unusual cases. See ill_set_inputfn().
375 */
376 (*ill->ill_inputfn)(mp, ipha, &ipha->ipha_dst, &iras, &rtc);
377
378 /* Any references to clean up? No hold on ira_ill */
379 if (iras.ira_flags & (IRAF_IPSEC_SECURE|IRAF_SYSTEM_LABELED))
380 ira_cleanup(&iras, B_FALSE);
381
382 if (iras.ira_target_sqp_mp != NULL) {
383 /* Better be called from ip_accept_tcp */
384 ASSERT(target_sqp != NULL);
385
386 /* Found one packet to accept */
387 mp = iras.ira_target_sqp_mp;
388 iras.ira_target_sqp_mp = NULL;
389 ASSERT(ip_recv_attr_is_mblk(mp));
390
391 if (atail != NULL)
392 atail->b_next = mp;
393 else
394 ahead = mp;
395 atail = mp;
396 acnt++;
397 mp = NULL;
398 }
399 /* mhip might point into 1st packet in the chain. */
400 iras.ira_mhip = NULL;
401 }
402 /* Any remaining references to the route cache? */
403 if (rtc.rtc_ire != NULL) {
404 ASSERT(rtc.rtc_ipaddr != INADDR_ANY);
405 ire_refrele(rtc.rtc_ire);
406 }
407
408 if (ahead != NULL) {
409 /* Better be called from ip_accept_tcp */
410 ASSERT(target_sqp != NULL);
411 *last = atail;
412 *cnt = acnt;
413 return (ahead);
414 }
415
416 return (NULL);
417 }
418
419 /*
420 * This input function is used when
421 * - is_system_labeled()
422 * - CGTP filtering
423 * - DHCP unicast before we have an IP address configured
424 * - there is an listener for IPPROTO_RSVP
425 */
426 void
ill_input_full_v4(mblk_t * mp,void * iph_arg,void * nexthop_arg,ip_recv_attr_t * ira,rtc_t * rtc)427 ill_input_full_v4(mblk_t *mp, void *iph_arg, void *nexthop_arg,
428 ip_recv_attr_t *ira, rtc_t *rtc)
429 {
430 ipha_t *ipha = (ipha_t *)iph_arg;
431 ipaddr_t nexthop = *(ipaddr_t *)nexthop_arg;
432 ill_t *ill = ira->ira_ill;
433 ip_stack_t *ipst = ill->ill_ipst;
434 int cgtp_flt_pkt;
435
436 ASSERT(ira->ira_tsl == NULL);
437
438 /*
439 * Attach any necessary label information to
440 * this packet
441 */
442 if (is_system_labeled()) {
443 ira->ira_flags |= IRAF_SYSTEM_LABELED;
444
445 /*
446 * This updates ira_cred, ira_tsl and ira_free_flags based
447 * on the label.
448 */
449 if (!tsol_get_pkt_label(mp, IPV4_VERSION, ira)) {
450 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
451 ip_drop_input("ipIfStatsInDiscards", mp, ill);
452 freemsg(mp);
453 return;
454 }
455 /* Note that ira_tsl can be NULL here. */
456
457 /* tsol_get_pkt_label sometimes does pullupmsg */
458 ipha = (ipha_t *)mp->b_rptr;
459 }
460
461 /*
462 * Invoke the CGTP (multirouting) filtering module to process
463 * the incoming packet. Packets identified as duplicates
464 * must be discarded. Filtering is active only if the
465 * the ip_cgtp_filter ndd variable is non-zero.
466 */
467 cgtp_flt_pkt = CGTP_IP_PKT_NOT_CGTP;
468 if (ipst->ips_ip_cgtp_filter &&
469 ipst->ips_ip_cgtp_filter_ops != NULL) {
470 netstackid_t stackid;
471
472 stackid = ipst->ips_netstack->netstack_stackid;
473 /*
474 * CGTP and IPMP are mutually exclusive so
475 * phyint_ifindex is fine here.
476 */
477 cgtp_flt_pkt =
478 ipst->ips_ip_cgtp_filter_ops->cfo_filter(stackid,
479 ill->ill_phyint->phyint_ifindex, mp);
480 if (cgtp_flt_pkt == CGTP_IP_PKT_DUPLICATE) {
481 ip_drop_input("CGTP_IP_PKT_DUPLICATE", mp, ill);
482 freemsg(mp);
483 return;
484 }
485 }
486
487 /*
488 * Brutal hack for DHCPv4 unicast: RFC2131 allows a DHCP
489 * server to unicast DHCP packets to a DHCP client using the
490 * IP address it is offering to the client. This can be
491 * disabled through the "broadcast bit", but not all DHCP
492 * servers honor that bit. Therefore, to interoperate with as
493 * many DHCP servers as possible, the DHCP client allows the
494 * server to unicast, but we treat those packets as broadcast
495 * here. Note that we don't rewrite the packet itself since
496 * (a) that would mess up the checksums and (b) the DHCP
497 * client conn is bound to INADDR_ANY so ip_fanout_udp() will
498 * hand it the packet regardless.
499 */
500 if (ill->ill_dhcpinit != 0 &&
501 ipha->ipha_version_and_hdr_length == IP_SIMPLE_HDR_VERSION &&
502 ipha->ipha_protocol == IPPROTO_UDP) {
503 udpha_t *udpha;
504
505 ipha = ip_pullup(mp, sizeof (ipha_t) + sizeof (udpha_t), ira);
506 if (ipha == NULL) {
507 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
508 ip_drop_input("ipIfStatsInDiscards - dhcp", mp, ill);
509 freemsg(mp);
510 return;
511 }
512 /* Reload since pullupmsg() can change b_rptr. */
513 udpha = (udpha_t *)&ipha[1];
514
515 if (ntohs(udpha->uha_dst_port) == IPPORT_BOOTPC) {
516 DTRACE_PROBE2(ip4__dhcpinit__pkt, ill_t *, ill,
517 mblk_t *, mp);
518 /*
519 * This assumes that we deliver to all conns for
520 * multicast and broadcast packets.
521 */
522 nexthop = INADDR_BROADCAST;
523 ira->ira_flags |= IRAF_DHCP_UNICAST;
524 }
525 }
526
527 /*
528 * If rsvpd is running, let RSVP daemon handle its processing
529 * and forwarding of RSVP multicast/unicast packets.
530 * If rsvpd is not running but mrouted is running, RSVP
531 * multicast packets are forwarded as multicast traffic
532 * and RSVP unicast packets are forwarded by unicast router.
533 * If neither rsvpd nor mrouted is running, RSVP multicast
534 * packets are not forwarded, but the unicast packets are
535 * forwarded like unicast traffic.
536 */
537 if (ipha->ipha_protocol == IPPROTO_RSVP &&
538 ipst->ips_ipcl_proto_fanout_v4[IPPROTO_RSVP].connf_head != NULL) {
539 /* RSVP packet and rsvpd running. Treat as ours */
540 ip2dbg(("ip_input: RSVP for us: 0x%x\n", ntohl(nexthop)));
541 /*
542 * We use a multicast address to get the packet to
543 * ire_recv_multicast_v4. There will not be a membership
544 * check since we set IRAF_RSVP
545 */
546 nexthop = htonl(INADDR_UNSPEC_GROUP);
547 ira->ira_flags |= IRAF_RSVP;
548 }
549
550 ill_input_short_v4(mp, ipha, &nexthop, ira, rtc);
551 }
552
553 /*
554 * This is the tail-end of the full receive side packet handling.
555 * It can be used directly when the configuration is simple.
556 */
557 void
ill_input_short_v4(mblk_t * mp,void * iph_arg,void * nexthop_arg,ip_recv_attr_t * ira,rtc_t * rtc)558 ill_input_short_v4(mblk_t *mp, void *iph_arg, void *nexthop_arg,
559 ip_recv_attr_t *ira, rtc_t *rtc)
560 {
561 ire_t *ire;
562 uint_t opt_len;
563 ill_t *ill = ira->ira_ill;
564 ip_stack_t *ipst = ill->ill_ipst;
565 uint_t pkt_len;
566 ssize_t len;
567 ipha_t *ipha = (ipha_t *)iph_arg;
568 ipaddr_t nexthop = *(ipaddr_t *)nexthop_arg;
569 ilb_stack_t *ilbs = ipst->ips_netstack->netstack_ilb;
570 uint_t irr_flags;
571 #define rptr ((uchar_t *)ipha)
572
573 ASSERT(DB_TYPE(mp) == M_DATA);
574
575 /*
576 * The following test for loopback is faster than
577 * IP_LOOPBACK_ADDR(), because it avoids any bitwise
578 * operations.
579 * Note that these addresses are always in network byte order
580 */
581 if (((*(uchar_t *)&ipha->ipha_dst) == IN_LOOPBACKNET) ||
582 ((*(uchar_t *)&ipha->ipha_src) == IN_LOOPBACKNET)) {
583 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInAddrErrors);
584 ip_drop_input("ipIfStatsInAddrErrors", mp, ill);
585 freemsg(mp);
586 return;
587 }
588
589 len = mp->b_wptr - rptr;
590 pkt_len = ira->ira_pktlen;
591
592 /* multiple mblk or too short */
593 len -= pkt_len;
594 if (len != 0) {
595 mp = ip_check_length(mp, rptr, len, pkt_len,
596 IP_SIMPLE_HDR_LENGTH, ira);
597 if (mp == NULL)
598 return;
599 ipha = (ipha_t *)mp->b_rptr;
600 }
601
602 DTRACE_IP7(receive, mblk_t *, mp, conn_t *, NULL, void_ip_t *,
603 ipha, __dtrace_ipsr_ill_t *, ill, ipha_t *, ipha, ip6_t *, NULL,
604 int, 0);
605
606 /*
607 * The event for packets being received from a 'physical'
608 * interface is placed after validation of the source and/or
609 * destination address as being local so that packets can be
610 * redirected to loopback addresses using ipnat.
611 */
612 DTRACE_PROBE4(ip4__physical__in__start,
613 ill_t *, ill, ill_t *, NULL,
614 ipha_t *, ipha, mblk_t *, mp);
615
616 if (HOOKS4_INTERESTED_PHYSICAL_IN(ipst)) {
617 int ll_multicast = 0;
618 int error;
619 ipaddr_t orig_dst = ipha->ipha_dst;
620
621 if (ira->ira_flags & IRAF_L2DST_MULTICAST)
622 ll_multicast = HPE_MULTICAST;
623 else if (ira->ira_flags & IRAF_L2DST_BROADCAST)
624 ll_multicast = HPE_BROADCAST;
625
626 FW_HOOKS(ipst->ips_ip4_physical_in_event,
627 ipst->ips_ipv4firewall_physical_in,
628 ill, NULL, ipha, mp, mp, ll_multicast, ipst, error);
629
630 DTRACE_PROBE1(ip4__physical__in__end, mblk_t *, mp);
631
632 if (mp == NULL)
633 return;
634 /* The length could have changed */
635 ipha = (ipha_t *)mp->b_rptr;
636 ira->ira_pktlen = ntohs(ipha->ipha_length);
637 pkt_len = ira->ira_pktlen;
638
639 /*
640 * In case the destination changed we override any previous
641 * change to nexthop.
642 */
643 if (orig_dst != ipha->ipha_dst)
644 nexthop = ipha->ipha_dst;
645 if (nexthop == INADDR_ANY) {
646 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInAddrErrors);
647 ip_drop_input("ipIfStatsInAddrErrors", mp, ill);
648 freemsg(mp);
649 return;
650 }
651 }
652
653 if (ipst->ips_ip4_observe.he_interested) {
654 zoneid_t dzone;
655
656 /*
657 * On the inbound path the src zone will be unknown as
658 * this packet has come from the wire.
659 */
660 dzone = ip_get_zoneid_v4(nexthop, mp, ira, ALL_ZONES);
661 ipobs_hook(mp, IPOBS_HOOK_INBOUND, ALL_ZONES, dzone, ill, ipst);
662 }
663
664 /*
665 * If the packet originated from a same-machine sender or
666 * there is a good HW IP header checksum, we clear the need
667 * look at the IP header checksum.
668 */
669 if (((DB_CKSUMFLAGS(mp) & HCK_IPV4_HDRCKSUM) &&
670 ILL_HCKSUM_CAPABLE(ill) && dohwcksum)) {
671 /* Header checksum was ok. Clear the flag */
672 DB_CKSUMFLAGS(mp) &= ~HCK_IPV4_HDRCKSUM;
673 ira->ira_flags &= ~IRAF_VERIFY_IP_CKSUM;
674 }
675
676 /*
677 * Here we check to see if we machine is setup as
678 * L3 loadbalancer and if the incoming packet is for a VIP
679 *
680 * Check the following:
681 * - there is at least a rule
682 * - protocol of the packet is supported
683 */
684 if (ilb_has_rules(ilbs) && ILB_SUPP_L4(ipha->ipha_protocol)) {
685 ipaddr_t lb_dst;
686 int lb_ret;
687
688 /* For convenience, we pull up the mblk. */
689 if (mp->b_cont != NULL) {
690 if (pullupmsg(mp, -1) == 0) {
691 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
692 ip_drop_input("ipIfStatsInDiscards - pullupmsg",
693 mp, ill);
694 freemsg(mp);
695 return;
696 }
697 ipha = (ipha_t *)mp->b_rptr;
698 }
699
700 /*
701 * We just drop all fragments going to any VIP, at
702 * least for now....
703 */
704 if (ntohs(ipha->ipha_fragment_offset_and_flags) &
705 (IPH_MF | IPH_OFFSET)) {
706 if (!ilb_rule_match_vip_v4(ilbs, nexthop, NULL)) {
707 goto after_ilb;
708 }
709
710 ILB_KSTAT_UPDATE(ilbs, ip_frag_in, 1);
711 ILB_KSTAT_UPDATE(ilbs, ip_frag_dropped, 1);
712 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
713 ip_drop_input("ILB fragment", mp, ill);
714 freemsg(mp);
715 return;
716 }
717 lb_ret = ilb_check_v4(ilbs, ill, mp, ipha, ipha->ipha_protocol,
718 (uint8_t *)ipha + IPH_HDR_LENGTH(ipha), &lb_dst);
719
720 if (lb_ret == ILB_DROPPED) {
721 /* Is this the right counter to increase? */
722 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
723 ip_drop_input("ILB_DROPPED", mp, ill);
724 freemsg(mp);
725 return;
726 }
727 if (lb_ret == ILB_BALANCED) {
728 /* Set the dst to that of the chosen server */
729 nexthop = lb_dst;
730 DB_CKSUMFLAGS(mp) = 0;
731 }
732 }
733
734 after_ilb:
735 opt_len = ipha->ipha_version_and_hdr_length - IP_SIMPLE_HDR_VERSION;
736 ira->ira_ip_hdr_length = IP_SIMPLE_HDR_LENGTH;
737 if (opt_len != 0) {
738 int error = 0;
739
740 ira->ira_ip_hdr_length += (opt_len << 2);
741 ira->ira_flags |= IRAF_IPV4_OPTIONS;
742
743 /* IP Options present! Validate the length. */
744 mp = ip_check_optlen(mp, ipha, opt_len, pkt_len, ira);
745 if (mp == NULL)
746 return;
747
748 /* Might have changed */
749 ipha = (ipha_t *)mp->b_rptr;
750
751 /* Verify IP header checksum before parsing the options */
752 if ((ira->ira_flags & IRAF_VERIFY_IP_CKSUM) &&
753 ip_csum_hdr(ipha)) {
754 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInCksumErrs);
755 ip_drop_input("ipIfStatsInCksumErrs", mp, ill);
756 freemsg(mp);
757 return;
758 }
759 ira->ira_flags &= ~IRAF_VERIFY_IP_CKSUM;
760
761 /*
762 * Go off to ip_input_options which returns the next hop
763 * destination address, which may have been affected
764 * by source routing.
765 */
766 IP_STAT(ipst, ip_opt);
767
768 nexthop = ip_input_options(ipha, nexthop, mp, ira, &error);
769 if (error != 0) {
770 /*
771 * An ICMP error has been sent and the packet has
772 * been dropped.
773 */
774 return;
775 }
776 }
777
778 if (ill->ill_flags & ILLF_ROUTER)
779 irr_flags = IRR_ALLOCATE;
780 else
781 irr_flags = IRR_NONE;
782
783 /* Can not use route cache with TX since the labels can differ */
784 if (ira->ira_flags & IRAF_SYSTEM_LABELED) {
785 if (CLASSD(nexthop)) {
786 ire = ire_multicast(ill);
787 } else {
788 /* Match destination and label */
789 ire = ire_route_recursive_v4(nexthop, 0, NULL,
790 ALL_ZONES, ira->ira_tsl, MATCH_IRE_SECATTR,
791 irr_flags, ira->ira_xmit_hint, ipst, NULL, NULL,
792 NULL);
793 }
794 /* Update the route cache so we do the ire_refrele */
795 ASSERT(ire != NULL);
796 if (rtc->rtc_ire != NULL)
797 ire_refrele(rtc->rtc_ire);
798 rtc->rtc_ire = ire;
799 rtc->rtc_ipaddr = nexthop;
800 } else if (nexthop == rtc->rtc_ipaddr && rtc->rtc_ire != NULL) {
801 /* Use the route cache */
802 ire = rtc->rtc_ire;
803 } else {
804 /* Update the route cache */
805 if (CLASSD(nexthop)) {
806 ire = ire_multicast(ill);
807 } else {
808 /* Just match the destination */
809 ire = ire_route_recursive_dstonly_v4(nexthop, irr_flags,
810 ira->ira_xmit_hint, ipst);
811 }
812 ASSERT(ire != NULL);
813 if (rtc->rtc_ire != NULL)
814 ire_refrele(rtc->rtc_ire);
815 rtc->rtc_ire = ire;
816 rtc->rtc_ipaddr = nexthop;
817 }
818
819 ire->ire_ib_pkt_count++;
820
821 /*
822 * Based on ire_type and ire_flags call one of:
823 * ire_recv_local_v4 - for IRE_LOCAL
824 * ire_recv_loopback_v4 - for IRE_LOOPBACK
825 * ire_recv_multirt_v4 - if RTF_MULTIRT
826 * ire_recv_noroute_v4 - if RTF_REJECT or RTF_BLACHOLE
827 * ire_recv_multicast_v4 - for IRE_MULTICAST
828 * ire_recv_broadcast_v4 - for IRE_BROADCAST
829 * ire_recv_noaccept_v4 - for ire_noaccept ones
830 * ire_recv_forward_v4 - for the rest.
831 */
832 (*ire->ire_recvfn)(ire, mp, ipha, ira);
833 }
834 #undef rptr
835
836 /*
837 * ire_recvfn for IREs that need forwarding
838 */
839 void
ire_recv_forward_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)840 ire_recv_forward_v4(ire_t *ire, mblk_t *mp, void *iph_arg, ip_recv_attr_t *ira)
841 {
842 ipha_t *ipha = (ipha_t *)iph_arg;
843 ill_t *ill = ira->ira_ill;
844 ip_stack_t *ipst = ill->ill_ipst;
845 ill_t *dst_ill;
846 nce_t *nce;
847 ipaddr_t src = ipha->ipha_src;
848 uint32_t added_tx_len;
849 uint32_t mtu, iremtu;
850
851 if (ira->ira_flags & (IRAF_L2DST_MULTICAST|IRAF_L2DST_BROADCAST)) {
852 BUMP_MIB(ill->ill_ip_mib, ipIfStatsForwProhibits);
853 ip_drop_input("l2 multicast not forwarded", mp, ill);
854 freemsg(mp);
855 return;
856 }
857
858 if (!(ill->ill_flags & ILLF_ROUTER) && !ip_source_routed(ipha, ipst)) {
859 BUMP_MIB(ill->ill_ip_mib, ipIfStatsForwProhibits);
860 ip_drop_input("ipIfStatsForwProhibits", mp, ill);
861 freemsg(mp);
862 return;
863 }
864
865 /*
866 * Either ire_nce_capable or ire_dep_parent would be set for the IRE
867 * when it is found by ire_route_recursive, but that some other thread
868 * could have changed the routes with the effect of clearing
869 * ire_dep_parent. In that case we'd end up dropping the packet, or
870 * finding a new nce below.
871 * Get, allocate, or update the nce.
872 * We get a refhold on ire_nce_cache as a result of this to avoid races
873 * where ire_nce_cache is deleted.
874 *
875 * This ensures that we don't forward if the interface is down since
876 * ipif_down removes all the nces.
877 */
878 mutex_enter(&ire->ire_lock);
879 nce = ire->ire_nce_cache;
880 if (nce == NULL) {
881 /* Not yet set up - try to set one up */
882 mutex_exit(&ire->ire_lock);
883 (void) ire_revalidate_nce(ire);
884 mutex_enter(&ire->ire_lock);
885 nce = ire->ire_nce_cache;
886 if (nce == NULL) {
887 mutex_exit(&ire->ire_lock);
888 /* The ire_dep_parent chain went bad, or no memory */
889 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
890 ip_drop_input("No ire_dep_parent", mp, ill);
891 freemsg(mp);
892 return;
893 }
894 }
895 nce_refhold(nce);
896 mutex_exit(&ire->ire_lock);
897
898 if (nce->nce_is_condemned) {
899 nce_t *nce1;
900
901 nce1 = ire_handle_condemned_nce(nce, ire, ipha, NULL, B_FALSE);
902 nce_refrele(nce);
903 if (nce1 == NULL) {
904 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
905 ip_drop_input("No nce", mp, ill);
906 freemsg(mp);
907 return;
908 }
909 nce = nce1;
910 }
911 dst_ill = nce->nce_ill;
912
913 /*
914 * Unless we are forwarding, drop the packet.
915 * We have to let source routed packets through if they go out
916 * the same interface i.e., they are 'ping -l' packets.
917 */
918 if (!(dst_ill->ill_flags & ILLF_ROUTER) &&
919 !(ip_source_routed(ipha, ipst) && dst_ill == ill)) {
920 if (ip_source_routed(ipha, ipst)) {
921 ip_drop_input("ICMP_SOURCE_ROUTE_FAILED", mp, ill);
922 icmp_unreachable(mp, ICMP_SOURCE_ROUTE_FAILED, ira);
923 nce_refrele(nce);
924 return;
925 }
926 BUMP_MIB(ill->ill_ip_mib, ipIfStatsForwProhibits);
927 ip_drop_input("ipIfStatsForwProhibits", mp, ill);
928 freemsg(mp);
929 nce_refrele(nce);
930 return;
931 }
932
933 if (ire->ire_zoneid != GLOBAL_ZONEID && ire->ire_zoneid != ALL_ZONES) {
934 ipaddr_t dst = ipha->ipha_dst;
935
936 ire->ire_ib_pkt_count--;
937 /*
938 * Should only use IREs that are visible from the
939 * global zone for forwarding.
940 * Take a source route into account the same way as ip_input
941 * did.
942 */
943 if (ira->ira_flags & IRAF_IPV4_OPTIONS) {
944 int error = 0;
945
946 dst = ip_input_options(ipha, dst, mp, ira, &error);
947 ASSERT(error == 0); /* ip_input checked */
948 }
949 ire = ire_route_recursive_v4(dst, 0, NULL, GLOBAL_ZONEID,
950 ira->ira_tsl, MATCH_IRE_SECATTR,
951 (ill->ill_flags & ILLF_ROUTER) ? IRR_ALLOCATE : IRR_NONE,
952 ira->ira_xmit_hint, ipst, NULL, NULL, NULL);
953 ire->ire_ib_pkt_count++;
954 (*ire->ire_recvfn)(ire, mp, ipha, ira);
955 ire_refrele(ire);
956 nce_refrele(nce);
957 return;
958 }
959
960 /*
961 * ipIfStatsHCInForwDatagrams should only be increment if there
962 * will be an attempt to forward the packet, which is why we
963 * increment after the above condition has been checked.
964 */
965 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInForwDatagrams);
966
967 /* Initiate Read side IPPF processing */
968 if (IPP_ENABLED(IPP_FWD_IN, ipst)) {
969 /* ip_process translates an IS_UNDER_IPMP */
970 mp = ip_process(IPP_FWD_IN, mp, ill, ill);
971 if (mp == NULL) {
972 /* ip_drop_packet and MIB done */
973 ip2dbg(("ire_recv_forward_v4: pkt dropped/deferred "
974 "during IPPF processing\n"));
975 nce_refrele(nce);
976 return;
977 }
978 }
979
980 DTRACE_PROBE4(ip4__forwarding__start,
981 ill_t *, ill, ill_t *, dst_ill, ipha_t *, ipha, mblk_t *, mp);
982
983 if (HOOKS4_INTERESTED_FORWARDING(ipst)) {
984 int error;
985
986 FW_HOOKS(ipst->ips_ip4_forwarding_event,
987 ipst->ips_ipv4firewall_forwarding,
988 ill, dst_ill, ipha, mp, mp, 0, ipst, error);
989
990 DTRACE_PROBE1(ip4__forwarding__end, mblk_t *, mp);
991
992 if (mp == NULL) {
993 nce_refrele(nce);
994 return;
995 }
996 /*
997 * Even if the destination was changed by the filter we use the
998 * forwarding decision that was made based on the address
999 * in ip_input.
1000 */
1001
1002 /* Might have changed */
1003 ipha = (ipha_t *)mp->b_rptr;
1004 ira->ira_pktlen = ntohs(ipha->ipha_length);
1005 }
1006
1007 /* Packet is being forwarded. Turning off hwcksum flag. */
1008 DB_CKSUMFLAGS(mp) = 0;
1009
1010 /*
1011 * Martian Address Filtering [RFC 1812, Section 5.3.7]
1012 * The loopback address check for both src and dst has already
1013 * been checked in ip_input
1014 * In the future one can envision adding RPF checks using number 3.
1015 * If we already checked the same source address we can skip this.
1016 */
1017 if (!(ira->ira_flags & IRAF_VERIFIED_SRC) ||
1018 src != ira->ira_verified_src) {
1019 switch (ipst->ips_src_check) {
1020 case 0:
1021 break;
1022 case 2:
1023 if (ip_type_v4(src, ipst) == IRE_BROADCAST) {
1024 BUMP_MIB(ill->ill_ip_mib,
1025 ipIfStatsForwProhibits);
1026 BUMP_MIB(ill->ill_ip_mib,
1027 ipIfStatsInAddrErrors);
1028 ip_drop_input("ipIfStatsInAddrErrors", mp, ill);
1029 freemsg(mp);
1030 nce_refrele(nce);
1031 return;
1032 }
1033 /* FALLTHRU */
1034
1035 case 1:
1036 if (CLASSD(src)) {
1037 BUMP_MIB(ill->ill_ip_mib,
1038 ipIfStatsForwProhibits);
1039 BUMP_MIB(ill->ill_ip_mib,
1040 ipIfStatsInAddrErrors);
1041 ip_drop_input("ipIfStatsInAddrErrors", mp, ill);
1042 freemsg(mp);
1043 nce_refrele(nce);
1044 return;
1045 }
1046 break;
1047 }
1048 /* Remember for next packet */
1049 ira->ira_flags |= IRAF_VERIFIED_SRC;
1050 ira->ira_verified_src = src;
1051 }
1052
1053 /*
1054 * Check if packet is going out the same link on which it arrived.
1055 * Means we might need to send a redirect.
1056 */
1057 if (IS_ON_SAME_LAN(dst_ill, ill) && ipst->ips_ip_g_send_redirects) {
1058 ip_send_potential_redirect_v4(mp, ipha, ire, ira);
1059 }
1060
1061 added_tx_len = 0;
1062 if (ira->ira_flags & IRAF_SYSTEM_LABELED) {
1063 mblk_t *mp1;
1064 uint32_t old_pkt_len = ira->ira_pktlen;
1065
1066 /* Verify IP header checksum before adding/removing options */
1067 if ((ira->ira_flags & IRAF_VERIFY_IP_CKSUM) &&
1068 ip_csum_hdr(ipha)) {
1069 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInCksumErrs);
1070 ip_drop_input("ipIfStatsInCksumErrs", mp, ill);
1071 freemsg(mp);
1072 nce_refrele(nce);
1073 return;
1074 }
1075 ira->ira_flags &= ~IRAF_VERIFY_IP_CKSUM;
1076
1077 /*
1078 * Check if it can be forwarded and add/remove
1079 * CIPSO options as needed.
1080 */
1081 if ((mp1 = tsol_ip_forward(ire, mp, ira)) == NULL) {
1082 BUMP_MIB(ill->ill_ip_mib, ipIfStatsForwProhibits);
1083 ip_drop_input("tsol_ip_forward", mp, ill);
1084 freemsg(mp);
1085 nce_refrele(nce);
1086 return;
1087 }
1088 /*
1089 * Size may have changed. Remember amount added in case
1090 * IP needs to send an ICMP too big.
1091 */
1092 mp = mp1;
1093 ipha = (ipha_t *)mp->b_rptr;
1094 ira->ira_pktlen = ntohs(ipha->ipha_length);
1095 ira->ira_ip_hdr_length = IPH_HDR_LENGTH(ipha);
1096 if (ira->ira_pktlen > old_pkt_len)
1097 added_tx_len = ira->ira_pktlen - old_pkt_len;
1098
1099 /* Options can have been added or removed */
1100 if (ira->ira_ip_hdr_length != IP_SIMPLE_HDR_LENGTH)
1101 ira->ira_flags |= IRAF_IPV4_OPTIONS;
1102 else
1103 ira->ira_flags &= ~IRAF_IPV4_OPTIONS;
1104 }
1105
1106 mtu = dst_ill->ill_mtu;
1107 if ((iremtu = ire->ire_metrics.iulp_mtu) != 0 && iremtu < mtu)
1108 mtu = iremtu;
1109 ip_forward_xmit_v4(nce, ill, mp, ipha, ira, mtu, added_tx_len);
1110 nce_refrele(nce);
1111 }
1112
1113 /*
1114 * Used for sending out unicast and multicast packets that are
1115 * forwarded.
1116 */
1117 void
ip_forward_xmit_v4(nce_t * nce,ill_t * ill,mblk_t * mp,ipha_t * ipha,ip_recv_attr_t * ira,uint32_t mtu,uint32_t added_tx_len)1118 ip_forward_xmit_v4(nce_t *nce, ill_t *ill, mblk_t *mp, ipha_t *ipha,
1119 ip_recv_attr_t *ira, uint32_t mtu, uint32_t added_tx_len)
1120 {
1121 ill_t *dst_ill = nce->nce_ill;
1122 uint32_t pkt_len;
1123 uint32_t sum;
1124 iaflags_t iraflags = ira->ira_flags;
1125 ip_stack_t *ipst = ill->ill_ipst;
1126 iaflags_t ixaflags;
1127
1128 if (ipha->ipha_ttl <= 1) {
1129 /* Perhaps the checksum was bad */
1130 if ((iraflags & IRAF_VERIFY_IP_CKSUM) && ip_csum_hdr(ipha)) {
1131 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInCksumErrs);
1132 ip_drop_input("ipIfStatsInCksumErrs", mp, ill);
1133 freemsg(mp);
1134 return;
1135 }
1136 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
1137 ip_drop_input("ICMP_TTL_EXCEEDED", mp, ill);
1138 icmp_time_exceeded(mp, ICMP_TTL_EXCEEDED, ira);
1139 return;
1140 }
1141
1142 /*
1143 * Count the forward as a hop and update the checksum
1144 * accordingly.
1145 */
1146 ipha->ipha_ttl--;
1147 sum = (int)ipha->ipha_hdr_checksum + IP_HDR_CSUM_TTL_ADJUST;
1148 ipha->ipha_hdr_checksum = (uint16_t)(sum + (sum >> 16));
1149
1150 /* Check if there are options to update */
1151 if (iraflags & IRAF_IPV4_OPTIONS) {
1152 ASSERT(ipha->ipha_version_and_hdr_length !=
1153 IP_SIMPLE_HDR_VERSION);
1154 ASSERT(!(iraflags & IRAF_VERIFY_IP_CKSUM));
1155
1156 if (!ip_forward_options(mp, ipha, dst_ill, ira)) {
1157 /* ipIfStatsForwProhibits and ip_drop_input done */
1158 return;
1159 }
1160
1161 ipha->ipha_hdr_checksum = 0;
1162 ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
1163 }
1164
1165 /* Initiate Write side IPPF processing before any fragmentation */
1166 if (IPP_ENABLED(IPP_FWD_OUT, ipst)) {
1167 /* ip_process translates an IS_UNDER_IPMP */
1168 mp = ip_process(IPP_FWD_OUT, mp, dst_ill, dst_ill);
1169 if (mp == NULL) {
1170 /* ip_drop_packet and MIB done */
1171 ip2dbg(("ire_recv_forward_v4: pkt dropped/deferred" \
1172 " during IPPF processing\n"));
1173 return;
1174 }
1175 }
1176
1177 pkt_len = ira->ira_pktlen;
1178
1179 BUMP_MIB(dst_ill->ill_ip_mib, ipIfStatsHCOutForwDatagrams);
1180
1181 ixaflags = IXAF_IS_IPV4 | IXAF_NO_DEV_FLOW_CTL;
1182
1183 if (pkt_len > mtu) {
1184 /*
1185 * It needs fragging on its way out. If we haven't
1186 * verified the header checksum yet we do it now since
1187 * are going to put a surely good checksum in the
1188 * outgoing header, we have to make sure that it
1189 * was good coming in.
1190 */
1191 if ((iraflags & IRAF_VERIFY_IP_CKSUM) && ip_csum_hdr(ipha)) {
1192 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInCksumErrs);
1193 ip_drop_input("ipIfStatsInCksumErrs", mp, ill);
1194 freemsg(mp);
1195 return;
1196 }
1197 if (ipha->ipha_fragment_offset_and_flags & IPH_DF_HTONS) {
1198 BUMP_MIB(dst_ill->ill_ip_mib, ipIfStatsOutFragFails);
1199 ip_drop_output("ipIfStatsOutFragFails", mp, dst_ill);
1200 if (iraflags & IRAF_SYSTEM_LABELED) {
1201 /*
1202 * Remove any CIPSO option added by
1203 * tsol_ip_forward, and make sure we report
1204 * a path MTU so that there
1205 * is room to add such a CIPSO option for future
1206 * packets.
1207 */
1208 mtu = tsol_pmtu_adjust(mp, mtu, added_tx_len,
1209 AF_INET);
1210 }
1211
1212 icmp_frag_needed(mp, mtu, ira);
1213 return;
1214 }
1215
1216 (void) ip_fragment_v4(mp, nce, ixaflags, pkt_len, mtu,
1217 ira->ira_xmit_hint, GLOBAL_ZONEID, 0, ip_xmit, NULL);
1218 return;
1219 }
1220
1221 ASSERT(pkt_len == ntohs(((ipha_t *)mp->b_rptr)->ipha_length));
1222 if (iraflags & IRAF_LOOPBACK_COPY) {
1223 /*
1224 * IXAF_NO_LOOP_ZONEID is not set hence 7th arg
1225 * is don't care
1226 */
1227 (void) ip_postfrag_loopcheck(mp, nce,
1228 ixaflags | IXAF_LOOPBACK_COPY,
1229 pkt_len, ira->ira_xmit_hint, GLOBAL_ZONEID, 0, NULL);
1230 } else {
1231 (void) ip_xmit(mp, nce, ixaflags, pkt_len, ira->ira_xmit_hint,
1232 GLOBAL_ZONEID, 0, NULL);
1233 }
1234 }
1235
1236 /*
1237 * ire_recvfn for RTF_REJECT and RTF_BLACKHOLE routes, including IRE_NOROUTE,
1238 * which is what ire_route_recursive returns when there is no matching ire.
1239 * Send ICMP unreachable unless blackhole.
1240 */
1241 void
ire_recv_noroute_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)1242 ire_recv_noroute_v4(ire_t *ire, mblk_t *mp, void *iph_arg, ip_recv_attr_t *ira)
1243 {
1244 ipha_t *ipha = (ipha_t *)iph_arg;
1245 ill_t *ill = ira->ira_ill;
1246 ip_stack_t *ipst = ill->ill_ipst;
1247
1248 /* Would we have forwarded this packet if we had a route? */
1249 if (ira->ira_flags & (IRAF_L2DST_MULTICAST|IRAF_L2DST_BROADCAST)) {
1250 BUMP_MIB(ill->ill_ip_mib, ipIfStatsForwProhibits);
1251 ip_drop_input("l2 multicast not forwarded", mp, ill);
1252 freemsg(mp);
1253 return;
1254 }
1255
1256 if (!(ill->ill_flags & ILLF_ROUTER)) {
1257 BUMP_MIB(ill->ill_ip_mib, ipIfStatsForwProhibits);
1258 ip_drop_input("ipIfStatsForwProhibits", mp, ill);
1259 freemsg(mp);
1260 return;
1261 }
1262 /*
1263 * If we had a route this could have been forwarded. Count as such.
1264 *
1265 * ipIfStatsHCInForwDatagrams should only be increment if there
1266 * will be an attempt to forward the packet, which is why we
1267 * increment after the above condition has been checked.
1268 */
1269 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInForwDatagrams);
1270
1271 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInNoRoutes);
1272
1273 ip_rts_change(RTM_MISS, ipha->ipha_dst, 0, 0, 0, 0, 0, 0, RTA_DST,
1274 ipst);
1275
1276 if (ire->ire_flags & RTF_BLACKHOLE) {
1277 ip_drop_input("ipIfStatsInNoRoutes RTF_BLACKHOLE", mp, ill);
1278 freemsg(mp);
1279 } else {
1280 ip_drop_input("ipIfStatsInNoRoutes RTF_REJECT", mp, ill);
1281
1282 if (ip_source_routed(ipha, ipst)) {
1283 icmp_unreachable(mp, ICMP_SOURCE_ROUTE_FAILED, ira);
1284 } else {
1285 icmp_unreachable(mp, ICMP_HOST_UNREACHABLE, ira);
1286 }
1287 }
1288 }
1289
1290 /*
1291 * ire_recvfn for IRE_LOCALs marked with ire_noaccept. Such IREs are used for
1292 * VRRP when in noaccept mode.
1293 * We silently drop the packet. ARP handles packets even if noaccept is set.
1294 */
1295 /* ARGSUSED */
1296 void
ire_recv_noaccept_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)1297 ire_recv_noaccept_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
1298 ip_recv_attr_t *ira)
1299 {
1300 ill_t *ill = ira->ira_ill;
1301
1302 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
1303 ip_drop_input("ipIfStatsInDiscards - noaccept", mp, ill);
1304 freemsg(mp);
1305 }
1306
1307 /*
1308 * ire_recvfn for IRE_BROADCAST.
1309 */
1310 void
ire_recv_broadcast_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)1311 ire_recv_broadcast_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
1312 ip_recv_attr_t *ira)
1313 {
1314 ipha_t *ipha = (ipha_t *)iph_arg;
1315 ill_t *ill = ira->ira_ill;
1316 ill_t *dst_ill = ire->ire_ill;
1317 ip_stack_t *ipst = ill->ill_ipst;
1318 ire_t *alt_ire;
1319 nce_t *nce;
1320 ipaddr_t ipha_dst;
1321
1322 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInBcastPkts);
1323
1324 /* Tag for higher-level protocols */
1325 ira->ira_flags |= IRAF_BROADCAST;
1326
1327 /*
1328 * Whether local or directed broadcast forwarding: don't allow
1329 * for TCP.
1330 */
1331 if (ipha->ipha_protocol == IPPROTO_TCP) {
1332 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
1333 ip_drop_input("ipIfStatsInDiscards", mp, ill);
1334 freemsg(mp);
1335 return;
1336 }
1337
1338 /*
1339 * So that we don't end up with dups, only one ill an IPMP group is
1340 * nominated to receive broadcast traffic.
1341 * If we have no cast_ill we are liberal and accept everything.
1342 */
1343 if (IS_UNDER_IPMP(ill)) {
1344 /* For an under ill_grp can change under lock */
1345 rw_enter(&ipst->ips_ill_g_lock, RW_READER);
1346 if (!ill->ill_nom_cast && ill->ill_grp != NULL &&
1347 ill->ill_grp->ig_cast_ill != NULL) {
1348 rw_exit(&ipst->ips_ill_g_lock);
1349 /* No MIB since this is normal operation */
1350 ip_drop_input("not nom_cast", mp, ill);
1351 freemsg(mp);
1352 return;
1353 }
1354 rw_exit(&ipst->ips_ill_g_lock);
1355
1356 ira->ira_ruifindex = ill_get_upper_ifindex(ill);
1357 }
1358
1359 /*
1360 * After reassembly and IPsec we will need to duplicate the
1361 * broadcast packet for all matching zones on the ill.
1362 */
1363 ira->ira_zoneid = ALL_ZONES;
1364
1365 /*
1366 * Check for directed broadcast i.e. ire->ire_ill is different than
1367 * the incoming ill.
1368 * The same broadcast address can be assigned to multiple interfaces
1369 * so have to check explicitly for that case by looking up the alt_ire
1370 */
1371 if (dst_ill == ill && !(ire->ire_flags & RTF_MULTIRT)) {
1372 /* Reassemble on the ill on which the packet arrived */
1373 ip_input_local_v4(ire, mp, ipha, ira);
1374 /* Restore */
1375 ira->ira_ruifindex = ill->ill_phyint->phyint_ifindex;
1376 return;
1377 }
1378
1379 /* Is there an IRE_BROADCAST on the incoming ill? */
1380 ipha_dst = ((ira->ira_flags & IRAF_DHCP_UNICAST) ? INADDR_BROADCAST :
1381 ipha->ipha_dst);
1382 alt_ire = ire_ftable_lookup_v4(ipha_dst, 0, 0, IRE_BROADCAST, ill,
1383 ALL_ZONES, ira->ira_tsl,
1384 MATCH_IRE_TYPE|MATCH_IRE_ILL|MATCH_IRE_SECATTR, 0, ipst, NULL);
1385 if (alt_ire != NULL) {
1386 /* Not a directed broadcast */
1387 /*
1388 * In the special case of multirouted broadcast
1389 * packets, we unconditionally need to "gateway"
1390 * them to the appropriate interface here so that reassembly
1391 * works. We know that the IRE_BROADCAST on cgtp0 doesn't
1392 * have RTF_MULTIRT set so we look for such an IRE in the
1393 * bucket.
1394 */
1395 if (alt_ire->ire_flags & RTF_MULTIRT) {
1396 irb_t *irb;
1397 ire_t *ire1;
1398
1399 irb = ire->ire_bucket;
1400 irb_refhold(irb);
1401 for (ire1 = irb->irb_ire; ire1 != NULL;
1402 ire1 = ire1->ire_next) {
1403 if (IRE_IS_CONDEMNED(ire1))
1404 continue;
1405 if (!(ire1->ire_type & IRE_BROADCAST) ||
1406 (ire1->ire_flags & RTF_MULTIRT))
1407 continue;
1408 ill = ire1->ire_ill;
1409 ill_refhold(ill);
1410 break;
1411 }
1412 irb_refrele(irb);
1413 if (ire1 != NULL) {
1414 ill_t *orig_ill = ira->ira_ill;
1415
1416 ire_refrele(alt_ire);
1417 /* Reassemble on the new ill */
1418 ira->ira_ill = ill;
1419 ip_input_local_v4(ire, mp, ipha, ira);
1420 ill_refrele(ill);
1421 /* Restore */
1422 ira->ira_ill = orig_ill;
1423 ira->ira_ruifindex =
1424 orig_ill->ill_phyint->phyint_ifindex;
1425 return;
1426 }
1427 }
1428 ire_refrele(alt_ire);
1429 /* Reassemble on the ill on which the packet arrived */
1430 ip_input_local_v4(ire, mp, ipha, ira);
1431 goto done;
1432 }
1433
1434 /*
1435 * This is a directed broadcast
1436 *
1437 * If directed broadcast is allowed, then forward the packet out
1438 * the destination interface with IXAF_LOOPBACK_COPY set. That will
1439 * result in ip_input() receiving a copy of the packet on the
1440 * appropriate ill. (We could optimize this to avoid the extra trip
1441 * via ip_input(), but since directed broadcasts are normally disabled
1442 * it doesn't make sense to optimize it.)
1443 */
1444 if (!ipst->ips_ip_g_forward_directed_bcast ||
1445 (ira->ira_flags & (IRAF_L2DST_MULTICAST|IRAF_L2DST_BROADCAST))) {
1446 ip_drop_input("directed broadcast not allowed", mp, ill);
1447 freemsg(mp);
1448 goto done;
1449 }
1450 if ((ira->ira_flags & IRAF_VERIFY_IP_CKSUM) && ip_csum_hdr(ipha)) {
1451 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInCksumErrs);
1452 ip_drop_input("ipIfStatsInCksumErrs", mp, ill);
1453 freemsg(mp);
1454 goto done;
1455 }
1456
1457 /*
1458 * Clear the indication that this may have hardware
1459 * checksum as we are not using it for forwarding.
1460 */
1461 DB_CKSUMFLAGS(mp) = 0;
1462
1463 /*
1464 * Adjust ttl to 2 (1+1 - the forward engine will decrement it by one.
1465 */
1466 ipha->ipha_ttl = ipst->ips_ip_broadcast_ttl + 1;
1467 ipha->ipha_hdr_checksum = 0;
1468 ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
1469
1470 /*
1471 * We use ip_forward_xmit to do any fragmentation.
1472 * and loopback copy on the outbound interface.
1473 *
1474 * Make it so that IXAF_LOOPBACK_COPY to be set on transmit side.
1475 */
1476 ira->ira_flags |= IRAF_LOOPBACK_COPY;
1477
1478 nce = arp_nce_init(dst_ill, ipha->ipha_dst, IRE_BROADCAST);
1479 if (nce == NULL) {
1480 BUMP_MIB(dst_ill->ill_ip_mib, ipIfStatsOutDiscards);
1481 ip_drop_output("No nce", mp, dst_ill);
1482 freemsg(mp);
1483 goto done;
1484 }
1485
1486 ip_forward_xmit_v4(nce, ill, mp, ipha, ira, dst_ill->ill_mc_mtu, 0);
1487 nce_refrele(nce);
1488 done:
1489 /* Restore */
1490 ira->ira_ruifindex = ill->ill_phyint->phyint_ifindex;
1491 }
1492
1493 /*
1494 * ire_recvfn for IRE_MULTICAST.
1495 */
1496 void
ire_recv_multicast_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)1497 ire_recv_multicast_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
1498 ip_recv_attr_t *ira)
1499 {
1500 ipha_t *ipha = (ipha_t *)iph_arg;
1501 ill_t *ill = ira->ira_ill;
1502 ip_stack_t *ipst = ill->ill_ipst;
1503
1504 ASSERT(ire->ire_ill == ira->ira_ill);
1505
1506 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInMcastPkts);
1507 UPDATE_MIB(ill->ill_ip_mib, ipIfStatsHCInMcastOctets, ira->ira_pktlen);
1508
1509 /* RSVP hook */
1510 if (ira->ira_flags & IRAF_RSVP)
1511 goto forus;
1512
1513 /* Tag for higher-level protocols */
1514 ira->ira_flags |= IRAF_MULTICAST;
1515
1516 /*
1517 * So that we don't end up with dups, only one ill an IPMP group is
1518 * nominated to receive multicast traffic.
1519 * If we have no cast_ill we are liberal and accept everything.
1520 */
1521 if (IS_UNDER_IPMP(ill)) {
1522 ip_stack_t *ipst = ill->ill_ipst;
1523
1524 /* For an under ill_grp can change under lock */
1525 rw_enter(&ipst->ips_ill_g_lock, RW_READER);
1526 if (!ill->ill_nom_cast && ill->ill_grp != NULL &&
1527 ill->ill_grp->ig_cast_ill != NULL) {
1528 rw_exit(&ipst->ips_ill_g_lock);
1529 ip_drop_input("not on cast ill", mp, ill);
1530 freemsg(mp);
1531 return;
1532 }
1533 rw_exit(&ipst->ips_ill_g_lock);
1534 /*
1535 * We switch to the upper ill so that mrouter and hasmembers
1536 * can operate on upper here and in ip_input_multicast.
1537 */
1538 ill = ipmp_ill_hold_ipmp_ill(ill);
1539 if (ill != NULL) {
1540 ASSERT(ill != ira->ira_ill);
1541 ASSERT(ire->ire_ill == ira->ira_ill);
1542 ira->ira_ill = ill;
1543 ira->ira_ruifindex = ill->ill_phyint->phyint_ifindex;
1544 } else {
1545 ill = ira->ira_ill;
1546 }
1547 }
1548
1549 /*
1550 * Check if we are a multicast router - send ip_mforward a copy of
1551 * the packet.
1552 * Due to mroute_decap tunnels we consider forwarding packets even if
1553 * mrouted has not joined the allmulti group on this interface.
1554 */
1555 if (ipst->ips_ip_g_mrouter) {
1556 int retval;
1557
1558 /*
1559 * Clear the indication that this may have hardware
1560 * checksum as we are not using it for forwarding.
1561 */
1562 DB_CKSUMFLAGS(mp) = 0;
1563
1564 /*
1565 * ip_mforward helps us make these distinctions: If received
1566 * on tunnel and not IGMP, then drop.
1567 * If IGMP packet, then don't check membership
1568 * If received on a phyint and IGMP or PIM, then
1569 * don't check membership
1570 */
1571 retval = ip_mforward(mp, ira);
1572 /* ip_mforward updates mib variables if needed */
1573
1574 switch (retval) {
1575 case 0:
1576 /*
1577 * pkt is okay and arrived on phyint.
1578 *
1579 * If we are running as a multicast router
1580 * we need to see all IGMP and/or PIM packets.
1581 */
1582 if ((ipha->ipha_protocol == IPPROTO_IGMP) ||
1583 (ipha->ipha_protocol == IPPROTO_PIM)) {
1584 goto forus;
1585 }
1586 break;
1587 case -1:
1588 /* pkt is mal-formed, toss it */
1589 freemsg(mp);
1590 goto done;
1591 case 1:
1592 /*
1593 * pkt is okay and arrived on a tunnel
1594 *
1595 * If we are running a multicast router
1596 * we need to see all igmp packets.
1597 */
1598 if (ipha->ipha_protocol == IPPROTO_IGMP) {
1599 goto forus;
1600 }
1601 ip_drop_input("Multicast on tunnel ignored", mp, ill);
1602 freemsg(mp);
1603 goto done;
1604 }
1605 }
1606
1607 /*
1608 * Check if we have members on this ill. This is not necessary for
1609 * correctness because even if the NIC/GLD had a leaky filter, we
1610 * filter before passing to each conn_t.
1611 */
1612 if (!ill_hasmembers_v4(ill, ipha->ipha_dst)) {
1613 /*
1614 * Nobody interested
1615 *
1616 * This might just be caused by the fact that
1617 * multiple IP Multicast addresses map to the same
1618 * link layer multicast - no need to increment counter!
1619 */
1620 ip_drop_input("Multicast with no members", mp, ill);
1621 freemsg(mp);
1622 goto done;
1623 }
1624 forus:
1625 ip2dbg(("ire_recv_multicast_v4: multicast for us: 0x%x\n",
1626 ntohl(ipha->ipha_dst)));
1627
1628 /*
1629 * After reassembly and IPsec we will need to duplicate the
1630 * multicast packet for all matching zones on the ill.
1631 */
1632 ira->ira_zoneid = ALL_ZONES;
1633
1634 /* Reassemble on the ill on which the packet arrived */
1635 ip_input_local_v4(ire, mp, ipha, ira);
1636 done:
1637 if (ill != ire->ire_ill) {
1638 ill_refrele(ill);
1639 ira->ira_ill = ire->ire_ill;
1640 ira->ira_ruifindex = ira->ira_ill->ill_phyint->phyint_ifindex;
1641 }
1642 }
1643
1644 /*
1645 * ire_recvfn for IRE_OFFLINK with RTF_MULTIRT.
1646 * Drop packets since we don't forward out multirt routes.
1647 */
1648 /* ARGSUSED */
1649 void
ire_recv_multirt_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)1650 ire_recv_multirt_v4(ire_t *ire, mblk_t *mp, void *iph_arg, ip_recv_attr_t *ira)
1651 {
1652 ill_t *ill = ira->ira_ill;
1653
1654 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInNoRoutes);
1655 ip_drop_input("Not forwarding out MULTIRT", mp, ill);
1656 freemsg(mp);
1657 }
1658
1659 /*
1660 * ire_recvfn for IRE_LOOPBACK. This is only used when a FW_HOOK
1661 * has rewritten the packet to have a loopback destination address (We
1662 * filter out packet with a loopback destination from arriving over the wire).
1663 * We don't know what zone to use, thus we always use the GLOBAL_ZONEID.
1664 */
1665 void
ire_recv_loopback_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)1666 ire_recv_loopback_v4(ire_t *ire, mblk_t *mp, void *iph_arg, ip_recv_attr_t *ira)
1667 {
1668 ipha_t *ipha = (ipha_t *)iph_arg;
1669 ill_t *ill = ira->ira_ill;
1670 ill_t *ire_ill = ire->ire_ill;
1671
1672 ira->ira_zoneid = GLOBAL_ZONEID;
1673
1674 /* Switch to the lo0 ill for further processing */
1675 if (ire_ill != ill) {
1676 /*
1677 * Update ira_ill to be the ILL on which the IP address
1678 * is hosted.
1679 * No need to hold the ill since we have a hold on the ire
1680 */
1681 ASSERT(ira->ira_ill == ira->ira_rill);
1682 ira->ira_ill = ire_ill;
1683
1684 ip_input_local_v4(ire, mp, ipha, ira);
1685
1686 /* Restore */
1687 ASSERT(ira->ira_ill == ire_ill);
1688 ira->ira_ill = ill;
1689 return;
1690
1691 }
1692 ip_input_local_v4(ire, mp, ipha, ira);
1693 }
1694
1695 /*
1696 * ire_recvfn for IRE_LOCAL.
1697 */
1698 void
ire_recv_local_v4(ire_t * ire,mblk_t * mp,void * iph_arg,ip_recv_attr_t * ira)1699 ire_recv_local_v4(ire_t *ire, mblk_t *mp, void *iph_arg, ip_recv_attr_t *ira)
1700 {
1701 ipha_t *ipha = (ipha_t *)iph_arg;
1702 ill_t *ill = ira->ira_ill;
1703 ill_t *ire_ill = ire->ire_ill;
1704
1705 /* Make a note for DAD that this address is in use */
1706 ire->ire_last_used_time = LBOLT_FASTPATH;
1707
1708 /* Only target the IRE_LOCAL with the right zoneid. */
1709 ira->ira_zoneid = ire->ire_zoneid;
1710
1711 /*
1712 * If the packet arrived on the wrong ill, we check that
1713 * this is ok.
1714 * If it is, then we ensure that we do the reassembly on
1715 * the ill on which the address is hosted. We keep ira_rill as
1716 * the one on which the packet arrived, so that IP_PKTINFO and
1717 * friends can report this.
1718 */
1719 if (ire_ill != ill) {
1720 ire_t *new_ire;
1721
1722 new_ire = ip_check_multihome(&ipha->ipha_dst, ire, ill);
1723 if (new_ire == NULL) {
1724 /* Drop packet */
1725 BUMP_MIB(ill->ill_ip_mib, ipIfStatsForwProhibits);
1726 ip_drop_input("ipIfStatsInForwProhibits", mp, ill);
1727 freemsg(mp);
1728 return;
1729 }
1730 /*
1731 * Update ira_ill to be the ILL on which the IP address
1732 * is hosted. No need to hold the ill since we have a
1733 * hold on the ire. Note that we do the switch even if
1734 * new_ire == ire (for IPMP, ire would be the one corresponding
1735 * to the IPMP ill).
1736 */
1737 ASSERT(ira->ira_ill == ira->ira_rill);
1738 ira->ira_ill = new_ire->ire_ill;
1739
1740 /* ira_ruifindex tracks the upper for ira_rill */
1741 if (IS_UNDER_IPMP(ill))
1742 ira->ira_ruifindex = ill_get_upper_ifindex(ill);
1743
1744 ip_input_local_v4(new_ire, mp, ipha, ira);
1745
1746 /* Restore */
1747 ASSERT(ira->ira_ill == new_ire->ire_ill);
1748 ira->ira_ill = ill;
1749 ira->ira_ruifindex = ill->ill_phyint->phyint_ifindex;
1750
1751 if (new_ire != ire)
1752 ire_refrele(new_ire);
1753 return;
1754 }
1755
1756 ip_input_local_v4(ire, mp, ipha, ira);
1757 }
1758
1759 /*
1760 * Common function for packets arriving for the host. Handles
1761 * checksum verification, reassembly checks, etc.
1762 */
1763 static void
ip_input_local_v4(ire_t * ire,mblk_t * mp,ipha_t * ipha,ip_recv_attr_t * ira)1764 ip_input_local_v4(ire_t *ire, mblk_t *mp, ipha_t *ipha, ip_recv_attr_t *ira)
1765 {
1766 ill_t *ill = ira->ira_ill;
1767 iaflags_t iraflags = ira->ira_flags;
1768
1769 /*
1770 * Verify IP header checksum. If the packet was AH or ESP then
1771 * this flag has already been cleared. Likewise if the packet
1772 * had a hardware checksum.
1773 */
1774 if ((iraflags & IRAF_VERIFY_IP_CKSUM) && ip_csum_hdr(ipha)) {
1775 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInCksumErrs);
1776 ip_drop_input("ipIfStatsInCksumErrs", mp, ill);
1777 freemsg(mp);
1778 return;
1779 }
1780
1781 if (iraflags & IRAF_IPV4_OPTIONS) {
1782 if (!ip_input_local_options(mp, ipha, ira)) {
1783 /* Error has been sent and mp consumed */
1784 return;
1785 }
1786 /*
1787 * Some old hardware does partial checksum by including the
1788 * whole IP header, so the partial checksum value might have
1789 * become invalid if any option in the packet have been
1790 * updated. Always clear partial checksum flag here.
1791 */
1792 DB_CKSUMFLAGS(mp) &= ~HCK_PARTIALCKSUM;
1793 }
1794
1795 /*
1796 * Is packet part of fragmented IP packet?
1797 * We compare against defined values in network byte order
1798 */
1799 if (ipha->ipha_fragment_offset_and_flags &
1800 (IPH_MF_HTONS | IPH_OFFSET_HTONS)) {
1801 /*
1802 * Make sure we have ira_l2src before we loose the original
1803 * mblk
1804 */
1805 if (!(ira->ira_flags & IRAF_L2SRC_SET))
1806 ip_setl2src(mp, ira, ira->ira_rill);
1807
1808 mp = ip_input_fragment(mp, ipha, ira);
1809 if (mp == NULL)
1810 return;
1811 /* Completed reassembly */
1812 ipha = (ipha_t *)mp->b_rptr;
1813 }
1814
1815 /*
1816 * For broadcast and multicast we need some extra work before
1817 * we call ip_fanout_v4(), since in the case of shared-IP zones
1818 * we need to pretend that a packet arrived for each zoneid.
1819 */
1820 if (iraflags & IRAF_MULTIBROADCAST) {
1821 if (iraflags & IRAF_BROADCAST)
1822 ip_input_broadcast_v4(ire, mp, ipha, ira);
1823 else
1824 ip_input_multicast_v4(ire, mp, ipha, ira);
1825 return;
1826 }
1827 ip_fanout_v4(mp, ipha, ira);
1828 }
1829
1830
1831 /*
1832 * Handle multiple zones which match the same broadcast address
1833 * and ill by delivering a packet to each of them.
1834 * Walk the bucket and look for different ire_zoneid but otherwise
1835 * the same IRE (same ill/addr/mask/type).
1836 * Note that ire_add() tracks IREs that are identical in all
1837 * fields (addr/mask/type/gw/ill/zoneid) within a single IRE by
1838 * increasing ire_identical_cnt. Thus we don't need to be concerned
1839 * about those.
1840 */
1841 static void
ip_input_broadcast_v4(ire_t * ire,mblk_t * mp,ipha_t * ipha,ip_recv_attr_t * ira)1842 ip_input_broadcast_v4(ire_t *ire, mblk_t *mp, ipha_t *ipha, ip_recv_attr_t *ira)
1843 {
1844 ill_t *ill = ira->ira_ill;
1845 ip_stack_t *ipst = ill->ill_ipst;
1846 netstack_t *ns = ipst->ips_netstack;
1847 irb_t *irb;
1848 ire_t *ire1;
1849 mblk_t *mp1;
1850 ipha_t *ipha1;
1851 uint_t ira_pktlen = ira->ira_pktlen;
1852 uint16_t ira_ip_hdr_length = ira->ira_ip_hdr_length;
1853
1854 irb = ire->ire_bucket;
1855
1856 /*
1857 * If we don't have more than one shared-IP zone, or if
1858 * there can't be more than one IRE_BROADCAST for this
1859 * IP address, then just set the zoneid and proceed.
1860 */
1861 if (ns->netstack_numzones == 1 || irb->irb_ire_cnt == 1) {
1862 ira->ira_zoneid = ire->ire_zoneid;
1863
1864 ip_fanout_v4(mp, ipha, ira);
1865 return;
1866 }
1867 irb_refhold(irb);
1868 for (ire1 = irb->irb_ire; ire1 != NULL; ire1 = ire1->ire_next) {
1869 /* We do the main IRE after the end of the loop */
1870 if (ire1 == ire)
1871 continue;
1872
1873 /*
1874 * Only IREs for the same IP address should be in the same
1875 * bucket.
1876 * But could have IRE_HOSTs in the case of CGTP.
1877 */
1878 ASSERT(ire1->ire_addr == ire->ire_addr);
1879 if (!(ire1->ire_type & IRE_BROADCAST))
1880 continue;
1881
1882 if (IRE_IS_CONDEMNED(ire1))
1883 continue;
1884
1885 mp1 = copymsg(mp);
1886 if (mp1 == NULL) {
1887 /* Failed to deliver to one zone */
1888 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
1889 ip_drop_input("ipIfStatsInDiscards", mp, ill);
1890 continue;
1891 }
1892 ira->ira_zoneid = ire1->ire_zoneid;
1893 ipha1 = (ipha_t *)mp1->b_rptr;
1894 ip_fanout_v4(mp1, ipha1, ira);
1895 /*
1896 * IPsec might have modified ira_pktlen and ira_ip_hdr_length
1897 * so we restore them for a potential next iteration
1898 */
1899 ira->ira_pktlen = ira_pktlen;
1900 ira->ira_ip_hdr_length = ira_ip_hdr_length;
1901 }
1902 irb_refrele(irb);
1903 /* Do the main ire */
1904 ira->ira_zoneid = ire->ire_zoneid;
1905 ip_fanout_v4(mp, ipha, ira);
1906 }
1907
1908 /*
1909 * Handle multiple zones which want to receive the same multicast packets
1910 * on this ill by delivering a packet to each of them.
1911 *
1912 * Note that for packets delivered to transports we could instead do this
1913 * as part of the fanout code, but since we need to handle icmp_inbound
1914 * it is simpler to have multicast work the same as broadcast.
1915 *
1916 * The ip_fanout matching for multicast matches based on ilm independent of
1917 * zoneid since the zoneid restriction is applied when joining a multicast
1918 * group.
1919 */
1920 /* ARGSUSED */
1921 static void
ip_input_multicast_v4(ire_t * ire,mblk_t * mp,ipha_t * ipha,ip_recv_attr_t * ira)1922 ip_input_multicast_v4(ire_t *ire, mblk_t *mp, ipha_t *ipha, ip_recv_attr_t *ira)
1923 {
1924 ill_t *ill = ira->ira_ill;
1925 iaflags_t iraflags = ira->ira_flags;
1926 ip_stack_t *ipst = ill->ill_ipst;
1927 netstack_t *ns = ipst->ips_netstack;
1928 zoneid_t zoneid;
1929 mblk_t *mp1;
1930 ipha_t *ipha1;
1931 uint_t ira_pktlen = ira->ira_pktlen;
1932 uint16_t ira_ip_hdr_length = ira->ira_ip_hdr_length;
1933
1934 /* ire_recv_multicast has switched to the upper ill for IPMP */
1935 ASSERT(!IS_UNDER_IPMP(ill));
1936
1937 /*
1938 * If we don't have more than one shared-IP zone, or if
1939 * there are no members in anything but the global zone,
1940 * then just set the zoneid and proceed.
1941 */
1942 if (ns->netstack_numzones == 1 ||
1943 !ill_hasmembers_otherzones_v4(ill, ipha->ipha_dst,
1944 GLOBAL_ZONEID)) {
1945 ira->ira_zoneid = GLOBAL_ZONEID;
1946
1947 /* If sender didn't want this zone to receive it, drop */
1948 if ((iraflags & IRAF_NO_LOOP_ZONEID_SET) &&
1949 ira->ira_no_loop_zoneid == ira->ira_zoneid) {
1950 ip_drop_input("Multicast but wrong zoneid", mp, ill);
1951 freemsg(mp);
1952 return;
1953 }
1954 ip_fanout_v4(mp, ipha, ira);
1955 return;
1956 }
1957
1958 /*
1959 * Here we loop over all zoneids that have members in the group
1960 * and deliver a packet to ip_fanout for each zoneid.
1961 *
1962 * First find any members in the lowest numeric zoneid by looking for
1963 * first zoneid larger than -1 (ALL_ZONES).
1964 * We terminate the loop when we receive -1 (ALL_ZONES).
1965 */
1966 zoneid = ill_hasmembers_nextzone_v4(ill, ipha->ipha_dst, ALL_ZONES);
1967 for (; zoneid != ALL_ZONES;
1968 zoneid = ill_hasmembers_nextzone_v4(ill, ipha->ipha_dst, zoneid)) {
1969 /*
1970 * Avoid an extra copymsg/freemsg by skipping global zone here
1971 * and doing that at the end.
1972 */
1973 if (zoneid == GLOBAL_ZONEID)
1974 continue;
1975
1976 ira->ira_zoneid = zoneid;
1977
1978 /* If sender didn't want this zone to receive it, skip */
1979 if ((iraflags & IRAF_NO_LOOP_ZONEID_SET) &&
1980 ira->ira_no_loop_zoneid == ira->ira_zoneid)
1981 continue;
1982
1983 mp1 = copymsg(mp);
1984 if (mp1 == NULL) {
1985 /* Failed to deliver to one zone */
1986 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
1987 ip_drop_input("ipIfStatsInDiscards", mp, ill);
1988 continue;
1989 }
1990 ipha1 = (ipha_t *)mp1->b_rptr;
1991 ip_fanout_v4(mp1, ipha1, ira);
1992 /*
1993 * IPsec might have modified ira_pktlen and ira_ip_hdr_length
1994 * so we restore them for a potential next iteration
1995 */
1996 ira->ira_pktlen = ira_pktlen;
1997 ira->ira_ip_hdr_length = ira_ip_hdr_length;
1998 }
1999
2000 /* Do the main ire */
2001 ira->ira_zoneid = GLOBAL_ZONEID;
2002 /* If sender didn't want this zone to receive it, drop */
2003 if ((iraflags & IRAF_NO_LOOP_ZONEID_SET) &&
2004 ira->ira_no_loop_zoneid == ira->ira_zoneid) {
2005 ip_drop_input("Multicast but wrong zoneid", mp, ill);
2006 freemsg(mp);
2007 } else {
2008 ip_fanout_v4(mp, ipha, ira);
2009 }
2010 }
2011
2012
2013 /*
2014 * Determine the zoneid and IRAF_TX_* flags if trusted extensions
2015 * is in use. Updates ira_zoneid and ira_flags as a result.
2016 */
2017 static void
ip_fanout_tx_v4(mblk_t * mp,ipha_t * ipha,uint8_t protocol,uint_t ip_hdr_length,ip_recv_attr_t * ira)2018 ip_fanout_tx_v4(mblk_t *mp, ipha_t *ipha, uint8_t protocol,
2019 uint_t ip_hdr_length, ip_recv_attr_t *ira)
2020 {
2021 uint16_t *up;
2022 uint16_t lport;
2023 zoneid_t zoneid;
2024
2025 ASSERT(ira->ira_flags & IRAF_SYSTEM_LABELED);
2026
2027 /*
2028 * If the packet is unlabeled we might allow read-down
2029 * for MAC_EXEMPT. Below we clear this if it is a multi-level
2030 * port (MLP).
2031 * Note that ira_tsl can be NULL here.
2032 */
2033 if (ira->ira_tsl != NULL && ira->ira_tsl->tsl_flags & TSLF_UNLABELED)
2034 ira->ira_flags |= IRAF_TX_MAC_EXEMPTABLE;
2035
2036 if (ira->ira_zoneid != ALL_ZONES)
2037 return;
2038
2039 ira->ira_flags |= IRAF_TX_SHARED_ADDR;
2040
2041 up = (uint16_t *)((uchar_t *)ipha + ip_hdr_length);
2042 switch (protocol) {
2043 case IPPROTO_TCP:
2044 case IPPROTO_SCTP:
2045 case IPPROTO_UDP:
2046 /* Caller ensures this */
2047 ASSERT(((uchar_t *)ipha) + ip_hdr_length +4 <= mp->b_wptr);
2048
2049 /*
2050 * Only these transports support MLP.
2051 * We know their destination port numbers is in
2052 * the same place in the header.
2053 */
2054 lport = up[1];
2055
2056 /*
2057 * No need to handle exclusive-stack zones
2058 * since ALL_ZONES only applies to the shared IP instance.
2059 */
2060 zoneid = tsol_mlp_findzone(protocol, lport);
2061 /*
2062 * If no shared MLP is found, tsol_mlp_findzone returns
2063 * ALL_ZONES. In that case, we assume it's SLP, and
2064 * search for the zone based on the packet label.
2065 *
2066 * If there is such a zone, we prefer to find a
2067 * connection in it. Otherwise, we look for a
2068 * MAC-exempt connection in any zone whose label
2069 * dominates the default label on the packet.
2070 */
2071 if (zoneid == ALL_ZONES)
2072 zoneid = tsol_attr_to_zoneid(ira);
2073 else
2074 ira->ira_flags &= ~IRAF_TX_MAC_EXEMPTABLE;
2075 break;
2076 default:
2077 /* Handle shared address for other protocols */
2078 zoneid = tsol_attr_to_zoneid(ira);
2079 break;
2080 }
2081 ira->ira_zoneid = zoneid;
2082 }
2083
2084 /*
2085 * Increment checksum failure statistics
2086 */
2087 static void
ip_input_cksum_err_v4(uint8_t protocol,uint16_t hck_flags,ill_t * ill)2088 ip_input_cksum_err_v4(uint8_t protocol, uint16_t hck_flags, ill_t *ill)
2089 {
2090 ip_stack_t *ipst = ill->ill_ipst;
2091
2092 switch (protocol) {
2093 case IPPROTO_TCP:
2094 BUMP_MIB(ill->ill_ip_mib, tcpIfStatsInErrs);
2095
2096 if (hck_flags & HCK_FULLCKSUM)
2097 IP_STAT(ipst, ip_tcp_in_full_hw_cksum_err);
2098 else if (hck_flags & HCK_PARTIALCKSUM)
2099 IP_STAT(ipst, ip_tcp_in_part_hw_cksum_err);
2100 else
2101 IP_STAT(ipst, ip_tcp_in_sw_cksum_err);
2102 break;
2103 case IPPROTO_UDP:
2104 BUMP_MIB(ill->ill_ip_mib, udpIfStatsInCksumErrs);
2105 if (hck_flags & HCK_FULLCKSUM)
2106 IP_STAT(ipst, ip_udp_in_full_hw_cksum_err);
2107 else if (hck_flags & HCK_PARTIALCKSUM)
2108 IP_STAT(ipst, ip_udp_in_part_hw_cksum_err);
2109 else
2110 IP_STAT(ipst, ip_udp_in_sw_cksum_err);
2111 break;
2112 case IPPROTO_ICMP:
2113 BUMP_MIB(&ipst->ips_icmp_mib, icmpInCksumErrs);
2114 break;
2115 default:
2116 ASSERT(0);
2117 break;
2118 }
2119 }
2120
2121 /* Calculate the IPv4 pseudo-header checksum */
2122 uint32_t
ip_input_cksum_pseudo_v4(ipha_t * ipha,ip_recv_attr_t * ira)2123 ip_input_cksum_pseudo_v4(ipha_t *ipha, ip_recv_attr_t *ira)
2124 {
2125 uint_t ulp_len;
2126 uint32_t cksum;
2127 uint8_t protocol = ira->ira_protocol;
2128 uint16_t ip_hdr_length = ira->ira_ip_hdr_length;
2129
2130 #define iphs ((uint16_t *)ipha)
2131
2132 switch (protocol) {
2133 case IPPROTO_TCP:
2134 ulp_len = ira->ira_pktlen - ip_hdr_length;
2135
2136 /* Protocol and length */
2137 cksum = htons(ulp_len) + IP_TCP_CSUM_COMP;
2138 /* IP addresses */
2139 cksum += iphs[6] + iphs[7] + iphs[8] + iphs[9];
2140 break;
2141
2142 case IPPROTO_UDP: {
2143 udpha_t *udpha;
2144
2145 udpha = (udpha_t *)((uchar_t *)ipha + ip_hdr_length);
2146
2147 /* Protocol and length */
2148 cksum = udpha->uha_length + IP_UDP_CSUM_COMP;
2149 /* IP addresses */
2150 cksum += iphs[6] + iphs[7] + iphs[8] + iphs[9];
2151 break;
2152 }
2153
2154 default:
2155 cksum = 0;
2156 break;
2157 }
2158 #undef iphs
2159 return (cksum);
2160 }
2161
2162
2163 /*
2164 * Software verification of the ULP checksums.
2165 * Returns B_TRUE if ok.
2166 * Increments statistics of failed.
2167 */
2168 static boolean_t
ip_input_sw_cksum_v4(mblk_t * mp,ipha_t * ipha,ip_recv_attr_t * ira)2169 ip_input_sw_cksum_v4(mblk_t *mp, ipha_t *ipha, ip_recv_attr_t *ira)
2170 {
2171 ip_stack_t *ipst = ira->ira_ill->ill_ipst;
2172 uint32_t cksum;
2173 uint8_t protocol = ira->ira_protocol;
2174 uint16_t ip_hdr_length = ira->ira_ip_hdr_length;
2175
2176 IP_STAT(ipst, ip_in_sw_cksum);
2177
2178 ASSERT(protocol == IPPROTO_TCP || protocol == IPPROTO_UDP);
2179
2180 cksum = ip_input_cksum_pseudo_v4(ipha, ira);
2181 cksum = IP_CSUM(mp, ip_hdr_length, cksum);
2182 if (cksum == 0)
2183 return (B_TRUE);
2184
2185 ip_input_cksum_err_v4(protocol, 0, ira->ira_ill);
2186 return (B_FALSE);
2187 }
2188
2189 /*
2190 * Verify the ULP checksums.
2191 * Returns B_TRUE if ok, or if the ULP doesn't have a well-defined checksum
2192 * algorithm.
2193 * Increments statistics if failed.
2194 */
2195 static boolean_t
ip_input_cksum_v4(iaflags_t iraflags,mblk_t * mp,ipha_t * ipha,ip_recv_attr_t * ira)2196 ip_input_cksum_v4(iaflags_t iraflags, mblk_t *mp, ipha_t *ipha,
2197 ip_recv_attr_t *ira)
2198 {
2199 ill_t *ill = ira->ira_rill;
2200 uint16_t hck_flags;
2201 uint32_t cksum;
2202 mblk_t *mp1;
2203 int32_t len;
2204 uint8_t protocol = ira->ira_protocol;
2205 uint16_t ip_hdr_length = ira->ira_ip_hdr_length;
2206
2207
2208 switch (protocol) {
2209 case IPPROTO_TCP:
2210 break;
2211
2212 case IPPROTO_UDP: {
2213 udpha_t *udpha;
2214
2215 udpha = (udpha_t *)((uchar_t *)ipha + ip_hdr_length);
2216 if (udpha->uha_checksum == 0) {
2217 /* Packet doesn't have a UDP checksum */
2218 return (B_TRUE);
2219 }
2220 break;
2221 }
2222 case IPPROTO_SCTP: {
2223 sctp_hdr_t *sctph;
2224 uint32_t pktsum;
2225
2226 sctph = (sctp_hdr_t *)((uchar_t *)ipha + ip_hdr_length);
2227 #ifdef DEBUG
2228 if (skip_sctp_cksum)
2229 return (B_TRUE);
2230 #endif
2231 pktsum = sctph->sh_chksum;
2232 sctph->sh_chksum = 0;
2233 cksum = sctp_cksum(mp, ip_hdr_length);
2234 sctph->sh_chksum = pktsum;
2235 if (cksum == pktsum)
2236 return (B_TRUE);
2237
2238 /*
2239 * Defer until later whether a bad checksum is ok
2240 * in order to allow RAW sockets to use Adler checksum
2241 * with SCTP.
2242 */
2243 ira->ira_flags |= IRAF_SCTP_CSUM_ERR;
2244 return (B_TRUE);
2245 }
2246
2247 default:
2248 /* No ULP checksum to verify. */
2249 return (B_TRUE);
2250 }
2251
2252 /*
2253 * Revert to software checksum calculation if the interface
2254 * isn't capable of checksum offload.
2255 * We clear DB_CKSUMFLAGS when going through IPsec in ip_fanout.
2256 * Note: IRAF_NO_HW_CKSUM is not currently used.
2257 */
2258 ASSERT(!IS_IPMP(ill));
2259 if ((iraflags & IRAF_NO_HW_CKSUM) || !ILL_HCKSUM_CAPABLE(ill) ||
2260 !dohwcksum) {
2261 return (ip_input_sw_cksum_v4(mp, ipha, ira));
2262 }
2263
2264 hck_flags = DB_CKSUMFLAGS(mp);
2265
2266 /*
2267 * We apply this for all ULP protocols. Does the HW know to
2268 * not set the flags for SCTP and other protocols.
2269 */
2270 if (hck_flags & HCK_FULLCKSUM_OK) {
2271 /*
2272 * Hardware has already verified the checksum.
2273 */
2274 return (B_TRUE);
2275 }
2276
2277 if (hck_flags & HCK_FULLCKSUM) {
2278 /*
2279 * Full checksum has been computed by the hardware
2280 * and has been attached. If the driver wants us to
2281 * verify the correctness of the attached value, in
2282 * order to protect against faulty hardware, compare
2283 * it against -0 (0xFFFF) to see if it's valid.
2284 */
2285 cksum = DB_CKSUM16(mp);
2286 if (cksum == 0xFFFF)
2287 return (B_TRUE);
2288 ip_input_cksum_err_v4(protocol, hck_flags, ira->ira_ill);
2289 return (B_FALSE);
2290 }
2291
2292 mp1 = mp->b_cont;
2293 if ((hck_flags & HCK_PARTIALCKSUM) &&
2294 (mp1 == NULL || mp1->b_cont == NULL) &&
2295 ip_hdr_length >= DB_CKSUMSTART(mp) &&
2296 ((len = ip_hdr_length - DB_CKSUMSTART(mp)) & 1) == 0) {
2297 uint32_t adj;
2298 uchar_t *cksum_start;
2299
2300 cksum = ip_input_cksum_pseudo_v4(ipha, ira);
2301
2302 cksum_start = ((uchar_t *)ipha + DB_CKSUMSTART(mp));
2303
2304 /*
2305 * Partial checksum has been calculated by hardware
2306 * and attached to the packet; in addition, any
2307 * prepended extraneous data is even byte aligned,
2308 * and there are at most two mblks associated with
2309 * the packet. If any such data exists, we adjust
2310 * the checksum; also take care any postpended data.
2311 */
2312 IP_ADJCKSUM_PARTIAL(cksum_start, mp, mp1, len, adj);
2313 /*
2314 * One's complement subtract extraneous checksum
2315 */
2316 cksum += DB_CKSUM16(mp);
2317 if (adj >= cksum)
2318 cksum = ~(adj - cksum) & 0xFFFF;
2319 else
2320 cksum -= adj;
2321 cksum = (cksum & 0xFFFF) + ((int)cksum >> 16);
2322 cksum = (cksum & 0xFFFF) + ((int)cksum >> 16);
2323 if (!(~cksum & 0xFFFF))
2324 return (B_TRUE);
2325
2326 ip_input_cksum_err_v4(protocol, hck_flags, ira->ira_ill);
2327 return (B_FALSE);
2328 }
2329 return (ip_input_sw_cksum_v4(mp, ipha, ira));
2330 }
2331
2332
2333 /*
2334 * Handle fanout of received packets.
2335 * Unicast packets that are looped back (from ire_send_local_v4) and packets
2336 * from the wire are differentiated by checking IRAF_VERIFY_ULP_CKSUM.
2337 *
2338 * IPQoS Notes
2339 * Before sending it to the client, invoke IPPF processing. Policy processing
2340 * takes place only if the callout_position, IPP_LOCAL_IN, is enabled.
2341 */
2342 void
ip_fanout_v4(mblk_t * mp,ipha_t * ipha,ip_recv_attr_t * ira)2343 ip_fanout_v4(mblk_t *mp, ipha_t *ipha, ip_recv_attr_t *ira)
2344 {
2345 ill_t *ill = ira->ira_ill;
2346 iaflags_t iraflags = ira->ira_flags;
2347 ip_stack_t *ipst = ill->ill_ipst;
2348 uint8_t protocol = ipha->ipha_protocol;
2349 conn_t *connp;
2350 #define rptr ((uchar_t *)ipha)
2351 uint_t ip_hdr_length;
2352 uint_t min_ulp_header_length;
2353 int offset;
2354 ssize_t len;
2355 netstack_t *ns = ipst->ips_netstack;
2356 ipsec_stack_t *ipss = ns->netstack_ipsec;
2357 ill_t *rill = ira->ira_rill;
2358
2359 ASSERT(ira->ira_pktlen == ntohs(ipha->ipha_length));
2360
2361 ip_hdr_length = ira->ira_ip_hdr_length;
2362 ira->ira_protocol = protocol;
2363
2364 /*
2365 * Time for IPP once we've done reassembly and IPsec.
2366 * We skip this for loopback packets since we don't do IPQoS
2367 * on loopback.
2368 */
2369 if (IPP_ENABLED(IPP_LOCAL_IN, ipst) &&
2370 !(iraflags & IRAF_LOOPBACK) &&
2371 (protocol != IPPROTO_ESP && protocol != IPPROTO_AH)) {
2372 /*
2373 * Use the interface on which the packet arrived - not where
2374 * the IP address is hosted.
2375 */
2376 /* ip_process translates an IS_UNDER_IPMP */
2377 mp = ip_process(IPP_LOCAL_IN, mp, rill, ill);
2378 if (mp == NULL) {
2379 /* ip_drop_packet and MIB done */
2380 return;
2381 }
2382 }
2383
2384 /* Determine the minimum required size of the upper-layer header */
2385 /* Need to do this for at least the set of ULPs that TX handles. */
2386 switch (protocol) {
2387 case IPPROTO_TCP:
2388 min_ulp_header_length = TCP_MIN_HEADER_LENGTH;
2389 break;
2390 case IPPROTO_SCTP:
2391 min_ulp_header_length = SCTP_COMMON_HDR_LENGTH;
2392 break;
2393 case IPPROTO_UDP:
2394 min_ulp_header_length = UDPH_SIZE;
2395 break;
2396 case IPPROTO_ICMP:
2397 min_ulp_header_length = ICMPH_SIZE;
2398 break;
2399 default:
2400 min_ulp_header_length = 0;
2401 break;
2402 }
2403 /* Make sure we have the min ULP header length */
2404 len = mp->b_wptr - rptr;
2405 if (len < ip_hdr_length + min_ulp_header_length) {
2406 if (ira->ira_pktlen < ip_hdr_length + min_ulp_header_length) {
2407 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInTruncatedPkts);
2408 ip_drop_input("ipIfStatsInTruncatedPkts", mp, ill);
2409 freemsg(mp);
2410 return;
2411 }
2412 IP_STAT(ipst, ip_recv_pullup);
2413 ipha = ip_pullup(mp, ip_hdr_length + min_ulp_header_length,
2414 ira);
2415 if (ipha == NULL)
2416 goto discard;
2417 len = mp->b_wptr - rptr;
2418 }
2419
2420 /*
2421 * If trusted extensions then determine the zoneid and TX specific
2422 * ira_flags.
2423 */
2424 if (iraflags & IRAF_SYSTEM_LABELED) {
2425 /* This can update ira->ira_flags and ira->ira_zoneid */
2426 ip_fanout_tx_v4(mp, ipha, protocol, ip_hdr_length, ira);
2427 iraflags = ira->ira_flags;
2428 }
2429
2430
2431 /* Verify ULP checksum. Handles TCP, UDP, and SCTP */
2432 if (iraflags & IRAF_VERIFY_ULP_CKSUM) {
2433 if (!ip_input_cksum_v4(iraflags, mp, ipha, ira)) {
2434 /* Bad checksum. Stats are already incremented */
2435 ip_drop_input("Bad ULP checksum", mp, ill);
2436 freemsg(mp);
2437 return;
2438 }
2439 /* IRAF_SCTP_CSUM_ERR could have been set */
2440 iraflags = ira->ira_flags;
2441 }
2442 switch (protocol) {
2443 case IPPROTO_TCP:
2444 /* For TCP, discard broadcast and multicast packets. */
2445 if (iraflags & IRAF_MULTIBROADCAST)
2446 goto discard;
2447
2448 /* First mblk contains IP+TCP headers per above check */
2449 ASSERT(len >= ip_hdr_length + TCP_MIN_HEADER_LENGTH);
2450
2451 /* TCP options present? */
2452 offset = ((uchar_t *)ipha)[ip_hdr_length + 12] >> 4;
2453 if (offset != 5) {
2454 if (offset < 5)
2455 goto discard;
2456
2457 /*
2458 * There must be TCP options.
2459 * Make sure we can grab them.
2460 */
2461 offset <<= 2;
2462 offset += ip_hdr_length;
2463 if (len < offset) {
2464 if (ira->ira_pktlen < offset) {
2465 BUMP_MIB(ill->ill_ip_mib,
2466 ipIfStatsInTruncatedPkts);
2467 ip_drop_input(
2468 "ipIfStatsInTruncatedPkts",
2469 mp, ill);
2470 freemsg(mp);
2471 return;
2472 }
2473 IP_STAT(ipst, ip_recv_pullup);
2474 ipha = ip_pullup(mp, offset, ira);
2475 if (ipha == NULL)
2476 goto discard;
2477 len = mp->b_wptr - rptr;
2478 }
2479 }
2480
2481 /*
2482 * Pass up a squeue hint to tcp.
2483 * If ira_sqp is already set (this is loopback) we leave it
2484 * alone.
2485 */
2486 if (ira->ira_sqp == NULL) {
2487 ira->ira_sqp = ip_squeue_get(ira->ira_ring);
2488 }
2489
2490 /* Look for AF_INET or AF_INET6 that matches */
2491 connp = ipcl_classify_v4(mp, IPPROTO_TCP, ip_hdr_length,
2492 ira, ipst);
2493 if (connp == NULL) {
2494 /* Send the TH_RST */
2495 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2496 tcp_xmit_listeners_reset(mp, ira, ipst, NULL);
2497 return;
2498 }
2499 if (connp->conn_min_ttl != 0 &&
2500 connp->conn_min_ttl > ira->ira_ttl) {
2501 CONN_DEC_REF(connp);
2502 goto discard;
2503 }
2504 if (CONN_INBOUND_POLICY_PRESENT(connp, ipss) ||
2505 (iraflags & IRAF_IPSEC_SECURE)) {
2506 mp = ipsec_check_inbound_policy(mp, connp,
2507 ipha, NULL, ira);
2508 if (mp == NULL) {
2509 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2510 /* Note that mp is NULL */
2511 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2512 CONN_DEC_REF(connp);
2513 return;
2514 }
2515 }
2516 /* Found a client; up it goes */
2517 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2518 ira->ira_ill = ira->ira_rill = NULL;
2519 if (!IPCL_IS_TCP(connp)) {
2520 /* Not TCP; must be SOCK_RAW, IPPROTO_TCP */
2521 (connp->conn_recv)(connp, mp, NULL, ira);
2522 CONN_DEC_REF(connp);
2523 ira->ira_ill = ill;
2524 ira->ira_rill = rill;
2525 return;
2526 }
2527
2528 /*
2529 * We do different processing whether called from
2530 * ip_accept_tcp and we match the target, don't match
2531 * the target, and when we are called by ip_input.
2532 */
2533 if (iraflags & IRAF_TARGET_SQP) {
2534 if (ira->ira_target_sqp == connp->conn_sqp) {
2535 mblk_t *attrmp;
2536
2537 attrmp = ip_recv_attr_to_mblk(ira);
2538 if (attrmp == NULL) {
2539 BUMP_MIB(ill->ill_ip_mib,
2540 ipIfStatsInDiscards);
2541 ip_drop_input("ipIfStatsInDiscards",
2542 mp, ill);
2543 freemsg(mp);
2544 CONN_DEC_REF(connp);
2545 } else {
2546 SET_SQUEUE(attrmp, connp->conn_recv,
2547 connp);
2548 attrmp->b_cont = mp;
2549 ASSERT(ira->ira_target_sqp_mp == NULL);
2550 ira->ira_target_sqp_mp = attrmp;
2551 /*
2552 * Conn ref release when drained from
2553 * the squeue.
2554 */
2555 }
2556 } else {
2557 SQUEUE_ENTER_ONE(connp->conn_sqp, mp,
2558 connp->conn_recv, connp, ira, SQ_FILL,
2559 SQTAG_IP_TCP_INPUT);
2560 }
2561 } else {
2562 SQUEUE_ENTER_ONE(connp->conn_sqp, mp, connp->conn_recv,
2563 connp, ira, ip_squeue_flag, SQTAG_IP_TCP_INPUT);
2564 }
2565 ira->ira_ill = ill;
2566 ira->ira_rill = rill;
2567 return;
2568
2569 case IPPROTO_SCTP: {
2570 sctp_hdr_t *sctph;
2571 in6_addr_t map_src, map_dst;
2572 uint32_t ports; /* Source and destination ports */
2573 sctp_stack_t *sctps = ipst->ips_netstack->netstack_sctp;
2574
2575 /* For SCTP, discard broadcast and multicast packets. */
2576 if (iraflags & IRAF_MULTIBROADCAST)
2577 goto discard;
2578
2579 /*
2580 * Since there is no SCTP h/w cksum support yet, just
2581 * clear the flag.
2582 */
2583 DB_CKSUMFLAGS(mp) = 0;
2584
2585 /* Length ensured above */
2586 ASSERT(MBLKL(mp) >= ip_hdr_length + SCTP_COMMON_HDR_LENGTH);
2587 sctph = (sctp_hdr_t *)(rptr + ip_hdr_length);
2588
2589 /* get the ports */
2590 ports = *(uint32_t *)&sctph->sh_sport;
2591
2592 IN6_IPADDR_TO_V4MAPPED(ipha->ipha_dst, &map_dst);
2593 IN6_IPADDR_TO_V4MAPPED(ipha->ipha_src, &map_src);
2594 if (iraflags & IRAF_SCTP_CSUM_ERR) {
2595 /*
2596 * No potential sctp checksum errors go to the Sun
2597 * sctp stack however they might be Adler-32 summed
2598 * packets a userland stack bound to a raw IP socket
2599 * could reasonably use. Note though that Adler-32 is
2600 * a long deprecated algorithm and customer sctp
2601 * networks should eventually migrate to CRC-32 at
2602 * which time this facility should be removed.
2603 */
2604 ip_fanout_sctp_raw(mp, ipha, NULL, ports, ira);
2605 return;
2606 }
2607 connp = sctp_fanout(&map_src, &map_dst, ports, ira, mp,
2608 sctps, sctph);
2609 if (connp == NULL) {
2610 /* Check for raw socket or OOTB handling */
2611 ip_fanout_sctp_raw(mp, ipha, NULL, ports, ira);
2612 return;
2613 }
2614 if (connp->conn_incoming_ifindex != 0 &&
2615 connp->conn_incoming_ifindex != ira->ira_ruifindex) {
2616 CONN_DEC_REF(connp);
2617 /* Check for raw socket or OOTB handling */
2618 ip_fanout_sctp_raw(mp, ipha, NULL, ports, ira);
2619 return;
2620 }
2621
2622 /* Found a client; up it goes */
2623 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2624 sctp_input(connp, ipha, NULL, mp, ira);
2625 /* sctp_input does a rele of the sctp_t */
2626 return;
2627 }
2628
2629 case IPPROTO_UDP:
2630 /* First mblk contains IP+UDP headers as checked above */
2631 ASSERT(MBLKL(mp) >= ip_hdr_length + UDPH_SIZE);
2632
2633 if (iraflags & IRAF_MULTIBROADCAST) {
2634 uint16_t *up; /* Pointer to ports in ULP header */
2635
2636 up = (uint16_t *)((uchar_t *)ipha + ip_hdr_length);
2637 ip_fanout_udp_multi_v4(mp, ipha, up[1], up[0], ira);
2638 return;
2639 }
2640
2641 /* Look for AF_INET or AF_INET6 that matches */
2642 connp = ipcl_classify_v4(mp, IPPROTO_UDP, ip_hdr_length,
2643 ira, ipst);
2644 if (connp == NULL) {
2645 if (ipst->ips_ipcl_proto_fanout_v4[IPPROTO_UDP].
2646 connf_head != NULL) {
2647 ASSERT(ira->ira_protocol == IPPROTO_UDP);
2648 ip_fanout_proto_v4(mp, ipha, ira);
2649 } else {
2650 ip_fanout_send_icmp_v4(mp,
2651 ICMP_DEST_UNREACHABLE,
2652 ICMP_PORT_UNREACHABLE, ira);
2653 }
2654 return;
2655
2656 }
2657 if (connp->conn_min_ttl != 0 &&
2658 connp->conn_min_ttl > ira->ira_ttl) {
2659 CONN_DEC_REF(connp);
2660 goto discard;
2661 }
2662 if (IPCL_IS_NONSTR(connp) ? connp->conn_flow_cntrld :
2663 !canputnext(connp->conn_rq)) {
2664 CONN_DEC_REF(connp);
2665 BUMP_MIB(ill->ill_ip_mib, udpIfStatsInOverflows);
2666 ip_drop_input("udpIfStatsInOverflows", mp, ill);
2667 freemsg(mp);
2668 return;
2669 }
2670 if (CONN_INBOUND_POLICY_PRESENT(connp, ipss) ||
2671 (iraflags & IRAF_IPSEC_SECURE)) {
2672 mp = ipsec_check_inbound_policy(mp, connp,
2673 ipha, NULL, ira);
2674 if (mp == NULL) {
2675 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2676 /* Note that mp is NULL */
2677 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2678 CONN_DEC_REF(connp);
2679 return;
2680 }
2681 }
2682 /*
2683 * Remove 0-spi if it's 0, or move everything behind
2684 * the UDP header over it and forward to ESP via
2685 * ip_fanout_v4().
2686 */
2687 if (connp->conn_udp->udp_nat_t_endpoint) {
2688 if (iraflags & IRAF_IPSEC_SECURE) {
2689 ip_drop_packet(mp, B_TRUE, ira->ira_ill,
2690 DROPPER(ipss, ipds_esp_nat_t_ipsec),
2691 &ipss->ipsec_dropper);
2692 CONN_DEC_REF(connp);
2693 return;
2694 }
2695
2696 mp = zero_spi_check(mp, ira);
2697 if (mp == NULL) {
2698 /*
2699 * Packet was consumed - probably sent to
2700 * ip_fanout_v4.
2701 */
2702 CONN_DEC_REF(connp);
2703 return;
2704 }
2705 /* Else continue like a normal UDP packet. */
2706 ipha = (ipha_t *)mp->b_rptr;
2707 protocol = ipha->ipha_protocol;
2708 ira->ira_protocol = protocol;
2709 }
2710 /* Found a client; up it goes */
2711 IP_STAT(ipst, ip_udp_fannorm);
2712 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2713 ira->ira_ill = ira->ira_rill = NULL;
2714 (connp->conn_recv)(connp, mp, NULL, ira);
2715 CONN_DEC_REF(connp);
2716 ira->ira_ill = ill;
2717 ira->ira_rill = rill;
2718 return;
2719 default:
2720 break;
2721 }
2722
2723 /*
2724 * Clear hardware checksumming flag as it is currently only
2725 * used by TCP and UDP.
2726 */
2727 DB_CKSUMFLAGS(mp) = 0;
2728
2729 switch (protocol) {
2730 case IPPROTO_ICMP:
2731 /*
2732 * We need to accomodate icmp messages coming in clear
2733 * until we get everything secure from the wire. If
2734 * icmp_accept_clear_messages is zero we check with
2735 * the global policy and act accordingly. If it is
2736 * non-zero, we accept the message without any checks.
2737 * But *this does not mean* that this will be delivered
2738 * to RAW socket clients. By accepting we might send
2739 * replies back, change our MTU value etc.,
2740 * but delivery to the ULP/clients depends on their
2741 * policy dispositions.
2742 */
2743 if (ipst->ips_icmp_accept_clear_messages == 0) {
2744 mp = ipsec_check_global_policy(mp, NULL,
2745 ipha, NULL, ira, ns);
2746 if (mp == NULL)
2747 return;
2748 }
2749
2750 /*
2751 * On a labeled system, we have to check whether the zone
2752 * itself is permitted to receive raw traffic.
2753 */
2754 if (ira->ira_flags & IRAF_SYSTEM_LABELED) {
2755 if (!tsol_can_accept_raw(mp, ira, B_FALSE)) {
2756 BUMP_MIB(&ipst->ips_icmp_mib, icmpInErrors);
2757 ip_drop_input("tsol_can_accept_raw", mp, ill);
2758 freemsg(mp);
2759 return;
2760 }
2761 }
2762
2763 /*
2764 * ICMP header checksum, including checksum field,
2765 * should be zero.
2766 */
2767 if (IP_CSUM(mp, ip_hdr_length, 0)) {
2768 BUMP_MIB(&ipst->ips_icmp_mib, icmpInCksumErrs);
2769 ip_drop_input("icmpInCksumErrs", mp, ill);
2770 freemsg(mp);
2771 return;
2772 }
2773 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2774 mp = icmp_inbound_v4(mp, ira);
2775 if (mp == NULL) {
2776 /* No need to pass to RAW sockets */
2777 return;
2778 }
2779 break;
2780
2781 case IPPROTO_IGMP:
2782 /*
2783 * If we are not willing to accept IGMP packets in clear,
2784 * then check with global policy.
2785 */
2786 if (ipst->ips_igmp_accept_clear_messages == 0) {
2787 mp = ipsec_check_global_policy(mp, NULL,
2788 ipha, NULL, ira, ns);
2789 if (mp == NULL)
2790 return;
2791 }
2792 if ((ira->ira_flags & IRAF_SYSTEM_LABELED) &&
2793 !tsol_can_accept_raw(mp, ira, B_TRUE)) {
2794 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2795 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2796 freemsg(mp);
2797 return;
2798 }
2799 /*
2800 * Validate checksum
2801 */
2802 if (IP_CSUM(mp, ip_hdr_length, 0)) {
2803 ++ipst->ips_igmpstat.igps_rcv_badsum;
2804 ip_drop_input("igps_rcv_badsum", mp, ill);
2805 freemsg(mp);
2806 return;
2807 }
2808
2809 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2810 mp = igmp_input(mp, ira);
2811 if (mp == NULL) {
2812 /* Bad packet - discarded by igmp_input */
2813 return;
2814 }
2815 break;
2816 case IPPROTO_PIM:
2817 /*
2818 * If we are not willing to accept PIM packets in clear,
2819 * then check with global policy.
2820 */
2821 if (ipst->ips_pim_accept_clear_messages == 0) {
2822 mp = ipsec_check_global_policy(mp, NULL,
2823 ipha, NULL, ira, ns);
2824 if (mp == NULL)
2825 return;
2826 }
2827 if ((ira->ira_flags & IRAF_SYSTEM_LABELED) &&
2828 !tsol_can_accept_raw(mp, ira, B_TRUE)) {
2829 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2830 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2831 freemsg(mp);
2832 return;
2833 }
2834 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2835
2836 /* Checksum is verified in pim_input */
2837 mp = pim_input(mp, ira);
2838 if (mp == NULL) {
2839 /* Bad packet - discarded by pim_input */
2840 return;
2841 }
2842 break;
2843 case IPPROTO_AH:
2844 case IPPROTO_ESP: {
2845 /*
2846 * Fast path for AH/ESP.
2847 */
2848 netstack_t *ns = ipst->ips_netstack;
2849 ipsec_stack_t *ipss = ns->netstack_ipsec;
2850
2851 IP_STAT(ipst, ipsec_proto_ahesp);
2852
2853 if (!ipsec_loaded(ipss)) {
2854 ip_proto_not_sup(mp, ira);
2855 return;
2856 }
2857
2858 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
2859 /* select inbound SA and have IPsec process the pkt */
2860 if (protocol == IPPROTO_ESP) {
2861 esph_t *esph;
2862 boolean_t esp_in_udp_sa;
2863 boolean_t esp_in_udp_packet;
2864
2865 mp = ipsec_inbound_esp_sa(mp, ira, &esph);
2866 if (mp == NULL)
2867 return;
2868
2869 ASSERT(esph != NULL);
2870 ASSERT(ira->ira_flags & IRAF_IPSEC_SECURE);
2871 ASSERT(ira->ira_ipsec_esp_sa != NULL);
2872 ASSERT(ira->ira_ipsec_esp_sa->ipsa_input_func != NULL);
2873
2874 esp_in_udp_sa = ((ira->ira_ipsec_esp_sa->ipsa_flags &
2875 IPSA_F_NATT) != 0);
2876 esp_in_udp_packet =
2877 (ira->ira_flags & IRAF_ESP_UDP_PORTS) != 0;
2878
2879 /*
2880 * The following is a fancy, but quick, way of saying:
2881 * ESP-in-UDP SA and Raw ESP packet --> drop
2882 * OR
2883 * ESP SA and ESP-in-UDP packet --> drop
2884 */
2885 if (esp_in_udp_sa != esp_in_udp_packet) {
2886 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2887 ip_drop_packet(mp, B_TRUE, ira->ira_ill,
2888 DROPPER(ipss, ipds_esp_no_sa),
2889 &ipss->ipsec_dropper);
2890 return;
2891 }
2892 mp = ira->ira_ipsec_esp_sa->ipsa_input_func(mp, esph,
2893 ira);
2894 } else {
2895 ah_t *ah;
2896
2897 mp = ipsec_inbound_ah_sa(mp, ira, &ah);
2898 if (mp == NULL)
2899 return;
2900
2901 ASSERT(ah != NULL);
2902 ASSERT(ira->ira_flags & IRAF_IPSEC_SECURE);
2903 ASSERT(ira->ira_ipsec_ah_sa != NULL);
2904 ASSERT(ira->ira_ipsec_ah_sa->ipsa_input_func != NULL);
2905 mp = ira->ira_ipsec_ah_sa->ipsa_input_func(mp, ah,
2906 ira);
2907 }
2908
2909 if (mp == NULL) {
2910 /*
2911 * Either it failed or is pending. In the former case
2912 * ipIfStatsInDiscards was increased.
2913 */
2914 return;
2915 }
2916 /* we're done with IPsec processing, send it up */
2917 ip_input_post_ipsec(mp, ira);
2918 return;
2919 }
2920 case IPPROTO_ENCAP: {
2921 ipha_t *inner_ipha;
2922
2923 /*
2924 * Handle self-encapsulated packets (IP-in-IP where
2925 * the inner addresses == the outer addresses).
2926 */
2927 if ((uchar_t *)ipha + ip_hdr_length + sizeof (ipha_t) >
2928 mp->b_wptr) {
2929 if (ira->ira_pktlen <
2930 ip_hdr_length + sizeof (ipha_t)) {
2931 BUMP_MIB(ill->ill_ip_mib,
2932 ipIfStatsInTruncatedPkts);
2933 ip_drop_input("ipIfStatsInTruncatedPkts",
2934 mp, ill);
2935 freemsg(mp);
2936 return;
2937 }
2938 ipha = ip_pullup(mp, (uchar_t *)ipha + ip_hdr_length +
2939 sizeof (ipha_t) - mp->b_rptr, ira);
2940 if (ipha == NULL) {
2941 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2942 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2943 freemsg(mp);
2944 return;
2945 }
2946 }
2947 inner_ipha = (ipha_t *)((uchar_t *)ipha + ip_hdr_length);
2948 /*
2949 * Check the sanity of the inner IP header.
2950 */
2951 if ((IPH_HDR_VERSION(inner_ipha) != IPV4_VERSION)) {
2952 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2953 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2954 freemsg(mp);
2955 return;
2956 }
2957 if (IPH_HDR_LENGTH(inner_ipha) < sizeof (ipha_t)) {
2958 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2959 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2960 freemsg(mp);
2961 return;
2962 }
2963 if (inner_ipha->ipha_src != ipha->ipha_src ||
2964 inner_ipha->ipha_dst != ipha->ipha_dst) {
2965 /* We fallthru to iptun fanout below */
2966 goto iptun;
2967 }
2968
2969 /*
2970 * Self-encapsulated tunnel packet. Remove
2971 * the outer IP header and fanout again.
2972 * We also need to make sure that the inner
2973 * header is pulled up until options.
2974 */
2975 mp->b_rptr = (uchar_t *)inner_ipha;
2976 ipha = inner_ipha;
2977 ip_hdr_length = IPH_HDR_LENGTH(ipha);
2978 if ((uchar_t *)ipha + ip_hdr_length > mp->b_wptr) {
2979 if (ira->ira_pktlen <
2980 (uchar_t *)ipha + ip_hdr_length - mp->b_rptr) {
2981 BUMP_MIB(ill->ill_ip_mib,
2982 ipIfStatsInTruncatedPkts);
2983 ip_drop_input("ipIfStatsInTruncatedPkts",
2984 mp, ill);
2985 freemsg(mp);
2986 return;
2987 }
2988 ipha = ip_pullup(mp,
2989 (uchar_t *)ipha + ip_hdr_length - mp->b_rptr, ira);
2990 if (ipha == NULL) {
2991 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
2992 ip_drop_input("ipIfStatsInDiscards", mp, ill);
2993 freemsg(mp);
2994 return;
2995 }
2996 }
2997 if (ip_hdr_length > sizeof (ipha_t)) {
2998 /* We got options on the inner packet. */
2999 ipaddr_t dst = ipha->ipha_dst;
3000 int error = 0;
3001
3002 dst = ip_input_options(ipha, dst, mp, ira, &error);
3003 if (error != 0) {
3004 /*
3005 * An ICMP error has been sent and the packet
3006 * has been dropped.
3007 */
3008 return;
3009 }
3010 if (dst != ipha->ipha_dst) {
3011 /*
3012 * Someone put a source-route in
3013 * the inside header of a self-
3014 * encapsulated packet. Drop it
3015 * with extreme prejudice and let
3016 * the sender know.
3017 */
3018 ip_drop_input("ICMP_SOURCE_ROUTE_FAILED",
3019 mp, ill);
3020 icmp_unreachable(mp, ICMP_SOURCE_ROUTE_FAILED,
3021 ira);
3022 return;
3023 }
3024 }
3025 if (!(ira->ira_flags & IRAF_IPSEC_SECURE)) {
3026 /*
3027 * This means that somebody is sending
3028 * Self-encapsualted packets without AH/ESP.
3029 *
3030 * Send this packet to find a tunnel endpoint.
3031 * if I can't find one, an ICMP
3032 * PROTOCOL_UNREACHABLE will get sent.
3033 */
3034 protocol = ipha->ipha_protocol;
3035 ira->ira_protocol = protocol;
3036 goto iptun;
3037 }
3038
3039 /* Update based on removed IP header */
3040 ira->ira_ip_hdr_length = ip_hdr_length;
3041 ira->ira_pktlen = ntohs(ipha->ipha_length);
3042
3043 if (ira->ira_flags & IRAF_IPSEC_DECAPS) {
3044 /*
3045 * This packet is self-encapsulated multiple
3046 * times. We don't want to recurse infinitely.
3047 * To keep it simple, drop the packet.
3048 */
3049 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
3050 ip_drop_input("ipIfStatsInDiscards", mp, ill);
3051 freemsg(mp);
3052 return;
3053 }
3054 ASSERT(ira->ira_flags & IRAF_IPSEC_SECURE);
3055 ira->ira_flags |= IRAF_IPSEC_DECAPS;
3056
3057 ip_input_post_ipsec(mp, ira);
3058 return;
3059 }
3060
3061 iptun: /* IPPROTO_ENCAPS that is not self-encapsulated */
3062 case IPPROTO_IPV6:
3063 /* iptun will verify trusted label */
3064 connp = ipcl_classify_v4(mp, protocol, ip_hdr_length,
3065 ira, ipst);
3066 if (connp != NULL) {
3067 BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInDelivers);
3068 ira->ira_ill = ira->ira_rill = NULL;
3069 (connp->conn_recv)(connp, mp, NULL, ira);
3070 CONN_DEC_REF(connp);
3071 ira->ira_ill = ill;
3072 ira->ira_rill = rill;
3073 return;
3074 }
3075 /* FALLTHRU */
3076 default:
3077 /*
3078 * On a labeled system, we have to check whether the zone
3079 * itself is permitted to receive raw traffic.
3080 */
3081 if (ira->ira_flags & IRAF_SYSTEM_LABELED) {
3082 if (!tsol_can_accept_raw(mp, ira, B_FALSE)) {
3083 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
3084 ip_drop_input("ipIfStatsInDiscards", mp, ill);
3085 freemsg(mp);
3086 return;
3087 }
3088 }
3089 break;
3090 }
3091
3092 /*
3093 * The above input functions may have returned the pulled up message.
3094 * So ipha need to be reinitialized.
3095 */
3096 ipha = (ipha_t *)mp->b_rptr;
3097 ira->ira_protocol = protocol = ipha->ipha_protocol;
3098 if (ipst->ips_ipcl_proto_fanout_v4[protocol].connf_head == NULL) {
3099 /*
3100 * No user-level listener for these packets packets.
3101 * Check for IPPROTO_ENCAP...
3102 */
3103 if (protocol == IPPROTO_ENCAP && ipst->ips_ip_g_mrouter) {
3104 /*
3105 * Check policy here,
3106 * THEN ship off to ip_mroute_decap().
3107 *
3108 * BTW, If I match a configured IP-in-IP
3109 * tunnel above, this path will not be reached, and
3110 * ip_mroute_decap will never be called.
3111 */
3112 mp = ipsec_check_global_policy(mp, connp,
3113 ipha, NULL, ira, ns);
3114 if (mp != NULL) {
3115 ip_mroute_decap(mp, ira);
3116 } /* Else we already freed everything! */
3117 } else {
3118 ip_proto_not_sup(mp, ira);
3119 }
3120 return;
3121 }
3122
3123 /*
3124 * Handle fanout to raw sockets. There
3125 * can be more than one stream bound to a particular
3126 * protocol. When this is the case, each one gets a copy
3127 * of any incoming packets.
3128 */
3129 ASSERT(ira->ira_protocol == ipha->ipha_protocol);
3130 ip_fanout_proto_v4(mp, ipha, ira);
3131 return;
3132
3133 discard:
3134 BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
3135 ip_drop_input("ipIfStatsInDiscards", mp, ill);
3136 freemsg(mp);
3137 #undef rptr
3138 }
3139