19454b2d8SWarner Losh /*- 251369649SPedro F. Giffuni * SPDX-License-Identifier: BSD-3-Clause 351369649SPedro F. Giffuni * 4df8bae1dSRodney W. Grimes * Copyright (c) 1982, 1986, 1988, 1991, 1993 5df8bae1dSRodney W. Grimes * The Regents of the University of California. All rights reserved. 6df8bae1dSRodney W. Grimes * 7df8bae1dSRodney W. Grimes * Redistribution and use in source and binary forms, with or without 8df8bae1dSRodney W. Grimes * modification, are permitted provided that the following conditions 9df8bae1dSRodney W. Grimes * are met: 10df8bae1dSRodney W. Grimes * 1. Redistributions of source code must retain the above copyright 11df8bae1dSRodney W. Grimes * notice, this list of conditions and the following disclaimer. 12df8bae1dSRodney W. Grimes * 2. Redistributions in binary form must reproduce the above copyright 13df8bae1dSRodney W. Grimes * notice, this list of conditions and the following disclaimer in the 14df8bae1dSRodney W. Grimes * documentation and/or other materials provided with the distribution. 1569a28758SEd Maste * 3. Neither the name of the University nor the names of its contributors 16df8bae1dSRodney W. Grimes * may be used to endorse or promote products derived from this software 17df8bae1dSRodney W. Grimes * without specific prior written permission. 18df8bae1dSRodney W. Grimes * 19df8bae1dSRodney W. Grimes * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 20df8bae1dSRodney W. Grimes * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 21df8bae1dSRodney W. Grimes * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 22df8bae1dSRodney W. Grimes * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 23df8bae1dSRodney W. Grimes * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 24df8bae1dSRodney W. Grimes * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 25df8bae1dSRodney W. Grimes * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 26df8bae1dSRodney W. Grimes * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 27df8bae1dSRodney W. Grimes * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 28df8bae1dSRodney W. Grimes * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 29df8bae1dSRodney W. Grimes * SUCH DAMAGE. 30df8bae1dSRodney W. Grimes */ 31df8bae1dSRodney W. Grimes 32677b542eSDavid E. O'Brien #include <sys/cdefs.h> 33240ef842SDavid E. O'Brien #include "opt_param.h" 34352d050eSMike Silbersack #include "opt_mbuf_stress_test.h" 356eeac1d9SJulian Elischer #include "opt_mbuf_profiling.h" 36e32a5b94SRobert Watson 37df8bae1dSRodney W. Grimes #include <sys/param.h> 38df8bae1dSRodney W. Grimes #include <sys/systm.h> 39fb919e4dSMark Murray #include <sys/kernel.h> 40beb699c7SMike Silbersack #include <sys/limits.h> 41fb919e4dSMark Murray #include <sys/lock.h> 42f9d0d524SRobert Watson #include <sys/malloc.h> 43df8bae1dSRodney W. Grimes #include <sys/mbuf.h> 44639acc13SGarrett Wollman #include <sys/sysctl.h> 45df8bae1dSRodney W. Grimes #include <sys/domain.h> 46df8bae1dSRodney W. Grimes #include <sys/protosw.h> 47beb699c7SMike Silbersack #include <sys/uio.h> 4882334850SJohn Baldwin #include <sys/vmmeter.h> 49c6c52d8eSAlexander Motin #include <sys/sbuf.h> 50480f4e94SGeorge V. Neville-Neil #include <sys/sdt.h> 5182334850SJohn Baldwin #include <vm/vm.h> 5282334850SJohn Baldwin #include <vm/vm_pageout.h> 5382334850SJohn Baldwin #include <vm/vm_page.h> 54480f4e94SGeorge V. Neville-Neil 55dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE5_XLATE(sdt, , , m__init, 56480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *", 57480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 58480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 59480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 60480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t"); 61480f4e94SGeorge V. Neville-Neil 62c2c34ee5SMateusz Guzik SDT_PROBE_DEFINE3_XLATE(sdt, , , m__gethdr_raw, 63c2c34ee5SMateusz Guzik "uint32_t", "uint32_t", 64c2c34ee5SMateusz Guzik "uint16_t", "uint16_t", 65c2c34ee5SMateusz Guzik "struct mbuf *", "mbufinfo_t *"); 66c2c34ee5SMateusz Guzik 67dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE3_XLATE(sdt, , , m__gethdr, 68480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 69480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 70480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 71480f4e94SGeorge V. Neville-Neil 72c2c34ee5SMateusz Guzik SDT_PROBE_DEFINE3_XLATE(sdt, , , m__get_raw, 73c2c34ee5SMateusz Guzik "uint32_t", "uint32_t", 74c2c34ee5SMateusz Guzik "uint16_t", "uint16_t", 75c2c34ee5SMateusz Guzik "struct mbuf *", "mbufinfo_t *"); 76c2c34ee5SMateusz Guzik 77dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE3_XLATE(sdt, , , m__get, 78480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 79480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 80480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 81480f4e94SGeorge V. Neville-Neil 82dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE4_XLATE(sdt, , , m__getcl, 83480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 84480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 85480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 86480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 87480f4e94SGeorge V. Neville-Neil 88edde7a53SAndrey V. Elsukov SDT_PROBE_DEFINE5_XLATE(sdt, , , m__getjcl, 89edde7a53SAndrey V. Elsukov "uint32_t", "uint32_t", 90edde7a53SAndrey V. Elsukov "uint16_t", "uint16_t", 91edde7a53SAndrey V. Elsukov "uint32_t", "uint32_t", 92edde7a53SAndrey V. Elsukov "uint32_t", "uint32_t", 93edde7a53SAndrey V. Elsukov "struct mbuf *", "mbufinfo_t *"); 94edde7a53SAndrey V. Elsukov 95dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE3_XLATE(sdt, , , m__clget, 96480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *", 97480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 98480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t"); 99480f4e94SGeorge V. Neville-Neil 100dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE4_XLATE(sdt, , , m__cljget, 101480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *", 102480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 103480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 104480f4e94SGeorge V. Neville-Neil "void*", "void*"); 105480f4e94SGeorge V. Neville-Neil 106dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE(sdt, , , m__cljset); 107480f4e94SGeorge V. Neville-Neil 108dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE1_XLATE(sdt, , , m__free, 109480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 110480f4e94SGeorge V. Neville-Neil 111dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE1_XLATE(sdt, , , m__freem, 112480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 113fb919e4dSMark Murray 1145e4bc63bSGleb Smirnoff #include <security/mac/mac_framework.h> 1155e4bc63bSGleb Smirnoff 116c414347bSGleb Smirnoff /* 117c414347bSGleb Smirnoff * Provide minimum possible defaults for link and protocol header space, 118c414347bSGleb Smirnoff * assuming IPv4 over Ethernet. Enabling IPv6, IEEE802.11 or some other 119c414347bSGleb Smirnoff * protocol may grow these values. 120c414347bSGleb Smirnoff */ 121c414347bSGleb Smirnoff u_int max_linkhdr = 16; 122c414347bSGleb Smirnoff u_int max_protohdr = 40; 123c414347bSGleb Smirnoff u_int max_hdr = 16 + 40; 124c414347bSGleb Smirnoff SYSCTL_INT(_kern_ipc, KIPC_MAX_LINKHDR, max_linkhdr, CTLFLAG_RD, 125c414347bSGleb Smirnoff &max_linkhdr, 16, "Size of largest link layer header"); 126c414347bSGleb Smirnoff SYSCTL_INT(_kern_ipc, KIPC_MAX_PROTOHDR, max_protohdr, CTLFLAG_RD, 127c414347bSGleb Smirnoff &max_protohdr, 40, "Size of largest protocol layer header"); 128c414347bSGleb Smirnoff SYSCTL_INT(_kern_ipc, KIPC_MAX_HDR, max_hdr, CTLFLAG_RD, 129c414347bSGleb Smirnoff &max_hdr, 16 + 40, "Size of largest link plus protocol header"); 130c414347bSGleb Smirnoff 131c414347bSGleb Smirnoff static void 132c414347bSGleb Smirnoff max_hdr_grow(void) 133c414347bSGleb Smirnoff { 134c414347bSGleb Smirnoff 135c414347bSGleb Smirnoff max_hdr = max_linkhdr + max_protohdr; 136c414347bSGleb Smirnoff MPASS(max_hdr <= MHLEN); 137c414347bSGleb Smirnoff } 138c414347bSGleb Smirnoff 139c414347bSGleb Smirnoff void 140c414347bSGleb Smirnoff max_linkhdr_grow(u_int new) 141c414347bSGleb Smirnoff { 142c414347bSGleb Smirnoff 143c414347bSGleb Smirnoff if (new > max_linkhdr) { 144c414347bSGleb Smirnoff max_linkhdr = new; 145c414347bSGleb Smirnoff max_hdr_grow(); 146c414347bSGleb Smirnoff } 147c414347bSGleb Smirnoff } 148c414347bSGleb Smirnoff 149c414347bSGleb Smirnoff void 150c414347bSGleb Smirnoff max_protohdr_grow(u_int new) 151c414347bSGleb Smirnoff { 152c414347bSGleb Smirnoff 153c414347bSGleb Smirnoff if (new > max_protohdr) { 154c414347bSGleb Smirnoff max_protohdr = new; 155c414347bSGleb Smirnoff max_hdr_grow(); 156c414347bSGleb Smirnoff } 157c414347bSGleb Smirnoff } 158c414347bSGleb Smirnoff 15951710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 16055e9f80dSMike Silbersack int m_defragpackets; 16155e9f80dSMike Silbersack int m_defragbytes; 16255e9f80dSMike Silbersack int m_defraguseless; 16355e9f80dSMike Silbersack int m_defragfailure; 164352d050eSMike Silbersack int m_defragrandomfailures; 1657d032714SBosko Milekic 16655e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragpackets, CTLFLAG_RD, 16755e9f80dSMike Silbersack &m_defragpackets, 0, ""); 16855e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragbytes, CTLFLAG_RD, 16955e9f80dSMike Silbersack &m_defragbytes, 0, ""); 17055e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defraguseless, CTLFLAG_RD, 17155e9f80dSMike Silbersack &m_defraguseless, 0, ""); 17255e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragfailure, CTLFLAG_RD, 17355e9f80dSMike Silbersack &m_defragfailure, 0, ""); 174352d050eSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragrandomfailures, CTLFLAG_RW, 175352d050eSMike Silbersack &m_defragrandomfailures, 0, ""); 176352d050eSMike Silbersack #endif 177df8bae1dSRodney W. Grimes 178df8bae1dSRodney W. Grimes /* 179f729ede6SAndre Oppermann * Ensure the correct size of various mbuf parameters. It could be off due 180f729ede6SAndre Oppermann * to compiler-induced padding and alignment artifacts. 181f729ede6SAndre Oppermann */ 182f729ede6SAndre Oppermann CTASSERT(MSIZE - offsetof(struct mbuf, m_dat) == MLEN); 183f729ede6SAndre Oppermann CTASSERT(MSIZE - offsetof(struct mbuf, m_pktdat) == MHLEN); 184f729ede6SAndre Oppermann 185f729ede6SAndre Oppermann /* 1863d1a9ed3SRobert Watson * mbuf data storage should be 64-bit aligned regardless of architectural 1873d1a9ed3SRobert Watson * pointer size; check this is the case with and without a packet header. 1883d1a9ed3SRobert Watson */ 1893d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_dat) % 8 == 0); 1903d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_pktdat) % 8 == 0); 1913d1a9ed3SRobert Watson 1923d1a9ed3SRobert Watson /* 1933d1a9ed3SRobert Watson * While the specific values here don't matter too much (i.e., +/- a few 1943d1a9ed3SRobert Watson * words), we do want to ensure that changes to these values are carefully 1953d1a9ed3SRobert Watson * reasoned about and properly documented. This is especially the case as 1963d1a9ed3SRobert Watson * network-protocol and device-driver modules encode these layouts, and must 1973d1a9ed3SRobert Watson * be recompiled if the structures change. Check these values at compile time 1983d1a9ed3SRobert Watson * against the ones documented in comments in mbuf.h. 1993d1a9ed3SRobert Watson * 2003d1a9ed3SRobert Watson * NB: Possibly they should be documented there via #define's and not just 2013d1a9ed3SRobert Watson * comments. 2023d1a9ed3SRobert Watson */ 2033d1a9ed3SRobert Watson #if defined(__LP64__) 2043d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_dat) == 32); 2054d88d81cSHans Petter Selasky CTASSERT(sizeof(struct pkthdr) == 64); 2060c103266SGleb Smirnoff CTASSERT(sizeof(struct m_ext) == 160); 2073d1a9ed3SRobert Watson #else 2083d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_dat) == 24); 209b8394039SHans Petter Selasky CTASSERT(sizeof(struct pkthdr) == 56); 210f57d153eSBrandon Bergren #if defined(__powerpc__) && defined(BOOKE) 211f57d153eSBrandon Bergren /* PowerPC booke has 64-bit physical pointers. */ 212b8394039SHans Petter Selasky CTASSERT(sizeof(struct m_ext) == 176); 213f57d153eSBrandon Bergren #else 214b8394039SHans Petter Selasky CTASSERT(sizeof(struct m_ext) == 172); 2153d1a9ed3SRobert Watson #endif 216f57d153eSBrandon Bergren #endif 2173d1a9ed3SRobert Watson 2183d1a9ed3SRobert Watson /* 219ec9d83ddSGleb Smirnoff * Assert that the queue(3) macros produce code of the same size as an old 220ec9d83ddSGleb Smirnoff * plain pointer does. 221ec9d83ddSGleb Smirnoff */ 222ec9d83ddSGleb Smirnoff #ifdef INVARIANTS 223d53d6fa9SMark Johnston static struct mbuf __used m_assertbuf; 224ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_slist) == sizeof(m_assertbuf.m_next)); 225ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_stailq) == sizeof(m_assertbuf.m_next)); 226ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_slistpkt) == sizeof(m_assertbuf.m_nextpkt)); 227ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_stailqpkt) == sizeof(m_assertbuf.m_nextpkt)); 228ec9d83ddSGleb Smirnoff #endif 229ec9d83ddSGleb Smirnoff 230ec9d83ddSGleb Smirnoff /* 2316bccea7cSRebecca Cran * Attach the cluster from *m to *n, set up m_ext in *n 23256a4e45aSAndre Oppermann * and bump the refcount of the cluster. 23356a4e45aSAndre Oppermann */ 23486a996e6SHiren Panchasara void 23556a5f52eSGleb Smirnoff mb_dupcl(struct mbuf *n, struct mbuf *m) 23656a4e45aSAndre Oppermann { 23756a5f52eSGleb Smirnoff volatile u_int *refcnt; 23856a4e45aSAndre Oppermann 23961664ee7SGleb Smirnoff KASSERT(m->m_flags & (M_EXT | M_EXTPG), 24061664ee7SGleb Smirnoff ("%s: M_EXT | M_EXTPG not set on %p", __func__, m)); 24161664ee7SGleb Smirnoff KASSERT(!(n->m_flags & (M_EXT | M_EXTPG)), 24261664ee7SGleb Smirnoff ("%s: M_EXT | M_EXTPG set on %p", __func__, n)); 2431fbe6a82SGleb Smirnoff 24407e87a1dSGleb Smirnoff /* 2450c103266SGleb Smirnoff * Cache access optimization. 2460c103266SGleb Smirnoff * 2470c103266SGleb Smirnoff * o Regular M_EXT storage doesn't need full copy of m_ext, since 2480c103266SGleb Smirnoff * the holder of the 'ext_count' is responsible to carry the free 2490c103266SGleb Smirnoff * routine and its arguments. 25061664ee7SGleb Smirnoff * o M_EXTPG data is split between main part of mbuf and m_ext, the 2510c103266SGleb Smirnoff * main part is copied in full, the m_ext part is similar to M_EXT. 2520c103266SGleb Smirnoff * o EXT_EXTREF, where 'ext_cnt' doesn't point into mbuf at all, is 2530c103266SGleb Smirnoff * special - it needs full copy of m_ext into each mbuf, since any 2540c103266SGleb Smirnoff * copy could end up as the last to free. 25507e87a1dSGleb Smirnoff */ 25661664ee7SGleb Smirnoff if (m->m_flags & M_EXTPG) { 2577b6c99d0SGleb Smirnoff bcopy(&m->m_epg_startcopy, &n->m_epg_startcopy, 2587b6c99d0SGleb Smirnoff __rangeof(struct mbuf, m_epg_startcopy, m_epg_endcopy)); 2597b6c99d0SGleb Smirnoff bcopy(&m->m_ext, &n->m_ext, m_epg_ext_copylen); 26061664ee7SGleb Smirnoff } else if (m->m_ext.ext_type == EXT_EXTREF) 2610c103266SGleb Smirnoff bcopy(&m->m_ext, &n->m_ext, sizeof(struct m_ext)); 26261664ee7SGleb Smirnoff else 26307e87a1dSGleb Smirnoff bcopy(&m->m_ext, &n->m_ext, m_ext_copylen); 2640c103266SGleb Smirnoff 26561664ee7SGleb Smirnoff n->m_flags |= m->m_flags & (M_RDONLY | M_EXT | M_EXTPG); 26656a5f52eSGleb Smirnoff 26756a5f52eSGleb Smirnoff /* See if this is the mbuf that holds the embedded refcount. */ 26856a5f52eSGleb Smirnoff if (m->m_ext.ext_flags & EXT_FLAG_EMBREF) { 26956a5f52eSGleb Smirnoff refcnt = n->m_ext.ext_cnt = &m->m_ext.ext_count; 27056a5f52eSGleb Smirnoff n->m_ext.ext_flags &= ~EXT_FLAG_EMBREF; 27156a5f52eSGleb Smirnoff } else { 27256a5f52eSGleb Smirnoff KASSERT(m->m_ext.ext_cnt != NULL, 27356a5f52eSGleb Smirnoff ("%s: no refcounting pointer on %p", __func__, m)); 27456a5f52eSGleb Smirnoff refcnt = m->m_ext.ext_cnt; 27556a5f52eSGleb Smirnoff } 27656a5f52eSGleb Smirnoff 27756a5f52eSGleb Smirnoff if (*refcnt == 1) 27856a5f52eSGleb Smirnoff *refcnt += 1; 27956a5f52eSGleb Smirnoff else 28056a5f52eSGleb Smirnoff atomic_add_int(refcnt, 1); 28156a4e45aSAndre Oppermann } 28256a4e45aSAndre Oppermann 2839523d1bfSNavdeep Parhar void 2849523d1bfSNavdeep Parhar m_demote_pkthdr(struct mbuf *m) 2859523d1bfSNavdeep Parhar { 2869523d1bfSNavdeep Parhar 2879523d1bfSNavdeep Parhar M_ASSERTPKTHDR(m); 28817cbcf33SHans Petter Selasky M_ASSERT_NO_SND_TAG(m); 2899523d1bfSNavdeep Parhar 2909523d1bfSNavdeep Parhar m_tag_delete_chain(m, NULL); 2919523d1bfSNavdeep Parhar m->m_flags &= ~M_PKTHDR; 2929523d1bfSNavdeep Parhar bzero(&m->m_pkthdr, sizeof(struct pkthdr)); 2939523d1bfSNavdeep Parhar } 2949523d1bfSNavdeep Parhar 29556a4e45aSAndre Oppermann /* 296ed111688SAndre Oppermann * Clean up mbuf (chain) from any tags and packet headers. 297e0068c3aSAndre Oppermann * If "all" is set then the first mbuf in the chain will be 298e0068c3aSAndre Oppermann * cleaned too. 299ed111688SAndre Oppermann */ 300ed111688SAndre Oppermann void 301651e4e6aSGleb Smirnoff m_demote(struct mbuf *m0, int all, int flags) 302ed111688SAndre Oppermann { 303ed111688SAndre Oppermann struct mbuf *m; 304ed111688SAndre Oppermann 305dd31400cSHans Petter Selasky flags |= M_DEMOTEFLAGS; 306dd31400cSHans Petter Selasky 307ed111688SAndre Oppermann for (m = all ? m0 : m0->m_next; m != NULL; m = m->m_next) { 3087ee2d058SGleb Smirnoff KASSERT(m->m_nextpkt == NULL, ("%s: m_nextpkt in m %p, m0 %p", 3097ee2d058SGleb Smirnoff __func__, m, m0)); 3109523d1bfSNavdeep Parhar if (m->m_flags & M_PKTHDR) 3119523d1bfSNavdeep Parhar m_demote_pkthdr(m); 312dd31400cSHans Petter Selasky m->m_flags &= flags; 313ed111688SAndre Oppermann } 314ed111688SAndre Oppermann } 315ed111688SAndre Oppermann 316ed111688SAndre Oppermann /* 317fdcc028dSAndre Oppermann * Sanity checks on mbuf (chain) for use in KASSERT() and general 318fdcc028dSAndre Oppermann * debugging. 319fdcc028dSAndre Oppermann * Returns 0 or panics when bad and 1 on all tests passed. 320fdcc028dSAndre Oppermann * Sanitize, 0 to run M_SANITY_ACTION, 1 to garble things so they 321fdcc028dSAndre Oppermann * blow up later. 322a048affbSAndre Oppermann */ 323a048affbSAndre Oppermann int 324a048affbSAndre Oppermann m_sanity(struct mbuf *m0, int sanitize) 325a048affbSAndre Oppermann { 326a048affbSAndre Oppermann struct mbuf *m; 327a048affbSAndre Oppermann caddr_t a, b; 328a048affbSAndre Oppermann int pktlen = 0; 329a048affbSAndre Oppermann 33021ee3e7aSKip Macy #ifdef INVARIANTS 33121ee3e7aSKip Macy #define M_SANITY_ACTION(s) panic("mbuf %p: " s, m) 33221ee3e7aSKip Macy #else 33321ee3e7aSKip Macy #define M_SANITY_ACTION(s) printf("mbuf %p: " s, m) 33421ee3e7aSKip Macy #endif 335a048affbSAndre Oppermann 336fdcc028dSAndre Oppermann for (m = m0; m != NULL; m = m->m_next) { 337a048affbSAndre Oppermann /* 338a048affbSAndre Oppermann * Basic pointer checks. If any of these fails then some 339a048affbSAndre Oppermann * unrelated kernel memory before or after us is trashed. 340a048affbSAndre Oppermann * No way to recover from that. 341a048affbSAndre Oppermann */ 342b66f2a48SRobert Watson a = M_START(m); 343b66f2a48SRobert Watson b = a + M_SIZE(m); 344a048affbSAndre Oppermann if ((caddr_t)m->m_data < a) 345a048affbSAndre Oppermann M_SANITY_ACTION("m_data outside mbuf data range left"); 346a048affbSAndre Oppermann if ((caddr_t)m->m_data > b) 347a048affbSAndre Oppermann M_SANITY_ACTION("m_data outside mbuf data range right"); 348a048affbSAndre Oppermann if ((caddr_t)m->m_data + m->m_len > b) 349a048affbSAndre Oppermann M_SANITY_ACTION("m_data + m_len exeeds mbuf space"); 350a048affbSAndre Oppermann 351a048affbSAndre Oppermann /* m->m_nextpkt may only be set on first mbuf in chain. */ 352fdcc028dSAndre Oppermann if (m != m0 && m->m_nextpkt != NULL) { 353a048affbSAndre Oppermann if (sanitize) { 354a048affbSAndre Oppermann m_freem(m->m_nextpkt); 355a048affbSAndre Oppermann m->m_nextpkt = (struct mbuf *)0xDEADC0DE; 356a048affbSAndre Oppermann } else 357a048affbSAndre Oppermann M_SANITY_ACTION("m->m_nextpkt on in-chain mbuf"); 358a048affbSAndre Oppermann } 359a048affbSAndre Oppermann 360a048affbSAndre Oppermann /* packet length (not mbuf length!) calculation */ 361a048affbSAndre Oppermann if (m0->m_flags & M_PKTHDR) 362a048affbSAndre Oppermann pktlen += m->m_len; 363a048affbSAndre Oppermann 364a048affbSAndre Oppermann /* m_tags may only be attached to first mbuf in chain. */ 365a048affbSAndre Oppermann if (m != m0 && m->m_flags & M_PKTHDR && 366a048affbSAndre Oppermann !SLIST_EMPTY(&m->m_pkthdr.tags)) { 367a048affbSAndre Oppermann if (sanitize) { 368a048affbSAndre Oppermann m_tag_delete_chain(m, NULL); 369a048affbSAndre Oppermann /* put in 0xDEADC0DE perhaps? */ 370fdcc028dSAndre Oppermann } else 371a048affbSAndre Oppermann M_SANITY_ACTION("m_tags on in-chain mbuf"); 372a048affbSAndre Oppermann } 373a048affbSAndre Oppermann 374a048affbSAndre Oppermann /* M_PKTHDR may only be set on first mbuf in chain */ 375a048affbSAndre Oppermann if (m != m0 && m->m_flags & M_PKTHDR) { 376a048affbSAndre Oppermann if (sanitize) { 377a048affbSAndre Oppermann bzero(&m->m_pkthdr, sizeof(m->m_pkthdr)); 378a048affbSAndre Oppermann m->m_flags &= ~M_PKTHDR; 379a048affbSAndre Oppermann /* put in 0xDEADCODE and leave hdr flag in */ 380a048affbSAndre Oppermann } else 381a048affbSAndre Oppermann M_SANITY_ACTION("M_PKTHDR on in-chain mbuf"); 382a048affbSAndre Oppermann } 383a048affbSAndre Oppermann } 384fdcc028dSAndre Oppermann m = m0; 385fdcc028dSAndre Oppermann if (pktlen && pktlen != m->m_pkthdr.len) { 386a048affbSAndre Oppermann if (sanitize) 387fdcc028dSAndre Oppermann m->m_pkthdr.len = 0; 388a048affbSAndre Oppermann else 389a048affbSAndre Oppermann M_SANITY_ACTION("m_pkthdr.len != mbuf chain length"); 390a048affbSAndre Oppermann } 391a048affbSAndre Oppermann return 1; 392fdcc028dSAndre Oppermann 393fdcc028dSAndre Oppermann #undef M_SANITY_ACTION 394a048affbSAndre Oppermann } 395a048affbSAndre Oppermann 3965e4bc63bSGleb Smirnoff /* 3975e4bc63bSGleb Smirnoff * Non-inlined part of m_init(). 3985e4bc63bSGleb Smirnoff */ 3995e4bc63bSGleb Smirnoff int 4005e4bc63bSGleb Smirnoff m_pkthdr_init(struct mbuf *m, int how) 4015e4bc63bSGleb Smirnoff { 4025e4bc63bSGleb Smirnoff #ifdef MAC 4035e4bc63bSGleb Smirnoff int error; 4045e4bc63bSGleb Smirnoff #endif 4055e4bc63bSGleb Smirnoff m->m_data = m->m_pktdat; 4065e4bc63bSGleb Smirnoff bzero(&m->m_pkthdr, sizeof(m->m_pkthdr)); 40750575ce1SAndrew Gallatin #ifdef NUMA 40850575ce1SAndrew Gallatin m->m_pkthdr.numa_domain = M_NODOM; 40950575ce1SAndrew Gallatin #endif 4105e4bc63bSGleb Smirnoff #ifdef MAC 4115e4bc63bSGleb Smirnoff /* If the label init fails, fail the alloc */ 4125e4bc63bSGleb Smirnoff error = mac_mbuf_init(m, how); 4135e4bc63bSGleb Smirnoff if (error) 4145e4bc63bSGleb Smirnoff return (error); 4155e4bc63bSGleb Smirnoff #endif 4165e4bc63bSGleb Smirnoff 4175e4bc63bSGleb Smirnoff return (0); 4185e4bc63bSGleb Smirnoff } 419a048affbSAndre Oppermann 420a048affbSAndre Oppermann /* 4219967cafcSSam Leffler * "Move" mbuf pkthdr from "from" to "to". 422e37b1fcdSRobert Watson * "from" must have M_PKTHDR set, and "to" must be empty. 423e37b1fcdSRobert Watson */ 424e37b1fcdSRobert Watson void 4259967cafcSSam Leffler m_move_pkthdr(struct mbuf *to, struct mbuf *from) 426e37b1fcdSRobert Watson { 427e37b1fcdSRobert Watson 428e37b1fcdSRobert Watson #if 0 4299967cafcSSam Leffler /* see below for why these are not enabled */ 430fe584538SDag-Erling Smørgrav M_ASSERTPKTHDR(to); 431225bff6fSRobert Watson /* Note: with MAC, this may not be a good assertion. */ 4329967cafcSSam Leffler KASSERT(SLIST_EMPTY(&to->m_pkthdr.tags), 4339967cafcSSam Leffler ("m_move_pkthdr: to has tags")); 434e37b1fcdSRobert Watson #endif 435e32a5b94SRobert Watson #ifdef MAC 436225bff6fSRobert Watson /* 437225bff6fSRobert Watson * XXXMAC: It could be this should also occur for non-MAC? 438225bff6fSRobert Watson */ 439e32a5b94SRobert Watson if (to->m_flags & M_PKTHDR) 440225bff6fSRobert Watson m_tag_delete_chain(to, NULL); 441e32a5b94SRobert Watson #endif 44282334850SJohn Baldwin to->m_flags = (from->m_flags & M_COPYFLAGS) | 4436edfd179SGleb Smirnoff (to->m_flags & (M_EXT | M_EXTPG)); 444a4e71429SSam Leffler if ((to->m_flags & M_EXT) == 0) 4459967cafcSSam Leffler to->m_data = to->m_pktdat; 4469967cafcSSam Leffler to->m_pkthdr = from->m_pkthdr; /* especially tags */ 4479967cafcSSam Leffler SLIST_INIT(&from->m_pkthdr.tags); /* purge tags from src */ 4489967cafcSSam Leffler from->m_flags &= ~M_PKTHDR; 449fb3bc596SJohn Baldwin if (from->m_pkthdr.csum_flags & CSUM_SND_TAG) { 450fb3bc596SJohn Baldwin from->m_pkthdr.csum_flags &= ~CSUM_SND_TAG; 451fb3bc596SJohn Baldwin from->m_pkthdr.snd_tag = NULL; 452fb3bc596SJohn Baldwin } 4539967cafcSSam Leffler } 4549967cafcSSam Leffler 4559967cafcSSam Leffler /* 4569967cafcSSam Leffler * Duplicate "from"'s mbuf pkthdr in "to". 4579967cafcSSam Leffler * "from" must have M_PKTHDR set, and "to" must be empty. 4589967cafcSSam Leffler * In particular, this does a deep copy of the packet tags. 4599967cafcSSam Leffler */ 4609967cafcSSam Leffler int 4610cbefd30SAlexander V. Chernikov m_dup_pkthdr(struct mbuf *to, const struct mbuf *from, int how) 4629967cafcSSam Leffler { 4639967cafcSSam Leffler 4649967cafcSSam Leffler #if 0 4659967cafcSSam Leffler /* 4669967cafcSSam Leffler * The mbuf allocator only initializes the pkthdr 467c95be8b5SGleb Smirnoff * when the mbuf is allocated with m_gethdr(). Many users 468c95be8b5SGleb Smirnoff * (e.g. m_copy*, m_prepend) use m_get() and then 4699967cafcSSam Leffler * smash the pkthdr as needed causing these 4709967cafcSSam Leffler * assertions to trip. For now just disable them. 4719967cafcSSam Leffler */ 472fe584538SDag-Erling Smørgrav M_ASSERTPKTHDR(to); 473225bff6fSRobert Watson /* Note: with MAC, this may not be a good assertion. */ 4749967cafcSSam Leffler KASSERT(SLIST_EMPTY(&to->m_pkthdr.tags), ("m_dup_pkthdr: to has tags")); 4759967cafcSSam Leffler #endif 476063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 4779967cafcSSam Leffler #ifdef MAC 4789967cafcSSam Leffler if (to->m_flags & M_PKTHDR) 479225bff6fSRobert Watson m_tag_delete_chain(to, NULL); 4809967cafcSSam Leffler #endif 48182334850SJohn Baldwin to->m_flags = (from->m_flags & M_COPYFLAGS) | 4826edfd179SGleb Smirnoff (to->m_flags & (M_EXT | M_EXTPG)); 483df8c7fc9SMike Silbersack if ((to->m_flags & M_EXT) == 0) 4849967cafcSSam Leffler to->m_data = to->m_pktdat; 485e37b1fcdSRobert Watson to->m_pkthdr = from->m_pkthdr; 486fb3bc596SJohn Baldwin if (from->m_pkthdr.csum_flags & CSUM_SND_TAG) 487fb3bc596SJohn Baldwin m_snd_tag_ref(from->m_pkthdr.snd_tag); 4889967cafcSSam Leffler SLIST_INIT(&to->m_pkthdr.tags); 48994985f74SGleb Smirnoff return (m_tag_copy_chain(to, from, how)); 490e37b1fcdSRobert Watson } 491e37b1fcdSRobert Watson 492e37b1fcdSRobert Watson /* 493df8bae1dSRodney W. Grimes * Lesser-used path for M_PREPEND: 494df8bae1dSRodney W. Grimes * allocate new mbuf to prepend to chain, 495df8bae1dSRodney W. Grimes * copy junk along. 496df8bae1dSRodney W. Grimes */ 497df8bae1dSRodney W. Grimes struct mbuf * 498122a814aSBosko Milekic m_prepend(struct mbuf *m, int len, int how) 499df8bae1dSRodney W. Grimes { 500df8bae1dSRodney W. Grimes struct mbuf *mn; 501df8bae1dSRodney W. Grimes 502f8bf8e39SMike Silbersack if (m->m_flags & M_PKTHDR) 503c95be8b5SGleb Smirnoff mn = m_gethdr(how, m->m_type); 504f8bf8e39SMike Silbersack else 505c95be8b5SGleb Smirnoff mn = m_get(how, m->m_type); 506122a814aSBosko Milekic if (mn == NULL) { 507df8bae1dSRodney W. Grimes m_freem(m); 508122a814aSBosko Milekic return (NULL); 509df8bae1dSRodney W. Grimes } 510225bff6fSRobert Watson if (m->m_flags & M_PKTHDR) 511c95be8b5SGleb Smirnoff m_move_pkthdr(mn, m); 512df8bae1dSRodney W. Grimes mn->m_next = m; 513df8bae1dSRodney W. Grimes m = mn; 514ed6a66caSRobert Watson if (len < M_SIZE(m)) 5155288989fSRandall Stewart M_ALIGN(m, len); 516df8bae1dSRodney W. Grimes m->m_len = len; 517df8bae1dSRodney W. Grimes return (m); 518df8bae1dSRodney W. Grimes } 519df8bae1dSRodney W. Grimes 520df8bae1dSRodney W. Grimes /* 521df8bae1dSRodney W. Grimes * Make a copy of an mbuf chain starting "off0" bytes from the beginning, 522df8bae1dSRodney W. Grimes * continuing for "len" bytes. If len is M_COPYALL, copy to end of mbuf. 523eb1b1807SGleb Smirnoff * The wait parameter is a choice of M_WAITOK/M_NOWAIT from caller. 5241c38f2eaSArchie Cobbs * Note that the copy is read-only, because clusters are not copied, 5251c38f2eaSArchie Cobbs * only their reference counts are incremented. 526df8bae1dSRodney W. Grimes */ 527df8bae1dSRodney W. Grimes struct mbuf * 52856a5f52eSGleb Smirnoff m_copym(struct mbuf *m, int off0, int len, int wait) 529df8bae1dSRodney W. Grimes { 530122a814aSBosko Milekic struct mbuf *n, **np; 531122a814aSBosko Milekic int off = off0; 532df8bae1dSRodney W. Grimes struct mbuf *top; 533df8bae1dSRodney W. Grimes int copyhdr = 0; 534df8bae1dSRodney W. Grimes 535e0a653ddSAlfred Perlstein KASSERT(off >= 0, ("m_copym, negative off %d", off)); 536e0a653ddSAlfred Perlstein KASSERT(len >= 0, ("m_copym, negative len %d", len)); 537063d8114SAlfred Perlstein MBUF_CHECKSLEEP(wait); 538df8bae1dSRodney W. Grimes if (off == 0 && m->m_flags & M_PKTHDR) 539df8bae1dSRodney W. Grimes copyhdr = 1; 540df8bae1dSRodney W. Grimes while (off > 0) { 541e0a653ddSAlfred Perlstein KASSERT(m != NULL, ("m_copym, offset > size of mbuf chain")); 542df8bae1dSRodney W. Grimes if (off < m->m_len) 543df8bae1dSRodney W. Grimes break; 544df8bae1dSRodney W. Grimes off -= m->m_len; 545df8bae1dSRodney W. Grimes m = m->m_next; 546df8bae1dSRodney W. Grimes } 547df8bae1dSRodney W. Grimes np = ⊤ 548b85f65afSPedro F. Giffuni top = NULL; 549df8bae1dSRodney W. Grimes while (len > 0) { 550122a814aSBosko Milekic if (m == NULL) { 551e0a653ddSAlfred Perlstein KASSERT(len == M_COPYALL, 552e0a653ddSAlfred Perlstein ("m_copym, length > size of mbuf chain")); 553df8bae1dSRodney W. Grimes break; 554df8bae1dSRodney W. Grimes } 555f8bf8e39SMike Silbersack if (copyhdr) 556c95be8b5SGleb Smirnoff n = m_gethdr(wait, m->m_type); 557f8bf8e39SMike Silbersack else 558c95be8b5SGleb Smirnoff n = m_get(wait, m->m_type); 559df8bae1dSRodney W. Grimes *np = n; 560122a814aSBosko Milekic if (n == NULL) 561df8bae1dSRodney W. Grimes goto nospace; 562df8bae1dSRodney W. Grimes if (copyhdr) { 5639967cafcSSam Leffler if (!m_dup_pkthdr(n, m, wait)) 5649967cafcSSam Leffler goto nospace; 565df8bae1dSRodney W. Grimes if (len == M_COPYALL) 566df8bae1dSRodney W. Grimes n->m_pkthdr.len -= off0; 567df8bae1dSRodney W. Grimes else 568df8bae1dSRodney W. Grimes n->m_pkthdr.len = len; 569df8bae1dSRodney W. Grimes copyhdr = 0; 570df8bae1dSRodney W. Grimes } 571df8bae1dSRodney W. Grimes n->m_len = min(len, m->m_len - off); 57261664ee7SGleb Smirnoff if (m->m_flags & (M_EXT | M_EXTPG)) { 573df8bae1dSRodney W. Grimes n->m_data = m->m_data + off; 57456a4e45aSAndre Oppermann mb_dupcl(n, m); 575df8bae1dSRodney W. Grimes } else 576df8bae1dSRodney W. Grimes bcopy(mtod(m, caddr_t)+off, mtod(n, caddr_t), 577bd395ae8SBosko Milekic (u_int)n->m_len); 578df8bae1dSRodney W. Grimes if (len != M_COPYALL) 579df8bae1dSRodney W. Grimes len -= n->m_len; 580df8bae1dSRodney W. Grimes off = 0; 581df8bae1dSRodney W. Grimes m = m->m_next; 582df8bae1dSRodney W. Grimes np = &n->m_next; 583df8bae1dSRodney W. Grimes } 58408442f8aSBosko Milekic 585df8bae1dSRodney W. Grimes return (top); 586df8bae1dSRodney W. Grimes nospace: 587df8bae1dSRodney W. Grimes m_freem(top); 588122a814aSBosko Milekic return (NULL); 589df8bae1dSRodney W. Grimes } 590df8bae1dSRodney W. Grimes 591df8bae1dSRodney W. Grimes /* 5926a06dea0SGarrett Wollman * Copy an entire packet, including header (which must be present). 5936a06dea0SGarrett Wollman * An optimization of the common case `m_copym(m, 0, M_COPYALL, how)'. 5941c38f2eaSArchie Cobbs * Note that the copy is read-only, because clusters are not copied, 5951c38f2eaSArchie Cobbs * only their reference counts are incremented. 5965fe86675SLuigi Rizzo * Preserve alignment of the first mbuf so if the creator has left 5975fe86675SLuigi Rizzo * some room at the beginning (e.g. for inserting protocol headers) 5985fe86675SLuigi Rizzo * the copies still have the room available. 5996a06dea0SGarrett Wollman */ 6006a06dea0SGarrett Wollman struct mbuf * 601122a814aSBosko Milekic m_copypacket(struct mbuf *m, int how) 6026a06dea0SGarrett Wollman { 6036a06dea0SGarrett Wollman struct mbuf *top, *n, *o; 6046a06dea0SGarrett Wollman 605063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 606c95be8b5SGleb Smirnoff n = m_get(how, m->m_type); 6076a06dea0SGarrett Wollman top = n; 608122a814aSBosko Milekic if (n == NULL) 6096a06dea0SGarrett Wollman goto nospace; 6106a06dea0SGarrett Wollman 6119967cafcSSam Leffler if (!m_dup_pkthdr(n, m, how)) 6129967cafcSSam Leffler goto nospace; 6136a06dea0SGarrett Wollman n->m_len = m->m_len; 61461664ee7SGleb Smirnoff if (m->m_flags & (M_EXT | M_EXTPG)) { 6156a06dea0SGarrett Wollman n->m_data = m->m_data; 61656a4e45aSAndre Oppermann mb_dupcl(n, m); 6176a06dea0SGarrett Wollman } else { 6185fe86675SLuigi Rizzo n->m_data = n->m_pktdat + (m->m_data - m->m_pktdat ); 6196a06dea0SGarrett Wollman bcopy(mtod(m, char *), mtod(n, char *), n->m_len); 6206a06dea0SGarrett Wollman } 6216a06dea0SGarrett Wollman 6226a06dea0SGarrett Wollman m = m->m_next; 6236a06dea0SGarrett Wollman while (m) { 624c95be8b5SGleb Smirnoff o = m_get(how, m->m_type); 625122a814aSBosko Milekic if (o == NULL) 6266a06dea0SGarrett Wollman goto nospace; 6276a06dea0SGarrett Wollman 6286a06dea0SGarrett Wollman n->m_next = o; 6296a06dea0SGarrett Wollman n = n->m_next; 6306a06dea0SGarrett Wollman 6316a06dea0SGarrett Wollman n->m_len = m->m_len; 63261664ee7SGleb Smirnoff if (m->m_flags & (M_EXT | M_EXTPG)) { 6336a06dea0SGarrett Wollman n->m_data = m->m_data; 63456a4e45aSAndre Oppermann mb_dupcl(n, m); 6356a06dea0SGarrett Wollman } else { 6366a06dea0SGarrett Wollman bcopy(mtod(m, char *), mtod(n, char *), n->m_len); 6376a06dea0SGarrett Wollman } 6386a06dea0SGarrett Wollman 6396a06dea0SGarrett Wollman m = m->m_next; 6406a06dea0SGarrett Wollman } 6416a06dea0SGarrett Wollman return top; 6426a06dea0SGarrett Wollman nospace: 6436a06dea0SGarrett Wollman m_freem(top); 644122a814aSBosko Milekic return (NULL); 6456a06dea0SGarrett Wollman } 6466a06dea0SGarrett Wollman 64782334850SJohn Baldwin static void 64882334850SJohn Baldwin m_copyfromunmapped(const struct mbuf *m, int off, int len, caddr_t cp) 64982334850SJohn Baldwin { 65082334850SJohn Baldwin struct iovec iov; 65182334850SJohn Baldwin struct uio uio; 6523c39566dSMateusz Guzik int error __diagused; 65382334850SJohn Baldwin 65482334850SJohn Baldwin KASSERT(off >= 0, ("m_copyfromunmapped: negative off %d", off)); 65582334850SJohn Baldwin KASSERT(len >= 0, ("m_copyfromunmapped: negative len %d", len)); 65682334850SJohn Baldwin KASSERT(off < m->m_len, 65782334850SJohn Baldwin ("m_copyfromunmapped: len exceeds mbuf length")); 65882334850SJohn Baldwin iov.iov_base = cp; 65982334850SJohn Baldwin iov.iov_len = len; 66082334850SJohn Baldwin uio.uio_resid = len; 66182334850SJohn Baldwin uio.uio_iov = &iov; 66282334850SJohn Baldwin uio.uio_segflg = UIO_SYSSPACE; 66382334850SJohn Baldwin uio.uio_iovcnt = 1; 66482334850SJohn Baldwin uio.uio_offset = 0; 66582334850SJohn Baldwin uio.uio_rw = UIO_READ; 666aa341db3SJohn Baldwin error = m_unmapped_uiomove(m, off, &uio, len); 667aa341db3SJohn Baldwin KASSERT(error == 0, ("m_unmapped_uiomove failed: off %d, len %d", off, 66882334850SJohn Baldwin len)); 66982334850SJohn Baldwin } 67082334850SJohn Baldwin 6716a06dea0SGarrett Wollman /* 672df8bae1dSRodney W. Grimes * Copy data from an mbuf chain starting "off" bytes from the beginning, 673df8bae1dSRodney W. Grimes * continuing for "len" bytes, into the indicated buffer. 674df8bae1dSRodney W. Grimes */ 67526f9a767SRodney W. Grimes void 676a8cfc0eeSJulian Elischer m_copydata(const struct mbuf *m, int off, int len, caddr_t cp) 677df8bae1dSRodney W. Grimes { 678bd395ae8SBosko Milekic u_int count; 679df8bae1dSRodney W. Grimes 680e0a653ddSAlfred Perlstein KASSERT(off >= 0, ("m_copydata, negative off %d", off)); 681e0a653ddSAlfred Perlstein KASSERT(len >= 0, ("m_copydata, negative len %d", len)); 682df8bae1dSRodney W. Grimes while (off > 0) { 683e0a653ddSAlfred Perlstein KASSERT(m != NULL, ("m_copydata, offset > size of mbuf chain")); 684df8bae1dSRodney W. Grimes if (off < m->m_len) 685df8bae1dSRodney W. Grimes break; 686df8bae1dSRodney W. Grimes off -= m->m_len; 687df8bae1dSRodney W. Grimes m = m->m_next; 688df8bae1dSRodney W. Grimes } 689df8bae1dSRodney W. Grimes while (len > 0) { 690e0a653ddSAlfred Perlstein KASSERT(m != NULL, ("m_copydata, length > size of mbuf chain")); 691df8bae1dSRodney W. Grimes count = min(m->m_len - off, len); 6926edfd179SGleb Smirnoff if ((m->m_flags & M_EXTPG) != 0) 69382334850SJohn Baldwin m_copyfromunmapped(m, off, count, cp); 69482334850SJohn Baldwin else 695df8bae1dSRodney W. Grimes bcopy(mtod(m, caddr_t) + off, cp, count); 696df8bae1dSRodney W. Grimes len -= count; 697df8bae1dSRodney W. Grimes cp += count; 698df8bae1dSRodney W. Grimes off = 0; 699df8bae1dSRodney W. Grimes m = m->m_next; 700df8bae1dSRodney W. Grimes } 701df8bae1dSRodney W. Grimes } 702df8bae1dSRodney W. Grimes 703df8bae1dSRodney W. Grimes /* 7041c38f2eaSArchie Cobbs * Copy a packet header mbuf chain into a completely new chain, including 7051c38f2eaSArchie Cobbs * copying any mbuf clusters. Use this instead of m_copypacket() when 7061c38f2eaSArchie Cobbs * you need a writable copy of an mbuf chain. 7071c38f2eaSArchie Cobbs */ 7081c38f2eaSArchie Cobbs struct mbuf * 7090cbefd30SAlexander V. Chernikov m_dup(const struct mbuf *m, int how) 7101c38f2eaSArchie Cobbs { 7111c38f2eaSArchie Cobbs struct mbuf **p, *top = NULL; 7121c38f2eaSArchie Cobbs int remain, moff, nsize; 7131c38f2eaSArchie Cobbs 714063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 7151c38f2eaSArchie Cobbs /* Sanity check */ 7161c38f2eaSArchie Cobbs if (m == NULL) 717122a814aSBosko Milekic return (NULL); 718fe584538SDag-Erling Smørgrav M_ASSERTPKTHDR(m); 7191c38f2eaSArchie Cobbs 7201c38f2eaSArchie Cobbs /* While there's more data, get a new mbuf, tack it on, and fill it */ 7211c38f2eaSArchie Cobbs remain = m->m_pkthdr.len; 7221c38f2eaSArchie Cobbs moff = 0; 7231c38f2eaSArchie Cobbs p = ⊤ 7241c38f2eaSArchie Cobbs while (remain > 0 || top == NULL) { /* allow m->m_pkthdr.len == 0 */ 7251c38f2eaSArchie Cobbs struct mbuf *n; 7261c38f2eaSArchie Cobbs 7271c38f2eaSArchie Cobbs /* Get the next new mbuf */ 728099a0e58SBosko Milekic if (remain >= MINCLSIZE) { 729099a0e58SBosko Milekic n = m_getcl(how, m->m_type, 0); 730099a0e58SBosko Milekic nsize = MCLBYTES; 731099a0e58SBosko Milekic } else { 732099a0e58SBosko Milekic n = m_get(how, m->m_type); 733099a0e58SBosko Milekic nsize = MLEN; 734099a0e58SBosko Milekic } 7351c38f2eaSArchie Cobbs if (n == NULL) 7361c38f2eaSArchie Cobbs goto nospace; 737099a0e58SBosko Milekic 738099a0e58SBosko Milekic if (top == NULL) { /* First one, must be PKTHDR */ 739099a0e58SBosko Milekic if (!m_dup_pkthdr(n, m, how)) { 740099a0e58SBosko Milekic m_free(n); 7411c38f2eaSArchie Cobbs goto nospace; 7421c38f2eaSArchie Cobbs } 74363e6f390SEd Maste if ((n->m_flags & M_EXT) == 0) 744099a0e58SBosko Milekic nsize = MHLEN; 745089bb672SAndrey V. Elsukov n->m_flags &= ~M_RDONLY; 7461c38f2eaSArchie Cobbs } 7471c38f2eaSArchie Cobbs n->m_len = 0; 7481c38f2eaSArchie Cobbs 7491c38f2eaSArchie Cobbs /* Link it into the new chain */ 7501c38f2eaSArchie Cobbs *p = n; 7511c38f2eaSArchie Cobbs p = &n->m_next; 7521c38f2eaSArchie Cobbs 7531c38f2eaSArchie Cobbs /* Copy data from original mbuf(s) into new mbuf */ 7541c38f2eaSArchie Cobbs while (n->m_len < nsize && m != NULL) { 7551c38f2eaSArchie Cobbs int chunk = min(nsize - n->m_len, m->m_len - moff); 7561c38f2eaSArchie Cobbs 757be79f30dSJohn Baldwin m_copydata(m, moff, chunk, n->m_data + n->m_len); 7581c38f2eaSArchie Cobbs moff += chunk; 7591c38f2eaSArchie Cobbs n->m_len += chunk; 7601c38f2eaSArchie Cobbs remain -= chunk; 7611c38f2eaSArchie Cobbs if (moff == m->m_len) { 7621c38f2eaSArchie Cobbs m = m->m_next; 7631c38f2eaSArchie Cobbs moff = 0; 7641c38f2eaSArchie Cobbs } 7651c38f2eaSArchie Cobbs } 7661c38f2eaSArchie Cobbs 7671c38f2eaSArchie Cobbs /* Check correct total mbuf length */ 7681c38f2eaSArchie Cobbs KASSERT((remain > 0 && m != NULL) || (remain == 0 && m == NULL), 769a48740b6SDavid E. O'Brien ("%s: bogus m_pkthdr.len", __func__)); 7701c38f2eaSArchie Cobbs } 7711c38f2eaSArchie Cobbs return (top); 7721c38f2eaSArchie Cobbs 7731c38f2eaSArchie Cobbs nospace: 7741c38f2eaSArchie Cobbs m_freem(top); 775122a814aSBosko Milekic return (NULL); 7761c38f2eaSArchie Cobbs } 7771c38f2eaSArchie Cobbs 7781c38f2eaSArchie Cobbs /* 779df8bae1dSRodney W. Grimes * Concatenate mbuf chain n to m. 780df8bae1dSRodney W. Grimes * Both chains must be of the same type (e.g. MT_DATA). 781df8bae1dSRodney W. Grimes * Any m_pkthdr is not updated. 782df8bae1dSRodney W. Grimes */ 78326f9a767SRodney W. Grimes void 784122a814aSBosko Milekic m_cat(struct mbuf *m, struct mbuf *n) 785df8bae1dSRodney W. Grimes { 786df8bae1dSRodney W. Grimes while (m->m_next) 787df8bae1dSRodney W. Grimes m = m->m_next; 788df8bae1dSRodney W. Grimes while (n) { 78914d7c5b1SAndre Oppermann if (!M_WRITABLE(m) || 7906edfd179SGleb Smirnoff (n->m_flags & M_EXTPG) != 0 || 79114d7c5b1SAndre Oppermann M_TRAILINGSPACE(m) < n->m_len) { 792df8bae1dSRodney W. Grimes /* just join the two chains */ 793df8bae1dSRodney W. Grimes m->m_next = n; 794df8bae1dSRodney W. Grimes return; 795df8bae1dSRodney W. Grimes } 796df8bae1dSRodney W. Grimes /* splat the data from one into the other */ 797df8bae1dSRodney W. Grimes bcopy(mtod(n, caddr_t), mtod(m, caddr_t) + m->m_len, 798df8bae1dSRodney W. Grimes (u_int)n->m_len); 799df8bae1dSRodney W. Grimes m->m_len += n->m_len; 800df8bae1dSRodney W. Grimes n = m_free(n); 801df8bae1dSRodney W. Grimes } 802df8bae1dSRodney W. Grimes } 803df8bae1dSRodney W. Grimes 8041967edbaSGleb Smirnoff /* 8051967edbaSGleb Smirnoff * Concatenate two pkthdr mbuf chains. 8061967edbaSGleb Smirnoff */ 8071967edbaSGleb Smirnoff void 8081967edbaSGleb Smirnoff m_catpkt(struct mbuf *m, struct mbuf *n) 8091967edbaSGleb Smirnoff { 8101967edbaSGleb Smirnoff 8111967edbaSGleb Smirnoff M_ASSERTPKTHDR(m); 8121967edbaSGleb Smirnoff M_ASSERTPKTHDR(n); 8131967edbaSGleb Smirnoff 8141967edbaSGleb Smirnoff m->m_pkthdr.len += n->m_pkthdr.len; 815651e4e6aSGleb Smirnoff m_demote(n, 1, 0); 8161967edbaSGleb Smirnoff 8171967edbaSGleb Smirnoff m_cat(m, n); 8181967edbaSGleb Smirnoff } 8191967edbaSGleb Smirnoff 82026f9a767SRodney W. Grimes void 821122a814aSBosko Milekic m_adj(struct mbuf *mp, int req_len) 822df8bae1dSRodney W. Grimes { 823122a814aSBosko Milekic int len = req_len; 824122a814aSBosko Milekic struct mbuf *m; 825122a814aSBosko Milekic int count; 826df8bae1dSRodney W. Grimes 827df8bae1dSRodney W. Grimes if ((m = mp) == NULL) 828df8bae1dSRodney W. Grimes return; 829df8bae1dSRodney W. Grimes if (len >= 0) { 830df8bae1dSRodney W. Grimes /* 831df8bae1dSRodney W. Grimes * Trim from head. 832df8bae1dSRodney W. Grimes */ 833df8bae1dSRodney W. Grimes while (m != NULL && len > 0) { 834df8bae1dSRodney W. Grimes if (m->m_len <= len) { 835df8bae1dSRodney W. Grimes len -= m->m_len; 836df8bae1dSRodney W. Grimes m->m_len = 0; 837df8bae1dSRodney W. Grimes m = m->m_next; 838df8bae1dSRodney W. Grimes } else { 839df8bae1dSRodney W. Grimes m->m_len -= len; 840df8bae1dSRodney W. Grimes m->m_data += len; 841df8bae1dSRodney W. Grimes len = 0; 842df8bae1dSRodney W. Grimes } 843df8bae1dSRodney W. Grimes } 844df8bae1dSRodney W. Grimes if (mp->m_flags & M_PKTHDR) 845a83baab6SMarko Zec mp->m_pkthdr.len -= (req_len - len); 846df8bae1dSRodney W. Grimes } else { 847df8bae1dSRodney W. Grimes /* 848df8bae1dSRodney W. Grimes * Trim from tail. Scan the mbuf chain, 849df8bae1dSRodney W. Grimes * calculating its length and finding the last mbuf. 850df8bae1dSRodney W. Grimes * If the adjustment only affects this mbuf, then just 851df8bae1dSRodney W. Grimes * adjust and return. Otherwise, rescan and truncate 852df8bae1dSRodney W. Grimes * after the remaining size. 853df8bae1dSRodney W. Grimes */ 854df8bae1dSRodney W. Grimes len = -len; 855df8bae1dSRodney W. Grimes count = 0; 856df8bae1dSRodney W. Grimes for (;;) { 857df8bae1dSRodney W. Grimes count += m->m_len; 858df8bae1dSRodney W. Grimes if (m->m_next == (struct mbuf *)0) 859df8bae1dSRodney W. Grimes break; 860df8bae1dSRodney W. Grimes m = m->m_next; 861df8bae1dSRodney W. Grimes } 862df8bae1dSRodney W. Grimes if (m->m_len >= len) { 863df8bae1dSRodney W. Grimes m->m_len -= len; 864df8bae1dSRodney W. Grimes if (mp->m_flags & M_PKTHDR) 865df8bae1dSRodney W. Grimes mp->m_pkthdr.len -= len; 866df8bae1dSRodney W. Grimes return; 867df8bae1dSRodney W. Grimes } 868df8bae1dSRodney W. Grimes count -= len; 869df8bae1dSRodney W. Grimes if (count < 0) 870df8bae1dSRodney W. Grimes count = 0; 871df8bae1dSRodney W. Grimes /* 872df8bae1dSRodney W. Grimes * Correct length for chain is "count". 873df8bae1dSRodney W. Grimes * Find the mbuf with last data, adjust its length, 874df8bae1dSRodney W. Grimes * and toss data from remaining mbufs on chain. 875df8bae1dSRodney W. Grimes */ 876df8bae1dSRodney W. Grimes m = mp; 877df8bae1dSRodney W. Grimes if (m->m_flags & M_PKTHDR) 878df8bae1dSRodney W. Grimes m->m_pkthdr.len = count; 879df8bae1dSRodney W. Grimes for (; m; m = m->m_next) { 880df8bae1dSRodney W. Grimes if (m->m_len >= count) { 881df8bae1dSRodney W. Grimes m->m_len = count; 88259d8b310SSam Leffler if (m->m_next != NULL) { 88359d8b310SSam Leffler m_freem(m->m_next); 88459d8b310SSam Leffler m->m_next = NULL; 88559d8b310SSam Leffler } 886df8bae1dSRodney W. Grimes break; 887df8bae1dSRodney W. Grimes } 888df8bae1dSRodney W. Grimes count -= m->m_len; 889df8bae1dSRodney W. Grimes } 890df8bae1dSRodney W. Grimes } 891df8bae1dSRodney W. Grimes } 892df8bae1dSRodney W. Grimes 893e243367bSKonstantin Belousov void 894e243367bSKonstantin Belousov m_adj_decap(struct mbuf *mp, int len) 895e243367bSKonstantin Belousov { 896e243367bSKonstantin Belousov uint8_t rsstype; 897e243367bSKonstantin Belousov 898e243367bSKonstantin Belousov m_adj(mp, len); 899e243367bSKonstantin Belousov if ((mp->m_flags & M_PKTHDR) != 0) { 900e243367bSKonstantin Belousov /* 901e243367bSKonstantin Belousov * If flowid was calculated by card from the inner 902e243367bSKonstantin Belousov * headers, move flowid to the decapsulated mbuf 903e243367bSKonstantin Belousov * chain, otherwise clear. This depends on the 904e243367bSKonstantin Belousov * internals of m_adj, which keeps pkthdr as is, in 905e243367bSKonstantin Belousov * particular not changing rsstype and flowid. 906e243367bSKonstantin Belousov */ 907e243367bSKonstantin Belousov rsstype = mp->m_pkthdr.rsstype; 908e243367bSKonstantin Belousov if ((rsstype & M_HASHTYPE_INNER) != 0) { 909e243367bSKonstantin Belousov M_HASHTYPE_SET(mp, rsstype & ~M_HASHTYPE_INNER); 910e243367bSKonstantin Belousov } else { 911e243367bSKonstantin Belousov M_HASHTYPE_CLEAR(mp); 912e243367bSKonstantin Belousov } 913e243367bSKonstantin Belousov } 914e243367bSKonstantin Belousov } 915e243367bSKonstantin Belousov 916df8bae1dSRodney W. Grimes /* 917df8bae1dSRodney W. Grimes * Rearange an mbuf chain so that len bytes are contiguous 918a2c36a02SKevin Lo * and in the data area of an mbuf (so that mtod will work 919a2c36a02SKevin Lo * for a structure of size len). Returns the resulting 920df8bae1dSRodney W. Grimes * mbuf chain on success, frees it and returns null on failure. 921df8bae1dSRodney W. Grimes * If there is room, it will add up to max_protohdr-len extra bytes to the 922df8bae1dSRodney W. Grimes * contiguous region in an attempt to avoid being called next time. 923df8bae1dSRodney W. Grimes */ 924df8bae1dSRodney W. Grimes struct mbuf * 925122a814aSBosko Milekic m_pullup(struct mbuf *n, int len) 926df8bae1dSRodney W. Grimes { 927122a814aSBosko Milekic struct mbuf *m; 928122a814aSBosko Milekic int count; 929df8bae1dSRodney W. Grimes int space; 930df8bae1dSRodney W. Grimes 9316edfd179SGleb Smirnoff KASSERT((n->m_flags & M_EXTPG) == 0, 93282334850SJohn Baldwin ("%s: unmapped mbuf %p", __func__, n)); 93382334850SJohn Baldwin 934df8bae1dSRodney W. Grimes /* 935df8bae1dSRodney W. Grimes * If first mbuf has no cluster, and has room for len bytes 936df8bae1dSRodney W. Grimes * without shifting current data, pullup into it, 937df8bae1dSRodney W. Grimes * otherwise allocate a new mbuf to prepend to the chain. 938df8bae1dSRodney W. Grimes */ 939df8bae1dSRodney W. Grimes if ((n->m_flags & M_EXT) == 0 && 940df8bae1dSRodney W. Grimes n->m_data + len < &n->m_dat[MLEN] && n->m_next) { 941df8bae1dSRodney W. Grimes if (n->m_len >= len) 942df8bae1dSRodney W. Grimes return (n); 943df8bae1dSRodney W. Grimes m = n; 944df8bae1dSRodney W. Grimes n = n->m_next; 945df8bae1dSRodney W. Grimes len -= m->m_len; 946df8bae1dSRodney W. Grimes } else { 947df8bae1dSRodney W. Grimes if (len > MHLEN) 948df8bae1dSRodney W. Grimes goto bad; 949c95be8b5SGleb Smirnoff m = m_get(M_NOWAIT, n->m_type); 950122a814aSBosko Milekic if (m == NULL) 951df8bae1dSRodney W. Grimes goto bad; 9529967cafcSSam Leffler if (n->m_flags & M_PKTHDR) 953c95be8b5SGleb Smirnoff m_move_pkthdr(m, n); 954df8bae1dSRodney W. Grimes } 955df8bae1dSRodney W. Grimes space = &m->m_dat[MLEN] - (m->m_data + m->m_len); 956df8bae1dSRodney W. Grimes do { 957df8bae1dSRodney W. Grimes count = min(min(max(len, max_protohdr), space), n->m_len); 958df8bae1dSRodney W. Grimes bcopy(mtod(n, caddr_t), mtod(m, caddr_t) + m->m_len, 959bd395ae8SBosko Milekic (u_int)count); 960df8bae1dSRodney W. Grimes len -= count; 961df8bae1dSRodney W. Grimes m->m_len += count; 962df8bae1dSRodney W. Grimes n->m_len -= count; 963df8bae1dSRodney W. Grimes space -= count; 964df8bae1dSRodney W. Grimes if (n->m_len) 965df8bae1dSRodney W. Grimes n->m_data += count; 966df8bae1dSRodney W. Grimes else 967df8bae1dSRodney W. Grimes n = m_free(n); 968df8bae1dSRodney W. Grimes } while (len > 0 && n); 969df8bae1dSRodney W. Grimes if (len > 0) { 970df8bae1dSRodney W. Grimes (void) m_free(m); 971df8bae1dSRodney W. Grimes goto bad; 972df8bae1dSRodney W. Grimes } 973df8bae1dSRodney W. Grimes m->m_next = n; 974df8bae1dSRodney W. Grimes return (m); 975df8bae1dSRodney W. Grimes bad: 976df8bae1dSRodney W. Grimes m_freem(n); 977122a814aSBosko Milekic return (NULL); 978df8bae1dSRodney W. Grimes } 979df8bae1dSRodney W. Grimes 980df8bae1dSRodney W. Grimes /* 9817ac139a9SJohn-Mark Gurney * Like m_pullup(), except a new mbuf is always allocated, and we allow 9827ac139a9SJohn-Mark Gurney * the amount of empty space before the data in the new mbuf to be specified 9837ac139a9SJohn-Mark Gurney * (in the event that the caller expects to prepend later). 9847ac139a9SJohn-Mark Gurney */ 9857ac139a9SJohn-Mark Gurney struct mbuf * 9867ac139a9SJohn-Mark Gurney m_copyup(struct mbuf *n, int len, int dstoff) 9877ac139a9SJohn-Mark Gurney { 9887ac139a9SJohn-Mark Gurney struct mbuf *m; 9897ac139a9SJohn-Mark Gurney int count, space; 9907ac139a9SJohn-Mark Gurney 9917ac139a9SJohn-Mark Gurney if (len > (MHLEN - dstoff)) 9927ac139a9SJohn-Mark Gurney goto bad; 993c95be8b5SGleb Smirnoff m = m_get(M_NOWAIT, n->m_type); 9947ac139a9SJohn-Mark Gurney if (m == NULL) 9957ac139a9SJohn-Mark Gurney goto bad; 9967ac139a9SJohn-Mark Gurney if (n->m_flags & M_PKTHDR) 997c95be8b5SGleb Smirnoff m_move_pkthdr(m, n); 9987ac139a9SJohn-Mark Gurney m->m_data += dstoff; 9997ac139a9SJohn-Mark Gurney space = &m->m_dat[MLEN] - (m->m_data + m->m_len); 10007ac139a9SJohn-Mark Gurney do { 10017ac139a9SJohn-Mark Gurney count = min(min(max(len, max_protohdr), space), n->m_len); 10027ac139a9SJohn-Mark Gurney memcpy(mtod(m, caddr_t) + m->m_len, mtod(n, caddr_t), 10037ac139a9SJohn-Mark Gurney (unsigned)count); 10047ac139a9SJohn-Mark Gurney len -= count; 10057ac139a9SJohn-Mark Gurney m->m_len += count; 10067ac139a9SJohn-Mark Gurney n->m_len -= count; 10077ac139a9SJohn-Mark Gurney space -= count; 10087ac139a9SJohn-Mark Gurney if (n->m_len) 10097ac139a9SJohn-Mark Gurney n->m_data += count; 10107ac139a9SJohn-Mark Gurney else 10117ac139a9SJohn-Mark Gurney n = m_free(n); 10127ac139a9SJohn-Mark Gurney } while (len > 0 && n); 10137ac139a9SJohn-Mark Gurney if (len > 0) { 10147ac139a9SJohn-Mark Gurney (void) m_free(m); 10157ac139a9SJohn-Mark Gurney goto bad; 10167ac139a9SJohn-Mark Gurney } 10177ac139a9SJohn-Mark Gurney m->m_next = n; 10187ac139a9SJohn-Mark Gurney return (m); 10197ac139a9SJohn-Mark Gurney bad: 10207ac139a9SJohn-Mark Gurney m_freem(n); 10217ac139a9SJohn-Mark Gurney return (NULL); 10227ac139a9SJohn-Mark Gurney } 10237ac139a9SJohn-Mark Gurney 10247ac139a9SJohn-Mark Gurney /* 1025df8bae1dSRodney W. Grimes * Partition an mbuf chain in two pieces, returning the tail -- 1026df8bae1dSRodney W. Grimes * all but the first len0 bytes. In case of failure, it returns NULL and 1027df8bae1dSRodney W. Grimes * attempts to restore the chain to its original state. 102848d183faSArchie Cobbs * 102948d183faSArchie Cobbs * Note that the resulting mbufs might be read-only, because the new 103048d183faSArchie Cobbs * mbuf can end up sharing an mbuf cluster with the original mbuf if 103148d183faSArchie Cobbs * the "breaking point" happens to lie within a cluster mbuf. Use the 103248d183faSArchie Cobbs * M_WRITABLE() macro to check for this case. 1033df8bae1dSRodney W. Grimes */ 1034df8bae1dSRodney W. Grimes struct mbuf * 1035122a814aSBosko Milekic m_split(struct mbuf *m0, int len0, int wait) 1036df8bae1dSRodney W. Grimes { 1037122a814aSBosko Milekic struct mbuf *m, *n; 1038bd395ae8SBosko Milekic u_int len = len0, remain; 1039df8bae1dSRodney W. Grimes 1040063d8114SAlfred Perlstein MBUF_CHECKSLEEP(wait); 1041df8bae1dSRodney W. Grimes for (m = m0; m && len > m->m_len; m = m->m_next) 1042df8bae1dSRodney W. Grimes len -= m->m_len; 1043122a814aSBosko Milekic if (m == NULL) 1044122a814aSBosko Milekic return (NULL); 1045df8bae1dSRodney W. Grimes remain = m->m_len - len; 104621f39848SGleb Smirnoff if (m0->m_flags & M_PKTHDR && remain == 0) { 104721f39848SGleb Smirnoff n = m_gethdr(wait, m0->m_type); 104877badb18SGleb Smirnoff if (n == NULL) 104921f39848SGleb Smirnoff return (NULL); 105021f39848SGleb Smirnoff n->m_next = m->m_next; 105121f39848SGleb Smirnoff m->m_next = NULL; 1052fb3bc596SJohn Baldwin if (m0->m_pkthdr.csum_flags & CSUM_SND_TAG) { 1053fb3bc596SJohn Baldwin n->m_pkthdr.snd_tag = 1054fb3bc596SJohn Baldwin m_snd_tag_ref(m0->m_pkthdr.snd_tag); 1055fb3bc596SJohn Baldwin n->m_pkthdr.csum_flags |= CSUM_SND_TAG; 1056fb3bc596SJohn Baldwin } else 105721f39848SGleb Smirnoff n->m_pkthdr.rcvif = m0->m_pkthdr.rcvif; 105821f39848SGleb Smirnoff n->m_pkthdr.len = m0->m_pkthdr.len - len0; 105921f39848SGleb Smirnoff m0->m_pkthdr.len = len0; 106021f39848SGleb Smirnoff return (n); 106121f39848SGleb Smirnoff } else if (m0->m_flags & M_PKTHDR) { 1062c95be8b5SGleb Smirnoff n = m_gethdr(wait, m0->m_type); 1063122a814aSBosko Milekic if (n == NULL) 1064122a814aSBosko Milekic return (NULL); 1065fb3bc596SJohn Baldwin if (m0->m_pkthdr.csum_flags & CSUM_SND_TAG) { 1066fb3bc596SJohn Baldwin n->m_pkthdr.snd_tag = 1067fb3bc596SJohn Baldwin m_snd_tag_ref(m0->m_pkthdr.snd_tag); 1068fb3bc596SJohn Baldwin n->m_pkthdr.csum_flags |= CSUM_SND_TAG; 1069fb3bc596SJohn Baldwin } else 1070df8bae1dSRodney W. Grimes n->m_pkthdr.rcvif = m0->m_pkthdr.rcvif; 1071df8bae1dSRodney W. Grimes n->m_pkthdr.len = m0->m_pkthdr.len - len0; 1072df8bae1dSRodney W. Grimes m0->m_pkthdr.len = len0; 107361664ee7SGleb Smirnoff if (m->m_flags & (M_EXT | M_EXTPG)) 1074df8bae1dSRodney W. Grimes goto extpacket; 1075df8bae1dSRodney W. Grimes if (remain > MHLEN) { 1076df8bae1dSRodney W. Grimes /* m can't be the lead packet */ 1077ed6a66caSRobert Watson M_ALIGN(n, 0); 1078df8bae1dSRodney W. Grimes n->m_next = m_split(m, len, wait); 1079122a814aSBosko Milekic if (n->m_next == NULL) { 1080df8bae1dSRodney W. Grimes (void) m_free(n); 1081122a814aSBosko Milekic return (NULL); 108240376987SJeffrey Hsu } else { 108340376987SJeffrey Hsu n->m_len = 0; 1084df8bae1dSRodney W. Grimes return (n); 108540376987SJeffrey Hsu } 1086df8bae1dSRodney W. Grimes } else 1087ed6a66caSRobert Watson M_ALIGN(n, remain); 1088df8bae1dSRodney W. Grimes } else if (remain == 0) { 1089df8bae1dSRodney W. Grimes n = m->m_next; 1090122a814aSBosko Milekic m->m_next = NULL; 1091df8bae1dSRodney W. Grimes return (n); 1092df8bae1dSRodney W. Grimes } else { 1093c95be8b5SGleb Smirnoff n = m_get(wait, m->m_type); 1094122a814aSBosko Milekic if (n == NULL) 1095122a814aSBosko Milekic return (NULL); 1096df8bae1dSRodney W. Grimes M_ALIGN(n, remain); 1097df8bae1dSRodney W. Grimes } 1098df8bae1dSRodney W. Grimes extpacket: 109961664ee7SGleb Smirnoff if (m->m_flags & (M_EXT | M_EXTPG)) { 1100df8bae1dSRodney W. Grimes n->m_data = m->m_data + len; 110156a4e45aSAndre Oppermann mb_dupcl(n, m); 1102df8bae1dSRodney W. Grimes } else { 1103df8bae1dSRodney W. Grimes bcopy(mtod(m, caddr_t) + len, mtod(n, caddr_t), remain); 1104df8bae1dSRodney W. Grimes } 1105df8bae1dSRodney W. Grimes n->m_len = remain; 1106df8bae1dSRodney W. Grimes m->m_len = len; 1107df8bae1dSRodney W. Grimes n->m_next = m->m_next; 1108122a814aSBosko Milekic m->m_next = NULL; 1109df8bae1dSRodney W. Grimes return (n); 1110df8bae1dSRodney W. Grimes } 1111fd01798fSGleb Smirnoff 1112fd01798fSGleb Smirnoff /* 1113fd01798fSGleb Smirnoff * Partition mchain in two pieces, keeping len0 bytes in head and transferring 1114fd01798fSGleb Smirnoff * remainder to tail. In case of failure, both chains to be left untouched. 1115fd01798fSGleb Smirnoff * M_EOR is observed correctly. 1116fd01798fSGleb Smirnoff * Resulting mbufs might be read-only. 1117fd01798fSGleb Smirnoff */ 1118fd01798fSGleb Smirnoff int 1119fd01798fSGleb Smirnoff mc_split(struct mchain *head, struct mchain *tail, u_int len0, int wait) 1120fd01798fSGleb Smirnoff { 1121fd01798fSGleb Smirnoff struct mbuf *m, *n; 1122fd01798fSGleb Smirnoff u_int len, mlen, remain; 1123fd01798fSGleb Smirnoff 1124fd01798fSGleb Smirnoff MPASS(!(mc_first(head)->m_flags & M_PKTHDR)); 1125fd01798fSGleb Smirnoff MBUF_CHECKSLEEP(wait); 1126fd01798fSGleb Smirnoff 1127fd01798fSGleb Smirnoff mlen = 0; 1128fd01798fSGleb Smirnoff len = len0; 1129fd01798fSGleb Smirnoff STAILQ_FOREACH(m, &head->mc_q, m_stailq) { 1130fd01798fSGleb Smirnoff mlen += MSIZE; 1131fd01798fSGleb Smirnoff if (m->m_flags & M_EXT) 1132fd01798fSGleb Smirnoff mlen += m->m_ext.ext_size; 1133fd01798fSGleb Smirnoff if (len > m->m_len) 1134fd01798fSGleb Smirnoff len -= m->m_len; 1135fd01798fSGleb Smirnoff else 1136fd01798fSGleb Smirnoff break; 1137fd01798fSGleb Smirnoff } 1138fd01798fSGleb Smirnoff if (__predict_false(m == NULL)) { 1139fd01798fSGleb Smirnoff *tail = MCHAIN_INITIALIZER(tail); 1140fd01798fSGleb Smirnoff return (0); 1141fd01798fSGleb Smirnoff } 1142fd01798fSGleb Smirnoff remain = m->m_len - len; 1143fd01798fSGleb Smirnoff if (remain > 0) { 1144fd01798fSGleb Smirnoff if (__predict_false((n = m_get(wait, m->m_type)) == NULL)) 1145fd01798fSGleb Smirnoff return (ENOMEM); 1146fd01798fSGleb Smirnoff m_align(n, remain); 1147fd01798fSGleb Smirnoff if (m->m_flags & M_EXT) { 1148fd01798fSGleb Smirnoff n->m_data = m->m_data + len; 1149fd01798fSGleb Smirnoff mb_dupcl(n, m); 1150fd01798fSGleb Smirnoff } else 1151fd01798fSGleb Smirnoff bcopy(mtod(m, char *) + len, mtod(n, char *), remain); 1152fd01798fSGleb Smirnoff } 1153fd01798fSGleb Smirnoff 1154fd01798fSGleb Smirnoff /* XXXGL: need STAILQ_SPLIT */ 1155fd01798fSGleb Smirnoff STAILQ_FIRST(&tail->mc_q) = STAILQ_NEXT(m, m_stailq); 1156fd01798fSGleb Smirnoff tail->mc_q.stqh_last = head->mc_q.stqh_last; 1157fd01798fSGleb Smirnoff tail->mc_len = head->mc_len - len0; 1158fd01798fSGleb Smirnoff tail->mc_mlen = head->mc_mlen - mlen; 1159fd01798fSGleb Smirnoff if (remain > 0) { 1160fd01798fSGleb Smirnoff MPASS(n->m_len == 0); 1161fd01798fSGleb Smirnoff mc_prepend(tail, n); 1162fd01798fSGleb Smirnoff n->m_len = remain; 1163fd01798fSGleb Smirnoff m->m_len -= remain; 1164fd01798fSGleb Smirnoff if (m->m_flags & M_EOR) { 1165fd01798fSGleb Smirnoff m->m_flags &= ~M_EOR; 1166fd01798fSGleb Smirnoff n->m_flags |= M_EOR; 1167fd01798fSGleb Smirnoff } 1168fd01798fSGleb Smirnoff } 1169fd01798fSGleb Smirnoff head->mc_q.stqh_last = &STAILQ_NEXT(m, m_stailq); 1170fd01798fSGleb Smirnoff STAILQ_NEXT(m, m_stailq) = NULL; 1171fd01798fSGleb Smirnoff head->mc_len = len0; 1172fd01798fSGleb Smirnoff head->mc_mlen = mlen; 1173fd01798fSGleb Smirnoff 1174fd01798fSGleb Smirnoff return (0); 1175fd01798fSGleb Smirnoff } 1176fd01798fSGleb Smirnoff 1177df8bae1dSRodney W. Grimes /* 1178df8bae1dSRodney W. Grimes * Routine to copy from device local memory into mbufs. 1179f5eece3fSBosko Milekic * Note that `off' argument is offset into first mbuf of target chain from 1180f5eece3fSBosko Milekic * which to begin copying the data to. 1181df8bae1dSRodney W. Grimes */ 1182df8bae1dSRodney W. Grimes struct mbuf * 1183f5eece3fSBosko Milekic m_devget(char *buf, int totlen, int off, struct ifnet *ifp, 1184122a814aSBosko Milekic void (*copy)(char *from, caddr_t to, u_int len)) 1185df8bae1dSRodney W. Grimes { 1186122a814aSBosko Milekic struct mbuf *m; 1187099a0e58SBosko Milekic struct mbuf *top = NULL, **mp = ⊤ 1188f5eece3fSBosko Milekic int len; 1189df8bae1dSRodney W. Grimes 1190f5eece3fSBosko Milekic if (off < 0 || off > MHLEN) 1191f5eece3fSBosko Milekic return (NULL); 1192f5eece3fSBosko Milekic 1193df8bae1dSRodney W. Grimes while (totlen > 0) { 1194099a0e58SBosko Milekic if (top == NULL) { /* First one, must be PKTHDR */ 1195f5eece3fSBosko Milekic if (totlen + off >= MINCLSIZE) { 1196eb1b1807SGleb Smirnoff m = m_getcl(M_NOWAIT, MT_DATA, M_PKTHDR); 1197f5eece3fSBosko Milekic len = MCLBYTES; 1198df8bae1dSRodney W. Grimes } else { 1199eb1b1807SGleb Smirnoff m = m_gethdr(M_NOWAIT, MT_DATA); 1200099a0e58SBosko Milekic len = MHLEN; 1201099a0e58SBosko Milekic 1202099a0e58SBosko Milekic /* Place initial small packet/header at end of mbuf */ 1203cee4a056SKevin Lo if (m && totlen + off + max_linkhdr <= MHLEN) { 1204df8bae1dSRodney W. Grimes m->m_data += max_linkhdr; 1205f5eece3fSBosko Milekic len -= max_linkhdr; 1206df8bae1dSRodney W. Grimes } 1207f5eece3fSBosko Milekic } 1208099a0e58SBosko Milekic if (m == NULL) 1209099a0e58SBosko Milekic return NULL; 1210099a0e58SBosko Milekic m->m_pkthdr.rcvif = ifp; 1211099a0e58SBosko Milekic m->m_pkthdr.len = totlen; 1212099a0e58SBosko Milekic } else { 1213099a0e58SBosko Milekic if (totlen + off >= MINCLSIZE) { 1214eb1b1807SGleb Smirnoff m = m_getcl(M_NOWAIT, MT_DATA, 0); 1215099a0e58SBosko Milekic len = MCLBYTES; 1216099a0e58SBosko Milekic } else { 1217eb1b1807SGleb Smirnoff m = m_get(M_NOWAIT, MT_DATA); 1218099a0e58SBosko Milekic len = MLEN; 1219099a0e58SBosko Milekic } 1220099a0e58SBosko Milekic if (m == NULL) { 1221099a0e58SBosko Milekic m_freem(top); 1222099a0e58SBosko Milekic return NULL; 1223099a0e58SBosko Milekic } 1224099a0e58SBosko Milekic } 1225f5eece3fSBosko Milekic if (off) { 1226f5eece3fSBosko Milekic m->m_data += off; 1227f5eece3fSBosko Milekic len -= off; 1228f5eece3fSBosko Milekic off = 0; 1229f5eece3fSBosko Milekic } 1230f5eece3fSBosko Milekic m->m_len = len = min(totlen, len); 1231df8bae1dSRodney W. Grimes if (copy) 1232bd395ae8SBosko Milekic copy(buf, mtod(m, caddr_t), (u_int)len); 1233df8bae1dSRodney W. Grimes else 1234bd395ae8SBosko Milekic bcopy(buf, mtod(m, caddr_t), (u_int)len); 1235f5eece3fSBosko Milekic buf += len; 1236df8bae1dSRodney W. Grimes *mp = m; 1237df8bae1dSRodney W. Grimes mp = &m->m_next; 1238df8bae1dSRodney W. Grimes totlen -= len; 1239df8bae1dSRodney W. Grimes } 1240df8bae1dSRodney W. Grimes return (top); 1241df8bae1dSRodney W. Grimes } 1242c5789ba3SPoul-Henning Kamp 12433f9dac85SJohn Baldwin static void 12443f9dac85SJohn Baldwin m_copytounmapped(const struct mbuf *m, int off, int len, c_caddr_t cp) 12453f9dac85SJohn Baldwin { 12463f9dac85SJohn Baldwin struct iovec iov; 12473f9dac85SJohn Baldwin struct uio uio; 12483c39566dSMateusz Guzik int error __diagused; 12493f9dac85SJohn Baldwin 12503f9dac85SJohn Baldwin KASSERT(off >= 0, ("m_copytounmapped: negative off %d", off)); 12513f9dac85SJohn Baldwin KASSERT(len >= 0, ("m_copytounmapped: negative len %d", len)); 12523f9dac85SJohn Baldwin KASSERT(off < m->m_len, ("m_copytounmapped: len exceeds mbuf length")); 12533f9dac85SJohn Baldwin iov.iov_base = __DECONST(caddr_t, cp); 12543f9dac85SJohn Baldwin iov.iov_len = len; 12553f9dac85SJohn Baldwin uio.uio_resid = len; 12563f9dac85SJohn Baldwin uio.uio_iov = &iov; 12573f9dac85SJohn Baldwin uio.uio_segflg = UIO_SYSSPACE; 12583f9dac85SJohn Baldwin uio.uio_iovcnt = 1; 12593f9dac85SJohn Baldwin uio.uio_offset = 0; 12603f9dac85SJohn Baldwin uio.uio_rw = UIO_WRITE; 1261aa341db3SJohn Baldwin error = m_unmapped_uiomove(m, off, &uio, len); 1262aa341db3SJohn Baldwin KASSERT(error == 0, ("m_unmapped_uiomove failed: off %d, len %d", off, 12633f9dac85SJohn Baldwin len)); 12643f9dac85SJohn Baldwin } 12653f9dac85SJohn Baldwin 1266c5789ba3SPoul-Henning Kamp /* 1267c5789ba3SPoul-Henning Kamp * Copy data from a buffer back into the indicated mbuf chain, 1268c5789ba3SPoul-Henning Kamp * starting "off" bytes from the beginning, extending the mbuf 1269c5789ba3SPoul-Henning Kamp * chain if necessary. 1270c5789ba3SPoul-Henning Kamp */ 1271c5789ba3SPoul-Henning Kamp void 127224665342SLuigi Rizzo m_copyback(struct mbuf *m0, int off, int len, c_caddr_t cp) 1273c5789ba3SPoul-Henning Kamp { 1274122a814aSBosko Milekic int mlen; 1275122a814aSBosko Milekic struct mbuf *m = m0, *n; 1276c5789ba3SPoul-Henning Kamp int totlen = 0; 1277c5789ba3SPoul-Henning Kamp 1278122a814aSBosko Milekic if (m0 == NULL) 1279c5789ba3SPoul-Henning Kamp return; 1280c5789ba3SPoul-Henning Kamp while (off > (mlen = m->m_len)) { 1281c5789ba3SPoul-Henning Kamp off -= mlen; 1282c5789ba3SPoul-Henning Kamp totlen += mlen; 1283122a814aSBosko Milekic if (m->m_next == NULL) { 1284eb1b1807SGleb Smirnoff n = m_get(M_NOWAIT, m->m_type); 1285122a814aSBosko Milekic if (n == NULL) 1286c5789ba3SPoul-Henning Kamp goto out; 1287099a0e58SBosko Milekic bzero(mtod(n, caddr_t), MLEN); 1288c5789ba3SPoul-Henning Kamp n->m_len = min(MLEN, len + off); 1289c5789ba3SPoul-Henning Kamp m->m_next = n; 1290c5789ba3SPoul-Henning Kamp } 1291c5789ba3SPoul-Henning Kamp m = m->m_next; 1292c5789ba3SPoul-Henning Kamp } 1293c5789ba3SPoul-Henning Kamp while (len > 0) { 1294129c5c81SAlexander Motin if (m->m_next == NULL && (len > m->m_len - off)) { 1295129c5c81SAlexander Motin m->m_len += min(len - (m->m_len - off), 1296129c5c81SAlexander Motin M_TRAILINGSPACE(m)); 1297129c5c81SAlexander Motin } 1298c5789ba3SPoul-Henning Kamp mlen = min (m->m_len - off, len); 12993f9dac85SJohn Baldwin if ((m->m_flags & M_EXTPG) != 0) 13003f9dac85SJohn Baldwin m_copytounmapped(m, off, mlen, cp); 13013f9dac85SJohn Baldwin else 1302bd395ae8SBosko Milekic bcopy(cp, off + mtod(m, caddr_t), (u_int)mlen); 1303c5789ba3SPoul-Henning Kamp cp += mlen; 1304c5789ba3SPoul-Henning Kamp len -= mlen; 1305c5789ba3SPoul-Henning Kamp mlen += off; 1306c5789ba3SPoul-Henning Kamp off = 0; 1307c5789ba3SPoul-Henning Kamp totlen += mlen; 1308c5789ba3SPoul-Henning Kamp if (len == 0) 1309c5789ba3SPoul-Henning Kamp break; 1310122a814aSBosko Milekic if (m->m_next == NULL) { 1311eb1b1807SGleb Smirnoff n = m_get(M_NOWAIT, m->m_type); 1312122a814aSBosko Milekic if (n == NULL) 1313c5789ba3SPoul-Henning Kamp break; 1314c5789ba3SPoul-Henning Kamp n->m_len = min(MLEN, len); 1315c5789ba3SPoul-Henning Kamp m->m_next = n; 1316c5789ba3SPoul-Henning Kamp } 1317c5789ba3SPoul-Henning Kamp m = m->m_next; 1318c5789ba3SPoul-Henning Kamp } 1319c5789ba3SPoul-Henning Kamp out: if (((m = m0)->m_flags & M_PKTHDR) && (m->m_pkthdr.len < totlen)) 1320c5789ba3SPoul-Henning Kamp m->m_pkthdr.len = totlen; 1321c5789ba3SPoul-Henning Kamp } 1322ce4a64f7SPoul-Henning Kamp 132337621fd5SBruce M Simpson /* 13244873d175SSam Leffler * Append the specified data to the indicated mbuf chain, 13254873d175SSam Leffler * Extend the mbuf chain if the new data does not fit in 13264873d175SSam Leffler * existing space. 13274873d175SSam Leffler * 13284873d175SSam Leffler * Return 1 if able to complete the job; otherwise 0. 13294873d175SSam Leffler */ 13304873d175SSam Leffler int 13314873d175SSam Leffler m_append(struct mbuf *m0, int len, c_caddr_t cp) 13324873d175SSam Leffler { 13334873d175SSam Leffler struct mbuf *m, *n; 13344873d175SSam Leffler int remainder, space; 13354873d175SSam Leffler 13364873d175SSam Leffler for (m = m0; m->m_next != NULL; m = m->m_next) 13374873d175SSam Leffler ; 13384873d175SSam Leffler remainder = len; 13394873d175SSam Leffler space = M_TRAILINGSPACE(m); 13404873d175SSam Leffler if (space > 0) { 13414873d175SSam Leffler /* 13424873d175SSam Leffler * Copy into available space. 13434873d175SSam Leffler */ 13444873d175SSam Leffler if (space > remainder) 13454873d175SSam Leffler space = remainder; 13464873d175SSam Leffler bcopy(cp, mtod(m, caddr_t) + m->m_len, space); 13474873d175SSam Leffler m->m_len += space; 13484873d175SSam Leffler cp += space, remainder -= space; 13494873d175SSam Leffler } 13504873d175SSam Leffler while (remainder > 0) { 13514873d175SSam Leffler /* 13524873d175SSam Leffler * Allocate a new mbuf; could check space 13534873d175SSam Leffler * and allocate a cluster instead. 13544873d175SSam Leffler */ 1355eb1b1807SGleb Smirnoff n = m_get(M_NOWAIT, m->m_type); 13564873d175SSam Leffler if (n == NULL) 13574873d175SSam Leffler break; 13584873d175SSam Leffler n->m_len = min(MLEN, remainder); 1359a37c415eSSam Leffler bcopy(cp, mtod(n, caddr_t), n->m_len); 1360a37c415eSSam Leffler cp += n->m_len, remainder -= n->m_len; 13614873d175SSam Leffler m->m_next = n; 13624873d175SSam Leffler m = n; 13634873d175SSam Leffler } 13644873d175SSam Leffler if (m0->m_flags & M_PKTHDR) 13654873d175SSam Leffler m0->m_pkthdr.len += len - remainder; 13664873d175SSam Leffler return (remainder == 0); 13674873d175SSam Leffler } 13684873d175SSam Leffler 13693c7a01d7SJohn Baldwin static int 13703c7a01d7SJohn Baldwin m_apply_extpg_one(struct mbuf *m, int off, int len, 13713c7a01d7SJohn Baldwin int (*f)(void *, void *, u_int), void *arg) 13723c7a01d7SJohn Baldwin { 13733c7a01d7SJohn Baldwin void *p; 13743c7a01d7SJohn Baldwin u_int i, count, pgoff, pglen; 13753c7a01d7SJohn Baldwin int rval; 13763c7a01d7SJohn Baldwin 13773c7a01d7SJohn Baldwin KASSERT(PMAP_HAS_DMAP, 13783c7a01d7SJohn Baldwin ("m_apply_extpg_one does not support unmapped mbufs")); 13793c7a01d7SJohn Baldwin off += mtod(m, vm_offset_t); 13803c7a01d7SJohn Baldwin if (off < m->m_epg_hdrlen) { 13813c7a01d7SJohn Baldwin count = min(m->m_epg_hdrlen - off, len); 13823c7a01d7SJohn Baldwin rval = f(arg, m->m_epg_hdr + off, count); 13833c7a01d7SJohn Baldwin if (rval) 13843c7a01d7SJohn Baldwin return (rval); 13853c7a01d7SJohn Baldwin len -= count; 13863c7a01d7SJohn Baldwin off = 0; 13873c7a01d7SJohn Baldwin } else 13883c7a01d7SJohn Baldwin off -= m->m_epg_hdrlen; 13893c7a01d7SJohn Baldwin pgoff = m->m_epg_1st_off; 13903c7a01d7SJohn Baldwin for (i = 0; i < m->m_epg_npgs && len > 0; i++) { 13913c7a01d7SJohn Baldwin pglen = m_epg_pagelen(m, i, pgoff); 13923c7a01d7SJohn Baldwin if (off < pglen) { 13933c7a01d7SJohn Baldwin count = min(pglen - off, len); 1394ac2b5444SMark Johnston p = (void *)PHYS_TO_DMAP(m->m_epg_pa[i] + pgoff + off); 13953c7a01d7SJohn Baldwin rval = f(arg, p, count); 13963c7a01d7SJohn Baldwin if (rval) 13973c7a01d7SJohn Baldwin return (rval); 13983c7a01d7SJohn Baldwin len -= count; 13993c7a01d7SJohn Baldwin off = 0; 14003c7a01d7SJohn Baldwin } else 14013c7a01d7SJohn Baldwin off -= pglen; 14023c7a01d7SJohn Baldwin pgoff = 0; 14033c7a01d7SJohn Baldwin } 14043c7a01d7SJohn Baldwin if (len > 0) { 14053c7a01d7SJohn Baldwin KASSERT(off < m->m_epg_trllen, 14063c7a01d7SJohn Baldwin ("m_apply_extpg_one: offset beyond trailer")); 14073c7a01d7SJohn Baldwin KASSERT(len <= m->m_epg_trllen - off, 14083c7a01d7SJohn Baldwin ("m_apply_extpg_one: length beyond trailer")); 14093c7a01d7SJohn Baldwin return (f(arg, m->m_epg_trail + off, len)); 14103c7a01d7SJohn Baldwin } 14113c7a01d7SJohn Baldwin return (0); 14123c7a01d7SJohn Baldwin } 14133c7a01d7SJohn Baldwin 14143c7a01d7SJohn Baldwin /* Apply function f to the data in a single mbuf. */ 14153c7a01d7SJohn Baldwin static int 14163c7a01d7SJohn Baldwin m_apply_one(struct mbuf *m, int off, int len, 14173c7a01d7SJohn Baldwin int (*f)(void *, void *, u_int), void *arg) 14183c7a01d7SJohn Baldwin { 14193c7a01d7SJohn Baldwin if ((m->m_flags & M_EXTPG) != 0) 14203c7a01d7SJohn Baldwin return (m_apply_extpg_one(m, off, len, f, arg)); 14213c7a01d7SJohn Baldwin else 14223c7a01d7SJohn Baldwin return (f(arg, mtod(m, caddr_t) + off, len)); 14233c7a01d7SJohn Baldwin } 14243c7a01d7SJohn Baldwin 14254873d175SSam Leffler /* 142637621fd5SBruce M Simpson * Apply function f to the data in an mbuf chain starting "off" bytes from 142737621fd5SBruce M Simpson * the beginning, continuing for "len" bytes. 142837621fd5SBruce M Simpson */ 142937621fd5SBruce M Simpson int 143037621fd5SBruce M Simpson m_apply(struct mbuf *m, int off, int len, 143154065297SBruce M Simpson int (*f)(void *, void *, u_int), void *arg) 143237621fd5SBruce M Simpson { 143354065297SBruce M Simpson u_int count; 143437621fd5SBruce M Simpson int rval; 143537621fd5SBruce M Simpson 143637621fd5SBruce M Simpson KASSERT(off >= 0, ("m_apply, negative off %d", off)); 143737621fd5SBruce M Simpson KASSERT(len >= 0, ("m_apply, negative len %d", len)); 143837621fd5SBruce M Simpson while (off > 0) { 143937621fd5SBruce M Simpson KASSERT(m != NULL, ("m_apply, offset > size of mbuf chain")); 144037621fd5SBruce M Simpson if (off < m->m_len) 144137621fd5SBruce M Simpson break; 144237621fd5SBruce M Simpson off -= m->m_len; 144337621fd5SBruce M Simpson m = m->m_next; 144437621fd5SBruce M Simpson } 144537621fd5SBruce M Simpson while (len > 0) { 144637621fd5SBruce M Simpson KASSERT(m != NULL, ("m_apply, offset > size of mbuf chain")); 144737621fd5SBruce M Simpson count = min(m->m_len - off, len); 14483c7a01d7SJohn Baldwin rval = m_apply_one(m, off, count, f, arg); 144937621fd5SBruce M Simpson if (rval) 145037621fd5SBruce M Simpson return (rval); 145137621fd5SBruce M Simpson len -= count; 145237621fd5SBruce M Simpson off = 0; 145337621fd5SBruce M Simpson m = m->m_next; 145437621fd5SBruce M Simpson } 145537621fd5SBruce M Simpson return (0); 145637621fd5SBruce M Simpson } 145737621fd5SBruce M Simpson 145837621fd5SBruce M Simpson /* 145937621fd5SBruce M Simpson * Return a pointer to mbuf/offset of location in mbuf chain. 146037621fd5SBruce M Simpson */ 146137621fd5SBruce M Simpson struct mbuf * 146237621fd5SBruce M Simpson m_getptr(struct mbuf *m, int loc, int *off) 146337621fd5SBruce M Simpson { 146437621fd5SBruce M Simpson 146537621fd5SBruce M Simpson while (loc >= 0) { 146654065297SBruce M Simpson /* Normal end of search. */ 146737621fd5SBruce M Simpson if (m->m_len > loc) { 146837621fd5SBruce M Simpson *off = loc; 146937621fd5SBruce M Simpson return (m); 147037621fd5SBruce M Simpson } else { 147137621fd5SBruce M Simpson loc -= m->m_len; 147237621fd5SBruce M Simpson if (m->m_next == NULL) { 147337621fd5SBruce M Simpson if (loc == 0) { 147454065297SBruce M Simpson /* Point at the end of valid data. */ 147537621fd5SBruce M Simpson *off = m->m_len; 147637621fd5SBruce M Simpson return (m); 147754065297SBruce M Simpson } 147837621fd5SBruce M Simpson return (NULL); 147954065297SBruce M Simpson } 148037621fd5SBruce M Simpson m = m->m_next; 148137621fd5SBruce M Simpson } 148237621fd5SBruce M Simpson } 148337621fd5SBruce M Simpson return (NULL); 148437621fd5SBruce M Simpson } 148537621fd5SBruce M Simpson 1486ce4a64f7SPoul-Henning Kamp void 14877b125090SJohn-Mark Gurney m_print(const struct mbuf *m, int maxlen) 1488ce4a64f7SPoul-Henning Kamp { 1489ce4a64f7SPoul-Henning Kamp int len; 14907b125090SJohn-Mark Gurney int pdata; 14916357e7b5SEivind Eklund const struct mbuf *m2; 1492ce4a64f7SPoul-Henning Kamp 14937e949c46SKenneth D. Merry if (m == NULL) { 14947e949c46SKenneth D. Merry printf("mbuf: %p\n", m); 14957e949c46SKenneth D. Merry return; 14967e949c46SKenneth D. Merry } 14977e949c46SKenneth D. Merry 14987b125090SJohn-Mark Gurney if (m->m_flags & M_PKTHDR) 1499ce4a64f7SPoul-Henning Kamp len = m->m_pkthdr.len; 15007b125090SJohn-Mark Gurney else 15017b125090SJohn-Mark Gurney len = -1; 1502ce4a64f7SPoul-Henning Kamp m2 = m; 15037b125090SJohn-Mark Gurney while (m2 != NULL && (len == -1 || len)) { 15047b125090SJohn-Mark Gurney pdata = m2->m_len; 15057b125090SJohn-Mark Gurney if (maxlen != -1 && pdata > maxlen) 15067b125090SJohn-Mark Gurney pdata = maxlen; 15077b125090SJohn-Mark Gurney printf("mbuf: %p len: %d, next: %p, %b%s", m2, m2->m_len, 15087b125090SJohn-Mark Gurney m2->m_next, m2->m_flags, "\20\20freelist\17skipfw" 15097b125090SJohn-Mark Gurney "\11proto5\10proto4\7proto3\6proto2\5proto1\4rdonly" 15107b125090SJohn-Mark Gurney "\3eor\2pkthdr\1ext", pdata ? "" : "\n"); 15117b125090SJohn-Mark Gurney if (pdata) 151245e0d0aaSJohn-Mark Gurney printf(", %*D\n", pdata, (u_char *)m2->m_data, "-"); 15137b125090SJohn-Mark Gurney if (len != -1) 1514ce4a64f7SPoul-Henning Kamp len -= m2->m_len; 1515ce4a64f7SPoul-Henning Kamp m2 = m2->m_next; 1516ce4a64f7SPoul-Henning Kamp } 15177b125090SJohn-Mark Gurney if (len > 0) 15187b125090SJohn-Mark Gurney printf("%d bytes unaccounted for.\n", len); 1519ce4a64f7SPoul-Henning Kamp return; 1520ce4a64f7SPoul-Henning Kamp } 15213f2e06c5SPoul-Henning Kamp 1522bd395ae8SBosko Milekic u_int 15233f2e06c5SPoul-Henning Kamp m_fixhdr(struct mbuf *m0) 15243f2e06c5SPoul-Henning Kamp { 1525bd395ae8SBosko Milekic u_int len; 15263f2e06c5SPoul-Henning Kamp 1527ac6e585dSPoul-Henning Kamp len = m_length(m0, NULL); 15283f2e06c5SPoul-Henning Kamp m0->m_pkthdr.len = len; 1529ac6e585dSPoul-Henning Kamp return (len); 1530ac6e585dSPoul-Henning Kamp } 1531ac6e585dSPoul-Henning Kamp 1532bd395ae8SBosko Milekic u_int 1533ac6e585dSPoul-Henning Kamp m_length(struct mbuf *m0, struct mbuf **last) 1534ac6e585dSPoul-Henning Kamp { 1535ac6e585dSPoul-Henning Kamp struct mbuf *m; 1536bd395ae8SBosko Milekic u_int len; 1537ac6e585dSPoul-Henning Kamp 1538ac6e585dSPoul-Henning Kamp len = 0; 1539ac6e585dSPoul-Henning Kamp for (m = m0; m != NULL; m = m->m_next) { 1540ac6e585dSPoul-Henning Kamp len += m->m_len; 1541ac6e585dSPoul-Henning Kamp if (m->m_next == NULL) 1542ac6e585dSPoul-Henning Kamp break; 1543ac6e585dSPoul-Henning Kamp } 1544ac6e585dSPoul-Henning Kamp if (last != NULL) 1545ac6e585dSPoul-Henning Kamp *last = m; 1546ac6e585dSPoul-Henning Kamp return (len); 15473f2e06c5SPoul-Henning Kamp } 154855e9f80dSMike Silbersack 154955e9f80dSMike Silbersack /* 155055e9f80dSMike Silbersack * Defragment a mbuf chain, returning the shortest possible 155155e9f80dSMike Silbersack * chain of mbufs and clusters. If allocation fails and 155255e9f80dSMike Silbersack * this cannot be completed, NULL will be returned, but 155355e9f80dSMike Silbersack * the passed in chain will be unchanged. Upon success, 155455e9f80dSMike Silbersack * the original chain will be freed, and the new chain 155555e9f80dSMike Silbersack * will be returned. 155655e9f80dSMike Silbersack * 155755e9f80dSMike Silbersack * If a non-packet header is passed in, the original 155855e9f80dSMike Silbersack * mbuf (chain?) will be returned unharmed. 155955e9f80dSMike Silbersack */ 156055e9f80dSMike Silbersack struct mbuf * 156155e9f80dSMike Silbersack m_defrag(struct mbuf *m0, int how) 156255e9f80dSMike Silbersack { 156355e9f80dSMike Silbersack struct mbuf *m_new = NULL, *m_final = NULL; 156455e9f80dSMike Silbersack int progress = 0, length; 156555e9f80dSMike Silbersack 1566063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 156755e9f80dSMike Silbersack if (!(m0->m_flags & M_PKTHDR)) 156855e9f80dSMike Silbersack return (m0); 156955e9f80dSMike Silbersack 1570f8bf8e39SMike Silbersack m_fixhdr(m0); /* Needed sanity check */ 1571f8bf8e39SMike Silbersack 1572352d050eSMike Silbersack #ifdef MBUF_STRESS_TEST 1573352d050eSMike Silbersack if (m_defragrandomfailures) { 1574352d050eSMike Silbersack int temp = arc4random() & 0xff; 1575352d050eSMike Silbersack if (temp == 0xba) 1576352d050eSMike Silbersack goto nospace; 1577352d050eSMike Silbersack } 1578352d050eSMike Silbersack #endif 157955e9f80dSMike Silbersack 158055e9f80dSMike Silbersack if (m0->m_pkthdr.len > MHLEN) 158155e9f80dSMike Silbersack m_final = m_getcl(how, MT_DATA, M_PKTHDR); 158255e9f80dSMike Silbersack else 158355e9f80dSMike Silbersack m_final = m_gethdr(how, MT_DATA); 158455e9f80dSMike Silbersack 158555e9f80dSMike Silbersack if (m_final == NULL) 158655e9f80dSMike Silbersack goto nospace; 158755e9f80dSMike Silbersack 1588a89ec05eSPeter Wemm if (m_dup_pkthdr(m_final, m0, how) == 0) 158955e9f80dSMike Silbersack goto nospace; 159055e9f80dSMike Silbersack 159155e9f80dSMike Silbersack m_new = m_final; 159255e9f80dSMike Silbersack 159355e9f80dSMike Silbersack while (progress < m0->m_pkthdr.len) { 159455e9f80dSMike Silbersack length = m0->m_pkthdr.len - progress; 159555e9f80dSMike Silbersack if (length > MCLBYTES) 159655e9f80dSMike Silbersack length = MCLBYTES; 159755e9f80dSMike Silbersack 159855e9f80dSMike Silbersack if (m_new == NULL) { 159955e9f80dSMike Silbersack if (length > MLEN) 160055e9f80dSMike Silbersack m_new = m_getcl(how, MT_DATA, 0); 160155e9f80dSMike Silbersack else 160255e9f80dSMike Silbersack m_new = m_get(how, MT_DATA); 160355e9f80dSMike Silbersack if (m_new == NULL) 160455e9f80dSMike Silbersack goto nospace; 160555e9f80dSMike Silbersack } 160655e9f80dSMike Silbersack 160755e9f80dSMike Silbersack m_copydata(m0, progress, length, mtod(m_new, caddr_t)); 160855e9f80dSMike Silbersack progress += length; 160955e9f80dSMike Silbersack m_new->m_len = length; 161055e9f80dSMike Silbersack if (m_new != m_final) 161155e9f80dSMike Silbersack m_cat(m_final, m_new); 161255e9f80dSMike Silbersack m_new = NULL; 161355e9f80dSMike Silbersack } 161451710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 161555e9f80dSMike Silbersack if (m0->m_next == NULL) 161655e9f80dSMike Silbersack m_defraguseless++; 161751710a45SMike Silbersack #endif 161855e9f80dSMike Silbersack m_freem(m0); 161955e9f80dSMike Silbersack m0 = m_final; 162051710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 162155e9f80dSMike Silbersack m_defragpackets++; 162255e9f80dSMike Silbersack m_defragbytes += m0->m_pkthdr.len; 162351710a45SMike Silbersack #endif 162455e9f80dSMike Silbersack return (m0); 162555e9f80dSMike Silbersack nospace: 162651710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 162755e9f80dSMike Silbersack m_defragfailure++; 162851710a45SMike Silbersack #endif 162955e9f80dSMike Silbersack if (m_final) 163055e9f80dSMike Silbersack m_freem(m_final); 163155e9f80dSMike Silbersack return (NULL); 163255e9f80dSMike Silbersack } 16333390d476SMike Silbersack 1634eeb76a18SSam Leffler /* 163582334850SJohn Baldwin * Return the number of fragments an mbuf will use. This is usually 163682334850SJohn Baldwin * used as a proxy for the number of scatter/gather elements needed by 163782334850SJohn Baldwin * a DMA engine to access an mbuf. In general mapped mbufs are 163882334850SJohn Baldwin * assumed to be backed by physically contiguous buffers that only 163982334850SJohn Baldwin * need a single fragment. Unmapped mbufs, on the other hand, can 164082334850SJohn Baldwin * span disjoint physical pages. 164182334850SJohn Baldwin */ 164282334850SJohn Baldwin static int 164382334850SJohn Baldwin frags_per_mbuf(struct mbuf *m) 164482334850SJohn Baldwin { 164582334850SJohn Baldwin int frags; 164682334850SJohn Baldwin 16476edfd179SGleb Smirnoff if ((m->m_flags & M_EXTPG) == 0) 164882334850SJohn Baldwin return (1); 164982334850SJohn Baldwin 165082334850SJohn Baldwin /* 165182334850SJohn Baldwin * The header and trailer are counted as a single fragment 165282334850SJohn Baldwin * each when present. 165382334850SJohn Baldwin * 165482334850SJohn Baldwin * XXX: This overestimates the number of fragments by assuming 165582334850SJohn Baldwin * all the backing physical pages are disjoint. 165682334850SJohn Baldwin */ 165782334850SJohn Baldwin frags = 0; 16587b6c99d0SGleb Smirnoff if (m->m_epg_hdrlen != 0) 165982334850SJohn Baldwin frags++; 16607b6c99d0SGleb Smirnoff frags += m->m_epg_npgs; 16617b6c99d0SGleb Smirnoff if (m->m_epg_trllen != 0) 166282334850SJohn Baldwin frags++; 166382334850SJohn Baldwin 166482334850SJohn Baldwin return (frags); 166582334850SJohn Baldwin } 166682334850SJohn Baldwin 166782334850SJohn Baldwin /* 1668eeb76a18SSam Leffler * Defragment an mbuf chain, returning at most maxfrags separate 1669eeb76a18SSam Leffler * mbufs+clusters. If this is not possible NULL is returned and 167028323addSBryan Drewery * the original mbuf chain is left in its present (potentially 1671eeb76a18SSam Leffler * modified) state. We use two techniques: collapsing consecutive 1672eeb76a18SSam Leffler * mbufs and replacing consecutive mbufs by a cluster. 1673eeb76a18SSam Leffler * 1674eeb76a18SSam Leffler * NB: this should really be named m_defrag but that name is taken 1675eeb76a18SSam Leffler */ 1676eeb76a18SSam Leffler struct mbuf * 1677eeb76a18SSam Leffler m_collapse(struct mbuf *m0, int how, int maxfrags) 1678eeb76a18SSam Leffler { 1679eeb76a18SSam Leffler struct mbuf *m, *n, *n2, **prev; 1680eeb76a18SSam Leffler u_int curfrags; 1681eeb76a18SSam Leffler 1682eeb76a18SSam Leffler /* 1683eeb76a18SSam Leffler * Calculate the current number of frags. 1684eeb76a18SSam Leffler */ 1685eeb76a18SSam Leffler curfrags = 0; 1686eeb76a18SSam Leffler for (m = m0; m != NULL; m = m->m_next) 168782334850SJohn Baldwin curfrags += frags_per_mbuf(m); 1688eeb76a18SSam Leffler /* 1689eeb76a18SSam Leffler * First, try to collapse mbufs. Note that we always collapse 1690eeb76a18SSam Leffler * towards the front so we don't need to deal with moving the 1691eeb76a18SSam Leffler * pkthdr. This may be suboptimal if the first mbuf has much 1692eeb76a18SSam Leffler * less data than the following. 1693eeb76a18SSam Leffler */ 1694eeb76a18SSam Leffler m = m0; 1695eeb76a18SSam Leffler again: 1696eeb76a18SSam Leffler for (;;) { 1697eeb76a18SSam Leffler n = m->m_next; 1698eeb76a18SSam Leffler if (n == NULL) 1699eeb76a18SSam Leffler break; 170014d7c5b1SAndre Oppermann if (M_WRITABLE(m) && 1701eeb76a18SSam Leffler n->m_len < M_TRAILINGSPACE(m)) { 170282334850SJohn Baldwin m_copydata(n, 0, n->m_len, 170382334850SJohn Baldwin mtod(m, char *) + m->m_len); 1704eeb76a18SSam Leffler m->m_len += n->m_len; 1705eeb76a18SSam Leffler m->m_next = n->m_next; 170682334850SJohn Baldwin curfrags -= frags_per_mbuf(n); 1707eeb76a18SSam Leffler m_free(n); 170882334850SJohn Baldwin if (curfrags <= maxfrags) 1709eeb76a18SSam Leffler return m0; 1710eeb76a18SSam Leffler } else 1711eeb76a18SSam Leffler m = n; 1712eeb76a18SSam Leffler } 1713eeb76a18SSam Leffler KASSERT(maxfrags > 1, 1714eeb76a18SSam Leffler ("maxfrags %u, but normal collapse failed", maxfrags)); 1715eeb76a18SSam Leffler /* 1716eeb76a18SSam Leffler * Collapse consecutive mbufs to a cluster. 1717eeb76a18SSam Leffler */ 1718eeb76a18SSam Leffler prev = &m0->m_next; /* NB: not the first mbuf */ 1719eeb76a18SSam Leffler while ((n = *prev) != NULL) { 1720eeb76a18SSam Leffler if ((n2 = n->m_next) != NULL && 1721eeb76a18SSam Leffler n->m_len + n2->m_len < MCLBYTES) { 1722eeb76a18SSam Leffler m = m_getcl(how, MT_DATA, 0); 1723eeb76a18SSam Leffler if (m == NULL) 1724eeb76a18SSam Leffler goto bad; 172582334850SJohn Baldwin m_copydata(n, 0, n->m_len, mtod(m, char *)); 172682334850SJohn Baldwin m_copydata(n2, 0, n2->m_len, 172782334850SJohn Baldwin mtod(m, char *) + n->m_len); 1728eeb76a18SSam Leffler m->m_len = n->m_len + n2->m_len; 1729eeb76a18SSam Leffler m->m_next = n2->m_next; 1730eeb76a18SSam Leffler *prev = m; 173182334850SJohn Baldwin curfrags += 1; /* For the new cluster */ 173282334850SJohn Baldwin curfrags -= frags_per_mbuf(n); 173382334850SJohn Baldwin curfrags -= frags_per_mbuf(n2); 1734eeb76a18SSam Leffler m_free(n); 1735eeb76a18SSam Leffler m_free(n2); 173682334850SJohn Baldwin if (curfrags <= maxfrags) 1737eeb76a18SSam Leffler return m0; 1738eeb76a18SSam Leffler /* 1739eeb76a18SSam Leffler * Still not there, try the normal collapse 1740eeb76a18SSam Leffler * again before we allocate another cluster. 1741eeb76a18SSam Leffler */ 1742eeb76a18SSam Leffler goto again; 1743eeb76a18SSam Leffler } 1744eeb76a18SSam Leffler prev = &n->m_next; 1745eeb76a18SSam Leffler } 1746eeb76a18SSam Leffler /* 1747eeb76a18SSam Leffler * No place where we can collapse to a cluster; punt. 1748eeb76a18SSam Leffler * This can occur if, for example, you request 2 frags 1749eeb76a18SSam Leffler * but the packet requires that both be clusters (we 1750eeb76a18SSam Leffler * never reallocate the first mbuf to avoid moving the 1751eeb76a18SSam Leffler * packet header). 1752eeb76a18SSam Leffler */ 1753eeb76a18SSam Leffler bad: 1754eeb76a18SSam Leffler return NULL; 1755eeb76a18SSam Leffler } 1756eeb76a18SSam Leffler 17573390d476SMike Silbersack #ifdef MBUF_STRESS_TEST 17583390d476SMike Silbersack 17593390d476SMike Silbersack /* 17603390d476SMike Silbersack * Fragment an mbuf chain. There's no reason you'd ever want to do 17613390d476SMike Silbersack * this in normal usage, but it's great for stress testing various 17623390d476SMike Silbersack * mbuf consumers. 17633390d476SMike Silbersack * 17643390d476SMike Silbersack * If fragmentation is not possible, the original chain will be 17653390d476SMike Silbersack * returned. 17663390d476SMike Silbersack * 17673390d476SMike Silbersack * Possible length values: 17683390d476SMike Silbersack * 0 no fragmentation will occur 17693390d476SMike Silbersack * > 0 each fragment will be of the specified length 17703390d476SMike Silbersack * -1 each fragment will be the same random value in length 17713390d476SMike Silbersack * -2 each fragment's length will be entirely random 17723390d476SMike Silbersack * (Random values range from 1 to 256) 17733390d476SMike Silbersack */ 17743390d476SMike Silbersack struct mbuf * 17753390d476SMike Silbersack m_fragment(struct mbuf *m0, int how, int length) 17763390d476SMike Silbersack { 177766234298SAndriy Voskoboinyk struct mbuf *m_first, *m_last; 177866234298SAndriy Voskoboinyk int divisor = 255, progress = 0, fraglen; 17793390d476SMike Silbersack 17803390d476SMike Silbersack if (!(m0->m_flags & M_PKTHDR)) 17813390d476SMike Silbersack return (m0); 17823390d476SMike Silbersack 178366234298SAndriy Voskoboinyk if (length == 0 || length < -2) 17843390d476SMike Silbersack return (m0); 178566234298SAndriy Voskoboinyk if (length > MCLBYTES) 178666234298SAndriy Voskoboinyk length = MCLBYTES; 178766234298SAndriy Voskoboinyk if (length < 0 && divisor > MCLBYTES) 178866234298SAndriy Voskoboinyk divisor = MCLBYTES; 178966234298SAndriy Voskoboinyk if (length == -1) 179066234298SAndriy Voskoboinyk length = 1 + (arc4random() % divisor); 179166234298SAndriy Voskoboinyk if (length > 0) 179266234298SAndriy Voskoboinyk fraglen = length; 17933390d476SMike Silbersack 17943390d476SMike Silbersack m_fixhdr(m0); /* Needed sanity check */ 17953390d476SMike Silbersack 179666234298SAndriy Voskoboinyk m_first = m_getcl(how, MT_DATA, M_PKTHDR); 179766234298SAndriy Voskoboinyk if (m_first == NULL) 17983390d476SMike Silbersack goto nospace; 17993390d476SMike Silbersack 180066234298SAndriy Voskoboinyk if (m_dup_pkthdr(m_first, m0, how) == 0) 18013390d476SMike Silbersack goto nospace; 18023390d476SMike Silbersack 180366234298SAndriy Voskoboinyk m_last = m_first; 18043390d476SMike Silbersack 18053390d476SMike Silbersack while (progress < m0->m_pkthdr.len) { 180666234298SAndriy Voskoboinyk if (length == -2) 180766234298SAndriy Voskoboinyk fraglen = 1 + (arc4random() % divisor); 18083390d476SMike Silbersack if (fraglen > m0->m_pkthdr.len - progress) 18093390d476SMike Silbersack fraglen = m0->m_pkthdr.len - progress; 18103390d476SMike Silbersack 181166234298SAndriy Voskoboinyk if (progress != 0) { 181266234298SAndriy Voskoboinyk struct mbuf *m_new = m_getcl(how, MT_DATA, 0); 18133390d476SMike Silbersack if (m_new == NULL) 18143390d476SMike Silbersack goto nospace; 181566234298SAndriy Voskoboinyk 181666234298SAndriy Voskoboinyk m_last->m_next = m_new; 181766234298SAndriy Voskoboinyk m_last = m_new; 18183390d476SMike Silbersack } 18193390d476SMike Silbersack 182066234298SAndriy Voskoboinyk m_copydata(m0, progress, fraglen, mtod(m_last, caddr_t)); 18213390d476SMike Silbersack progress += fraglen; 182266234298SAndriy Voskoboinyk m_last->m_len = fraglen; 18233390d476SMike Silbersack } 18243390d476SMike Silbersack m_freem(m0); 182566234298SAndriy Voskoboinyk m0 = m_first; 18263390d476SMike Silbersack return (m0); 18273390d476SMike Silbersack nospace: 182866234298SAndriy Voskoboinyk if (m_first) 182966234298SAndriy Voskoboinyk m_freem(m_first); 18303390d476SMike Silbersack /* Return the original chain on failure */ 18313390d476SMike Silbersack return (m0); 18323390d476SMike Silbersack } 18333390d476SMike Silbersack 18343390d476SMike Silbersack #endif 1835beb699c7SMike Silbersack 18365e20f43dSAndre Oppermann /* 183782334850SJohn Baldwin * Free pages from mbuf_ext_pgs, assuming they were allocated via 183882334850SJohn Baldwin * vm_page_alloc() and aren't associated with any object. Complement 183982334850SJohn Baldwin * to allocator from m_uiotombuf_nomap(). 184082334850SJohn Baldwin */ 184182334850SJohn Baldwin void 184282334850SJohn Baldwin mb_free_mext_pgs(struct mbuf *m) 184382334850SJohn Baldwin { 184482334850SJohn Baldwin vm_page_t pg; 184582334850SJohn Baldwin 1846365e8da4SGleb Smirnoff M_ASSERTEXTPG(m); 18477b6c99d0SGleb Smirnoff for (int i = 0; i < m->m_epg_npgs; i++) { 18480c103266SGleb Smirnoff pg = PHYS_TO_VM_PAGE(m->m_epg_pa[i]); 18499fb7c918SMark Johnston vm_page_unwire_noq(pg); 18509fb7c918SMark Johnston vm_page_free(pg); 185182334850SJohn Baldwin } 185282334850SJohn Baldwin } 185382334850SJohn Baldwin 185482334850SJohn Baldwin static struct mbuf * 185582334850SJohn Baldwin m_uiotombuf_nomap(struct uio *uio, int how, int len, int maxseg, int flags) 185682334850SJohn Baldwin { 185782334850SJohn Baldwin struct mbuf *m, *mb, *prev; 185882334850SJohn Baldwin vm_page_t pg_array[MBUF_PEXT_MAX_PGS]; 18599fb7c918SMark Johnston int error, length, i, needed; 186082334850SJohn Baldwin ssize_t total; 1861a4667e09SMark Johnston int pflags = malloc2vm_flags(how) | VM_ALLOC_NODUMP | VM_ALLOC_WIRED; 186282334850SJohn Baldwin 1863c2a8fd6fSJohn Baldwin MPASS((flags & M_PKTHDR) == 0); 1864608c44f9SMark Johnston MPASS((how & M_ZERO) == 0); 1865c2a8fd6fSJohn Baldwin 186682334850SJohn Baldwin /* 186782334850SJohn Baldwin * len can be zero or an arbitrary large value bound by 186882334850SJohn Baldwin * the total data supplied by the uio. 186982334850SJohn Baldwin */ 187082334850SJohn Baldwin if (len > 0) 187182334850SJohn Baldwin total = MIN(uio->uio_resid, len); 187282334850SJohn Baldwin else 187382334850SJohn Baldwin total = uio->uio_resid; 187482334850SJohn Baldwin 187582334850SJohn Baldwin if (maxseg == 0) 187682334850SJohn Baldwin maxseg = MBUF_PEXT_MAX_PGS * PAGE_SIZE; 187782334850SJohn Baldwin 187882334850SJohn Baldwin /* 1879c2a8fd6fSJohn Baldwin * If total is zero, return an empty mbuf. This can occur 1880c2a8fd6fSJohn Baldwin * for TLS 1.0 connections which send empty fragments as 1881c2a8fd6fSJohn Baldwin * a countermeasure against the known-IV weakness in CBC 1882c2a8fd6fSJohn Baldwin * ciphersuites. 1883c2a8fd6fSJohn Baldwin */ 1884c2a8fd6fSJohn Baldwin if (__predict_false(total == 0)) { 1885c2a8fd6fSJohn Baldwin mb = mb_alloc_ext_pgs(how, mb_free_mext_pgs); 1886c2a8fd6fSJohn Baldwin if (mb == NULL) 1887c2a8fd6fSJohn Baldwin return (NULL); 1888c2a8fd6fSJohn Baldwin mb->m_epg_flags = EPG_FLAG_ANON; 1889c2a8fd6fSJohn Baldwin return (mb); 1890c2a8fd6fSJohn Baldwin } 1891c2a8fd6fSJohn Baldwin 1892c2a8fd6fSJohn Baldwin /* 189382334850SJohn Baldwin * Allocate the pages 189482334850SJohn Baldwin */ 189582334850SJohn Baldwin m = NULL; 189682334850SJohn Baldwin while (total > 0) { 189723feb563SAndrew Gallatin mb = mb_alloc_ext_pgs(how, mb_free_mext_pgs); 189882334850SJohn Baldwin if (mb == NULL) 189982334850SJohn Baldwin goto failed; 190082334850SJohn Baldwin if (m == NULL) 190182334850SJohn Baldwin m = mb; 190282334850SJohn Baldwin else 190382334850SJohn Baldwin prev->m_next = mb; 190482334850SJohn Baldwin prev = mb; 19057b6c99d0SGleb Smirnoff mb->m_epg_flags = EPG_FLAG_ANON; 190682334850SJohn Baldwin needed = length = MIN(maxseg, total); 190782334850SJohn Baldwin for (i = 0; needed > 0; i++, needed -= PAGE_SIZE) { 190882334850SJohn Baldwin retry_page: 1909a4667e09SMark Johnston pg_array[i] = vm_page_alloc_noobj(pflags); 191082334850SJohn Baldwin if (pg_array[i] == NULL) { 191182334850SJohn Baldwin if (how & M_NOWAIT) { 191282334850SJohn Baldwin goto failed; 191382334850SJohn Baldwin } else { 191482334850SJohn Baldwin vm_wait(NULL); 191582334850SJohn Baldwin goto retry_page; 191682334850SJohn Baldwin } 191782334850SJohn Baldwin } 19180c103266SGleb Smirnoff mb->m_epg_pa[i] = VM_PAGE_TO_PHYS(pg_array[i]); 19197b6c99d0SGleb Smirnoff mb->m_epg_npgs++; 192082334850SJohn Baldwin } 19217b6c99d0SGleb Smirnoff mb->m_epg_last_len = length - PAGE_SIZE * (mb->m_epg_npgs - 1); 19220c103266SGleb Smirnoff MBUF_EXT_PGS_ASSERT_SANITY(mb); 192382334850SJohn Baldwin total -= length; 192482334850SJohn Baldwin error = uiomove_fromphys(pg_array, 0, length, uio); 192582334850SJohn Baldwin if (error != 0) 192682334850SJohn Baldwin goto failed; 192782334850SJohn Baldwin mb->m_len = length; 19287b6c99d0SGleb Smirnoff mb->m_ext.ext_size += PAGE_SIZE * mb->m_epg_npgs; 192982334850SJohn Baldwin if (flags & M_PKTHDR) 193082334850SJohn Baldwin m->m_pkthdr.len += length; 193182334850SJohn Baldwin } 193282334850SJohn Baldwin return (m); 193382334850SJohn Baldwin 193482334850SJohn Baldwin failed: 193582334850SJohn Baldwin m_freem(m); 193682334850SJohn Baldwin return (NULL); 193782334850SJohn Baldwin } 193882334850SJohn Baldwin 193982334850SJohn Baldwin /* 19405e20f43dSAndre Oppermann * Copy the contents of uio into a properly sized mbuf chain. 1941*aba79b0fSGleb Smirnoff * A compat KPI. Users are recommended to use direct calls to backing 1942*aba79b0fSGleb Smirnoff * functions. 19435e20f43dSAndre Oppermann */ 1944beb699c7SMike Silbersack struct mbuf * 1945*aba79b0fSGleb Smirnoff m_uiotombuf(struct uio *uio, int how, int len, int lspace, int flags) 1946beb699c7SMike Silbersack { 1947beb699c7SMike Silbersack 1948*aba79b0fSGleb Smirnoff if (flags & M_EXTPG) { 1949*aba79b0fSGleb Smirnoff /* XXX: 'lspace' magically becomes maxseg! */ 1950*aba79b0fSGleb Smirnoff return (m_uiotombuf_nomap(uio, how, len, lspace, flags)); 1951*aba79b0fSGleb Smirnoff } else if (__predict_false(uio->uio_resid + len == 0)) { 1952*aba79b0fSGleb Smirnoff struct mbuf *m; 195382334850SJohn Baldwin 19545e20f43dSAndre Oppermann /* 1955*aba79b0fSGleb Smirnoff * m_uiotombuf() is known to return zero length buffer, keep 1956*aba79b0fSGleb Smirnoff * this compatibility. mc_uiotomc() won't do that. 19575e20f43dSAndre Oppermann */ 1958*aba79b0fSGleb Smirnoff if (flags & M_PKTHDR) { 1959*aba79b0fSGleb Smirnoff m = m_gethdr(how, MT_DATA); 1960*aba79b0fSGleb Smirnoff m->m_pkthdr.memlen = MSIZE; 1961*aba79b0fSGleb Smirnoff } else 1962*aba79b0fSGleb Smirnoff m = m_get(how, MT_DATA); 1963*aba79b0fSGleb Smirnoff if (m != NULL) 1964*aba79b0fSGleb Smirnoff m->m_data += lspace; 1965*aba79b0fSGleb Smirnoff return (m); 1966*aba79b0fSGleb Smirnoff } else { 1967*aba79b0fSGleb Smirnoff struct mchain mc; 1968*aba79b0fSGleb Smirnoff int error; 1969*aba79b0fSGleb Smirnoff 1970*aba79b0fSGleb Smirnoff error = mc_uiotomc(&mc, uio, len, lspace, how, flags); 1971*aba79b0fSGleb Smirnoff if (__predict_true(error == 0)) { 1972*aba79b0fSGleb Smirnoff if (flags & M_PKTHDR) { 1973*aba79b0fSGleb Smirnoff mc_first(&mc)->m_pkthdr.len = mc.mc_len; 1974*aba79b0fSGleb Smirnoff mc_first(&mc)->m_pkthdr.memlen = mc.mc_mlen; 1975*aba79b0fSGleb Smirnoff } 1976*aba79b0fSGleb Smirnoff return (mc_first(&mc)); 1977*aba79b0fSGleb Smirnoff } else 1978*aba79b0fSGleb Smirnoff return (NULL); 1979*aba79b0fSGleb Smirnoff } 1980*aba79b0fSGleb Smirnoff } 1981*aba79b0fSGleb Smirnoff 1982*aba79b0fSGleb Smirnoff /* 1983*aba79b0fSGleb Smirnoff * Copy the contents of uio into a properly sized mbuf chain. 1984*aba79b0fSGleb Smirnoff * In case of failure state of mchain is inconsistent. 1985*aba79b0fSGleb Smirnoff * @param length Limit copyout length. If 0 entire uio_resid is copied. 1986*aba79b0fSGleb Smirnoff * @param lspace Provide leading space in the first mbuf in the chain. 1987*aba79b0fSGleb Smirnoff */ 1988*aba79b0fSGleb Smirnoff int 1989*aba79b0fSGleb Smirnoff mc_uiotomc(struct mchain *mc, struct uio *uio, u_int length, u_int lspace, 1990*aba79b0fSGleb Smirnoff int how, int flags) 1991*aba79b0fSGleb Smirnoff { 1992*aba79b0fSGleb Smirnoff struct mbuf *mb; 1993*aba79b0fSGleb Smirnoff u_int total; 1994*aba79b0fSGleb Smirnoff int error; 1995*aba79b0fSGleb Smirnoff 1996*aba79b0fSGleb Smirnoff MPASS(lspace < MHLEN); 1997*aba79b0fSGleb Smirnoff MPASS(UINT_MAX - lspace >= length); 1998*aba79b0fSGleb Smirnoff MPASS(uio->uio_rw == UIO_WRITE); 1999*aba79b0fSGleb Smirnoff MPASS(uio->uio_resid >= 0); 2000*aba79b0fSGleb Smirnoff 2001*aba79b0fSGleb Smirnoff if (length > 0) { 2002*aba79b0fSGleb Smirnoff if (uio->uio_resid > length) { 2003*aba79b0fSGleb Smirnoff total = length; 2004*aba79b0fSGleb Smirnoff flags &= ~M_EOR; 2005*aba79b0fSGleb Smirnoff } else 2006*aba79b0fSGleb Smirnoff total = uio->uio_resid; 2007*aba79b0fSGleb Smirnoff } else if (__predict_false(uio->uio_resid + lspace > UINT_MAX)) 2008*aba79b0fSGleb Smirnoff return (EOVERFLOW); 2009beb699c7SMike Silbersack else 2010beb699c7SMike Silbersack total = uio->uio_resid; 20115e20f43dSAndre Oppermann 2012*aba79b0fSGleb Smirnoff if (__predict_false(total + lspace == 0)) { 2013*aba79b0fSGleb Smirnoff *mc = MCHAIN_INITIALIZER(mc); 2014*aba79b0fSGleb Smirnoff return (0); 2015*aba79b0fSGleb Smirnoff } 20165e20f43dSAndre Oppermann 2017*aba79b0fSGleb Smirnoff error = mc_get(mc, total + lspace, how, MT_DATA, flags); 2018*aba79b0fSGleb Smirnoff if (__predict_false(error)) 2019*aba79b0fSGleb Smirnoff return (error); 2020*aba79b0fSGleb Smirnoff mc_first(mc)->m_data += lspace; 20215e20f43dSAndre Oppermann 20225e20f43dSAndre Oppermann /* Fill all mbufs with uio data and update header information. */ 2023*aba79b0fSGleb Smirnoff STAILQ_FOREACH(mb, &mc->mc_q, m_stailq) { 2024*aba79b0fSGleb Smirnoff u_int mlen; 20255e20f43dSAndre Oppermann 2026*aba79b0fSGleb Smirnoff mlen = min(M_TRAILINGSPACE(mb), total - mc->mc_len); 2027*aba79b0fSGleb Smirnoff error = uiomove(mtod(mb, void *), mlen, uio); 2028*aba79b0fSGleb Smirnoff if (__predict_false(error)) { 2029*aba79b0fSGleb Smirnoff mc_freem(mc); 2030*aba79b0fSGleb Smirnoff return (error); 20315e20f43dSAndre Oppermann } 2032*aba79b0fSGleb Smirnoff mb->m_len = mlen; 2033*aba79b0fSGleb Smirnoff mc->mc_len += mlen; 2034*aba79b0fSGleb Smirnoff } 2035*aba79b0fSGleb Smirnoff MPASS(mc->mc_len == total); 20365e20f43dSAndre Oppermann 2037*aba79b0fSGleb Smirnoff return (0); 2038beb699c7SMike Silbersack } 2039ab8ab90cSSam Leffler 2040ab8ab90cSSam Leffler /* 20413f9dac85SJohn Baldwin * Copy data to/from an unmapped mbuf into a uio limited by len if set. 204282334850SJohn Baldwin */ 204382334850SJohn Baldwin int 2044aa341db3SJohn Baldwin m_unmapped_uiomove(const struct mbuf *m, int m_off, struct uio *uio, int len) 204582334850SJohn Baldwin { 204682334850SJohn Baldwin vm_page_t pg; 204782334850SJohn Baldwin int error, i, off, pglen, pgoff, seglen, segoff; 204882334850SJohn Baldwin 2049365e8da4SGleb Smirnoff M_ASSERTEXTPG(m); 205082334850SJohn Baldwin error = 0; 205182334850SJohn Baldwin 205282334850SJohn Baldwin /* Skip over any data removed from the front. */ 205382334850SJohn Baldwin off = mtod(m, vm_offset_t); 205482334850SJohn Baldwin 205582334850SJohn Baldwin off += m_off; 20567b6c99d0SGleb Smirnoff if (m->m_epg_hdrlen != 0) { 20577b6c99d0SGleb Smirnoff if (off >= m->m_epg_hdrlen) { 20587b6c99d0SGleb Smirnoff off -= m->m_epg_hdrlen; 205982334850SJohn Baldwin } else { 20607b6c99d0SGleb Smirnoff seglen = m->m_epg_hdrlen - off; 206182334850SJohn Baldwin segoff = off; 206282334850SJohn Baldwin seglen = min(seglen, len); 206382334850SJohn Baldwin off = 0; 206482334850SJohn Baldwin len -= seglen; 20650c103266SGleb Smirnoff error = uiomove(__DECONST(void *, 20660c103266SGleb Smirnoff &m->m_epg_hdr[segoff]), seglen, uio); 206782334850SJohn Baldwin } 206882334850SJohn Baldwin } 20697b6c99d0SGleb Smirnoff pgoff = m->m_epg_1st_off; 20707b6c99d0SGleb Smirnoff for (i = 0; i < m->m_epg_npgs && error == 0 && len > 0; i++) { 2071c4ee38f8SGleb Smirnoff pglen = m_epg_pagelen(m, i, pgoff); 207282334850SJohn Baldwin if (off >= pglen) { 207382334850SJohn Baldwin off -= pglen; 207482334850SJohn Baldwin pgoff = 0; 207582334850SJohn Baldwin continue; 207682334850SJohn Baldwin } 207782334850SJohn Baldwin seglen = pglen - off; 207882334850SJohn Baldwin segoff = pgoff + off; 207982334850SJohn Baldwin off = 0; 208082334850SJohn Baldwin seglen = min(seglen, len); 208182334850SJohn Baldwin len -= seglen; 20820c103266SGleb Smirnoff pg = PHYS_TO_VM_PAGE(m->m_epg_pa[i]); 208382334850SJohn Baldwin error = uiomove_fromphys(&pg, segoff, seglen, uio); 208482334850SJohn Baldwin pgoff = 0; 208582334850SJohn Baldwin }; 208682334850SJohn Baldwin if (len != 0 && error == 0) { 20877b6c99d0SGleb Smirnoff KASSERT((off + len) <= m->m_epg_trllen, 208882334850SJohn Baldwin ("off + len > trail (%d + %d > %d, m_off = %d)", off, len, 20897b6c99d0SGleb Smirnoff m->m_epg_trllen, m_off)); 20900c103266SGleb Smirnoff error = uiomove(__DECONST(void *, &m->m_epg_trail[off]), 20910c103266SGleb Smirnoff len, uio); 209282334850SJohn Baldwin } 209382334850SJohn Baldwin return (error); 209482334850SJohn Baldwin } 209582334850SJohn Baldwin 209682334850SJohn Baldwin /* 2097bc05b2f6SAndre Oppermann * Copy an mbuf chain into a uio limited by len if set. 2098bc05b2f6SAndre Oppermann */ 2099bc05b2f6SAndre Oppermann int 210014984031SGleb Smirnoff m_mbuftouio(struct uio *uio, const struct mbuf *m, int len) 2101bc05b2f6SAndre Oppermann { 2102bc05b2f6SAndre Oppermann int error, length, total; 2103bc05b2f6SAndre Oppermann int progress = 0; 2104bc05b2f6SAndre Oppermann 2105bc05b2f6SAndre Oppermann if (len > 0) 2106bc05b2f6SAndre Oppermann total = min(uio->uio_resid, len); 2107bc05b2f6SAndre Oppermann else 2108bc05b2f6SAndre Oppermann total = uio->uio_resid; 2109bc05b2f6SAndre Oppermann 2110bc05b2f6SAndre Oppermann /* Fill the uio with data from the mbufs. */ 2111bc05b2f6SAndre Oppermann for (; m != NULL; m = m->m_next) { 2112bc05b2f6SAndre Oppermann length = min(m->m_len, total - progress); 2113bc05b2f6SAndre Oppermann 21146edfd179SGleb Smirnoff if ((m->m_flags & M_EXTPG) != 0) 2115aa341db3SJohn Baldwin error = m_unmapped_uiomove(m, 0, uio, length); 211682334850SJohn Baldwin else 2117bc05b2f6SAndre Oppermann error = uiomove(mtod(m, void *), length, uio); 2118bc05b2f6SAndre Oppermann if (error) 2119bc05b2f6SAndre Oppermann return (error); 2120bc05b2f6SAndre Oppermann 2121bc05b2f6SAndre Oppermann progress += length; 2122bc05b2f6SAndre Oppermann } 2123bc05b2f6SAndre Oppermann 2124bc05b2f6SAndre Oppermann return (0); 2125bc05b2f6SAndre Oppermann } 2126bc05b2f6SAndre Oppermann 2127bc05b2f6SAndre Oppermann /* 212847e2996eSSam Leffler * Create a writable copy of the mbuf chain. While doing this 212947e2996eSSam Leffler * we compact the chain with a goal of producing a chain with 213047e2996eSSam Leffler * at most two mbufs. The second mbuf in this chain is likely 213147e2996eSSam Leffler * to be a cluster. The primary purpose of this work is to create 213247e2996eSSam Leffler * a writable packet for encryption, compression, etc. The 213347e2996eSSam Leffler * secondary goal is to linearize the data so the data can be 213447e2996eSSam Leffler * passed to crypto hardware in the most efficient manner possible. 213547e2996eSSam Leffler */ 213647e2996eSSam Leffler struct mbuf * 213747e2996eSSam Leffler m_unshare(struct mbuf *m0, int how) 213847e2996eSSam Leffler { 213947e2996eSSam Leffler struct mbuf *m, *mprev; 214047e2996eSSam Leffler struct mbuf *n, *mfirst, *mlast; 214147e2996eSSam Leffler int len, off; 214247e2996eSSam Leffler 214347e2996eSSam Leffler mprev = NULL; 214447e2996eSSam Leffler for (m = m0; m != NULL; m = mprev->m_next) { 214547e2996eSSam Leffler /* 214647e2996eSSam Leffler * Regular mbufs are ignored unless there's a cluster 214747e2996eSSam Leffler * in front of it that we can use to coalesce. We do 214847e2996eSSam Leffler * the latter mainly so later clusters can be coalesced 214947e2996eSSam Leffler * also w/o having to handle them specially (i.e. convert 215047e2996eSSam Leffler * mbuf+cluster -> cluster). This optimization is heavily 215147e2996eSSam Leffler * influenced by the assumption that we're running over 215247e2996eSSam Leffler * Ethernet where MCLBYTES is large enough that the max 215347e2996eSSam Leffler * packet size will permit lots of coalescing into a 215447e2996eSSam Leffler * single cluster. This in turn permits efficient 215547e2996eSSam Leffler * crypto operations, especially when using hardware. 215647e2996eSSam Leffler */ 215747e2996eSSam Leffler if ((m->m_flags & M_EXT) == 0) { 215847e2996eSSam Leffler if (mprev && (mprev->m_flags & M_EXT) && 215947e2996eSSam Leffler m->m_len <= M_TRAILINGSPACE(mprev)) { 216047e2996eSSam Leffler /* XXX: this ignores mbuf types */ 216147e2996eSSam Leffler memcpy(mtod(mprev, caddr_t) + mprev->m_len, 216247e2996eSSam Leffler mtod(m, caddr_t), m->m_len); 216347e2996eSSam Leffler mprev->m_len += m->m_len; 216447e2996eSSam Leffler mprev->m_next = m->m_next; /* unlink from chain */ 216547e2996eSSam Leffler m_free(m); /* reclaim mbuf */ 216647e2996eSSam Leffler } else { 216747e2996eSSam Leffler mprev = m; 216847e2996eSSam Leffler } 216947e2996eSSam Leffler continue; 217047e2996eSSam Leffler } 217147e2996eSSam Leffler /* 217247e2996eSSam Leffler * Writable mbufs are left alone (for now). 217347e2996eSSam Leffler */ 217447e2996eSSam Leffler if (M_WRITABLE(m)) { 217547e2996eSSam Leffler mprev = m; 217647e2996eSSam Leffler continue; 217747e2996eSSam Leffler } 217847e2996eSSam Leffler 217947e2996eSSam Leffler /* 218047e2996eSSam Leffler * Not writable, replace with a copy or coalesce with 218147e2996eSSam Leffler * the previous mbuf if possible (since we have to copy 218247e2996eSSam Leffler * it anyway, we try to reduce the number of mbufs and 218347e2996eSSam Leffler * clusters so that future work is easier). 218447e2996eSSam Leffler */ 218547e2996eSSam Leffler KASSERT(m->m_flags & M_EXT, ("m_flags 0x%x", m->m_flags)); 218647e2996eSSam Leffler /* NB: we only coalesce into a cluster or larger */ 218747e2996eSSam Leffler if (mprev != NULL && (mprev->m_flags & M_EXT) && 218847e2996eSSam Leffler m->m_len <= M_TRAILINGSPACE(mprev)) { 218947e2996eSSam Leffler /* XXX: this ignores mbuf types */ 219047e2996eSSam Leffler memcpy(mtod(mprev, caddr_t) + mprev->m_len, 219147e2996eSSam Leffler mtod(m, caddr_t), m->m_len); 219247e2996eSSam Leffler mprev->m_len += m->m_len; 219347e2996eSSam Leffler mprev->m_next = m->m_next; /* unlink from chain */ 219447e2996eSSam Leffler m_free(m); /* reclaim mbuf */ 219547e2996eSSam Leffler continue; 219647e2996eSSam Leffler } 219747e2996eSSam Leffler 219847e2996eSSam Leffler /* 21995368b81eSGleb Smirnoff * Allocate new space to hold the copy and copy the data. 22005368b81eSGleb Smirnoff * We deal with jumbo mbufs (i.e. m_len > MCLBYTES) by 22015368b81eSGleb Smirnoff * splitting them into clusters. We could just malloc a 22025368b81eSGleb Smirnoff * buffer and make it external but too many device drivers 22035368b81eSGleb Smirnoff * don't know how to break up the non-contiguous memory when 22045368b81eSGleb Smirnoff * doing DMA. 220547e2996eSSam Leffler */ 2206fddd4f62SNavdeep Parhar n = m_getcl(how, m->m_type, m->m_flags & M_COPYFLAGS); 220747e2996eSSam Leffler if (n == NULL) { 220847e2996eSSam Leffler m_freem(m0); 220947e2996eSSam Leffler return (NULL); 221047e2996eSSam Leffler } 2211e40e8705SGleb Smirnoff if (m->m_flags & M_PKTHDR) { 2212e40e8705SGleb Smirnoff KASSERT(mprev == NULL, ("%s: m0 %p, m %p has M_PKTHDR", 2213e40e8705SGleb Smirnoff __func__, m0, m)); 2214e40e8705SGleb Smirnoff m_move_pkthdr(n, m); 2215e40e8705SGleb Smirnoff } 221647e2996eSSam Leffler len = m->m_len; 221747e2996eSSam Leffler off = 0; 221847e2996eSSam Leffler mfirst = n; 221947e2996eSSam Leffler mlast = NULL; 222047e2996eSSam Leffler for (;;) { 222147e2996eSSam Leffler int cc = min(len, MCLBYTES); 222247e2996eSSam Leffler memcpy(mtod(n, caddr_t), mtod(m, caddr_t) + off, cc); 222347e2996eSSam Leffler n->m_len = cc; 222447e2996eSSam Leffler if (mlast != NULL) 222547e2996eSSam Leffler mlast->m_next = n; 222647e2996eSSam Leffler mlast = n; 222747e2996eSSam Leffler #if 0 222847e2996eSSam Leffler newipsecstat.ips_clcopied++; 222947e2996eSSam Leffler #endif 223047e2996eSSam Leffler 223147e2996eSSam Leffler len -= cc; 223247e2996eSSam Leffler if (len <= 0) 223347e2996eSSam Leffler break; 223447e2996eSSam Leffler off += cc; 223547e2996eSSam Leffler 2236fddd4f62SNavdeep Parhar n = m_getcl(how, m->m_type, m->m_flags & M_COPYFLAGS); 223747e2996eSSam Leffler if (n == NULL) { 223847e2996eSSam Leffler m_freem(mfirst); 223947e2996eSSam Leffler m_freem(m0); 224047e2996eSSam Leffler return (NULL); 224147e2996eSSam Leffler } 224247e2996eSSam Leffler } 224347e2996eSSam Leffler n->m_next = m->m_next; 224447e2996eSSam Leffler if (mprev == NULL) 224547e2996eSSam Leffler m0 = mfirst; /* new head of chain */ 224647e2996eSSam Leffler else 224747e2996eSSam Leffler mprev->m_next = mfirst; /* replace old mbuf */ 224847e2996eSSam Leffler m_free(m); /* release old mbuf */ 224947e2996eSSam Leffler mprev = mfirst; 225047e2996eSSam Leffler } 225147e2996eSSam Leffler return (m0); 225247e2996eSSam Leffler } 22536eeac1d9SJulian Elischer 22546eeac1d9SJulian Elischer #ifdef MBUF_PROFILING 22556eeac1d9SJulian Elischer 22566eeac1d9SJulian Elischer #define MP_BUCKETS 32 /* don't just change this as things may overflow.*/ 22576eeac1d9SJulian Elischer struct mbufprofile { 22582182c0cfSJulian Elischer uintmax_t wasted[MP_BUCKETS]; 22592182c0cfSJulian Elischer uintmax_t used[MP_BUCKETS]; 22602182c0cfSJulian Elischer uintmax_t segments[MP_BUCKETS]; 22616eeac1d9SJulian Elischer } mbprof; 22626eeac1d9SJulian Elischer 22636eeac1d9SJulian Elischer void 22646eeac1d9SJulian Elischer m_profile(struct mbuf *m) 22656eeac1d9SJulian Elischer { 22666eeac1d9SJulian Elischer int segments = 0; 22676eeac1d9SJulian Elischer int used = 0; 22686eeac1d9SJulian Elischer int wasted = 0; 22696eeac1d9SJulian Elischer 22706eeac1d9SJulian Elischer while (m) { 22716eeac1d9SJulian Elischer segments++; 22726eeac1d9SJulian Elischer used += m->m_len; 22736eeac1d9SJulian Elischer if (m->m_flags & M_EXT) { 22746eeac1d9SJulian Elischer wasted += MHLEN - sizeof(m->m_ext) + 22756eeac1d9SJulian Elischer m->m_ext.ext_size - m->m_len; 22766eeac1d9SJulian Elischer } else { 22776eeac1d9SJulian Elischer if (m->m_flags & M_PKTHDR) 22786eeac1d9SJulian Elischer wasted += MHLEN - m->m_len; 22796eeac1d9SJulian Elischer else 22806eeac1d9SJulian Elischer wasted += MLEN - m->m_len; 22816eeac1d9SJulian Elischer } 22826eeac1d9SJulian Elischer m = m->m_next; 22836eeac1d9SJulian Elischer } 22846eeac1d9SJulian Elischer /* be paranoid.. it helps */ 22856eeac1d9SJulian Elischer if (segments > MP_BUCKETS - 1) 22866eeac1d9SJulian Elischer segments = MP_BUCKETS - 1; 22876eeac1d9SJulian Elischer if (used > 100000) 22886eeac1d9SJulian Elischer used = 100000; 22896eeac1d9SJulian Elischer if (wasted > 100000) 22906eeac1d9SJulian Elischer wasted = 100000; 22916eeac1d9SJulian Elischer /* store in the appropriate bucket */ 22926eeac1d9SJulian Elischer /* don't bother locking. if it's slightly off, so what? */ 22936eeac1d9SJulian Elischer mbprof.segments[segments]++; 22946eeac1d9SJulian Elischer mbprof.used[fls(used)]++; 22956eeac1d9SJulian Elischer mbprof.wasted[fls(wasted)]++; 22966eeac1d9SJulian Elischer } 22976eeac1d9SJulian Elischer 2298c6c52d8eSAlexander Motin static int 2299c6c52d8eSAlexander Motin mbprof_handler(SYSCTL_HANDLER_ARGS) 23006eeac1d9SJulian Elischer { 2301c6c52d8eSAlexander Motin char buf[256]; 2302c6c52d8eSAlexander Motin struct sbuf sb; 2303c6c52d8eSAlexander Motin int error; 230460ae52f7SEd Schouten uint64_t *p; 23056eeac1d9SJulian Elischer 2306c6c52d8eSAlexander Motin sbuf_new_for_sysctl(&sb, buf, sizeof(buf), req); 2307c6c52d8eSAlexander Motin 23086eeac1d9SJulian Elischer p = &mbprof.wasted[0]; 2309c6c52d8eSAlexander Motin sbuf_printf(&sb, 23106eeac1d9SJulian Elischer "wasted:\n" 23112182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 23122182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 23136eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 23146eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 23156eeac1d9SJulian Elischer #ifdef BIG_ARRAY 23166eeac1d9SJulian Elischer p = &mbprof.wasted[16]; 2317c6c52d8eSAlexander Motin sbuf_printf(&sb, 23182182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 23192182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 23206eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 23216eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 23226eeac1d9SJulian Elischer #endif 23236eeac1d9SJulian Elischer p = &mbprof.used[0]; 2324c6c52d8eSAlexander Motin sbuf_printf(&sb, 23256eeac1d9SJulian Elischer "used:\n" 23262182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 23272182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 23286eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 23296eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 23306eeac1d9SJulian Elischer #ifdef BIG_ARRAY 23316eeac1d9SJulian Elischer p = &mbprof.used[16]; 2332c6c52d8eSAlexander Motin sbuf_printf(&sb, 23332182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 23342182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 23356eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 23366eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 23376eeac1d9SJulian Elischer #endif 23386eeac1d9SJulian Elischer p = &mbprof.segments[0]; 2339c6c52d8eSAlexander Motin sbuf_printf(&sb, 23406eeac1d9SJulian Elischer "segments:\n" 23412182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 23422182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 23436eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 23446eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 23456eeac1d9SJulian Elischer #ifdef BIG_ARRAY 23466eeac1d9SJulian Elischer p = &mbprof.segments[16]; 2347c6c52d8eSAlexander Motin sbuf_printf(&sb, 23482182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 23492182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %jju", 23506eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 23516eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 23526eeac1d9SJulian Elischer #endif 23536eeac1d9SJulian Elischer 2354c6c52d8eSAlexander Motin error = sbuf_finish(&sb); 2355c6c52d8eSAlexander Motin sbuf_delete(&sb); 23566eeac1d9SJulian Elischer return (error); 23576eeac1d9SJulian Elischer } 23586eeac1d9SJulian Elischer 23596eeac1d9SJulian Elischer static int 23606eeac1d9SJulian Elischer mbprof_clr_handler(SYSCTL_HANDLER_ARGS) 23616eeac1d9SJulian Elischer { 23626eeac1d9SJulian Elischer int clear, error; 23636eeac1d9SJulian Elischer 23646eeac1d9SJulian Elischer clear = 0; 23656eeac1d9SJulian Elischer error = sysctl_handle_int(oidp, &clear, 0, req); 23666eeac1d9SJulian Elischer if (error || !req->newptr) 23676eeac1d9SJulian Elischer return (error); 23686eeac1d9SJulian Elischer 23696eeac1d9SJulian Elischer if (clear) { 23706eeac1d9SJulian Elischer bzero(&mbprof, sizeof(mbprof)); 23716eeac1d9SJulian Elischer } 23726eeac1d9SJulian Elischer 23736eeac1d9SJulian Elischer return (error); 23746eeac1d9SJulian Elischer } 23756eeac1d9SJulian Elischer 23767029da5cSPawel Biernacki SYSCTL_PROC(_kern_ipc, OID_AUTO, mbufprofile, 2377c6c52d8eSAlexander Motin CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, 0, 23787029da5cSPawel Biernacki mbprof_handler, "A", 23797029da5cSPawel Biernacki "mbuf profiling statistics"); 23806eeac1d9SJulian Elischer 23817029da5cSPawel Biernacki SYSCTL_PROC(_kern_ipc, OID_AUTO, mbufprofileclr, 2382c6c52d8eSAlexander Motin CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_MPSAFE, NULL, 0, 23837029da5cSPawel Biernacki mbprof_clr_handler, "I", 23847029da5cSPawel Biernacki "clear mbuf profiling statistics"); 23856eeac1d9SJulian Elischer #endif 2386