19454b2d8SWarner Losh /*- 251369649SPedro F. Giffuni * SPDX-License-Identifier: BSD-3-Clause 351369649SPedro F. Giffuni * 4df8bae1dSRodney W. Grimes * Copyright (c) 1982, 1986, 1988, 1991, 1993 5df8bae1dSRodney W. Grimes * The Regents of the University of California. All rights reserved. 6df8bae1dSRodney W. Grimes * 7df8bae1dSRodney W. Grimes * Redistribution and use in source and binary forms, with or without 8df8bae1dSRodney W. Grimes * modification, are permitted provided that the following conditions 9df8bae1dSRodney W. Grimes * are met: 10df8bae1dSRodney W. Grimes * 1. Redistributions of source code must retain the above copyright 11df8bae1dSRodney W. Grimes * notice, this list of conditions and the following disclaimer. 12df8bae1dSRodney W. Grimes * 2. Redistributions in binary form must reproduce the above copyright 13df8bae1dSRodney W. Grimes * notice, this list of conditions and the following disclaimer in the 14df8bae1dSRodney W. Grimes * documentation and/or other materials provided with the distribution. 1569a28758SEd Maste * 3. Neither the name of the University nor the names of its contributors 16df8bae1dSRodney W. Grimes * may be used to endorse or promote products derived from this software 17df8bae1dSRodney W. Grimes * without specific prior written permission. 18df8bae1dSRodney W. Grimes * 19df8bae1dSRodney W. Grimes * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 20df8bae1dSRodney W. Grimes * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 21df8bae1dSRodney W. Grimes * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 22df8bae1dSRodney W. Grimes * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 23df8bae1dSRodney W. Grimes * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 24df8bae1dSRodney W. Grimes * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 25df8bae1dSRodney W. Grimes * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 26df8bae1dSRodney W. Grimes * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 27df8bae1dSRodney W. Grimes * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 28df8bae1dSRodney W. Grimes * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 29df8bae1dSRodney W. Grimes * SUCH DAMAGE. 30df8bae1dSRodney W. Grimes * 31df8bae1dSRodney W. Grimes * @(#)uipc_mbuf.c 8.2 (Berkeley) 1/4/94 32df8bae1dSRodney W. Grimes */ 33df8bae1dSRodney W. Grimes 34677b542eSDavid E. O'Brien #include <sys/cdefs.h> 35677b542eSDavid E. O'Brien __FBSDID("$FreeBSD$"); 36677b542eSDavid E. O'Brien 37240ef842SDavid E. O'Brien #include "opt_param.h" 38352d050eSMike Silbersack #include "opt_mbuf_stress_test.h" 396eeac1d9SJulian Elischer #include "opt_mbuf_profiling.h" 40e32a5b94SRobert Watson 41df8bae1dSRodney W. Grimes #include <sys/param.h> 42df8bae1dSRodney W. Grimes #include <sys/systm.h> 43fb919e4dSMark Murray #include <sys/kernel.h> 44beb699c7SMike Silbersack #include <sys/limits.h> 45fb919e4dSMark Murray #include <sys/lock.h> 46f9d0d524SRobert Watson #include <sys/malloc.h> 47df8bae1dSRodney W. Grimes #include <sys/mbuf.h> 48639acc13SGarrett Wollman #include <sys/sysctl.h> 49df8bae1dSRodney W. Grimes #include <sys/domain.h> 50df8bae1dSRodney W. Grimes #include <sys/protosw.h> 51beb699c7SMike Silbersack #include <sys/uio.h> 5282334850SJohn Baldwin #include <sys/vmmeter.h> 53480f4e94SGeorge V. Neville-Neil #include <sys/sdt.h> 5482334850SJohn Baldwin #include <vm/vm.h> 5582334850SJohn Baldwin #include <vm/vm_pageout.h> 5682334850SJohn Baldwin #include <vm/vm_page.h> 57480f4e94SGeorge V. Neville-Neil 58dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE5_XLATE(sdt, , , m__init, 59480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *", 60480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 61480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 62480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 63480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t"); 64480f4e94SGeorge V. Neville-Neil 65dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE3_XLATE(sdt, , , m__gethdr, 66480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 67480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 68480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 69480f4e94SGeorge V. Neville-Neil 70dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE3_XLATE(sdt, , , m__get, 71480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 72480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 73480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 74480f4e94SGeorge V. Neville-Neil 75dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE4_XLATE(sdt, , , m__getcl, 76480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 77480f4e94SGeorge V. Neville-Neil "uint16_t", "uint16_t", 78480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 79480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 80480f4e94SGeorge V. Neville-Neil 81edde7a53SAndrey V. Elsukov SDT_PROBE_DEFINE5_XLATE(sdt, , , m__getjcl, 82edde7a53SAndrey V. Elsukov "uint32_t", "uint32_t", 83edde7a53SAndrey V. Elsukov "uint16_t", "uint16_t", 84edde7a53SAndrey V. Elsukov "uint32_t", "uint32_t", 85edde7a53SAndrey V. Elsukov "uint32_t", "uint32_t", 86edde7a53SAndrey V. Elsukov "struct mbuf *", "mbufinfo_t *"); 87edde7a53SAndrey V. Elsukov 88dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE3_XLATE(sdt, , , m__clget, 89480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *", 90480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 91480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t"); 92480f4e94SGeorge V. Neville-Neil 93dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE4_XLATE(sdt, , , m__cljget, 94480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *", 95480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 96480f4e94SGeorge V. Neville-Neil "uint32_t", "uint32_t", 97480f4e94SGeorge V. Neville-Neil "void*", "void*"); 98480f4e94SGeorge V. Neville-Neil 99dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE(sdt, , , m__cljset); 100480f4e94SGeorge V. Neville-Neil 101dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE1_XLATE(sdt, , , m__free, 102480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 103480f4e94SGeorge V. Neville-Neil 104dcd070d8SGeorge V. Neville-Neil SDT_PROBE_DEFINE1_XLATE(sdt, , , m__freem, 105480f4e94SGeorge V. Neville-Neil "struct mbuf *", "mbufinfo_t *"); 106fb919e4dSMark Murray 1075e4bc63bSGleb Smirnoff #include <security/mac/mac_framework.h> 1085e4bc63bSGleb Smirnoff 10928f8db14SBruce Evans int max_linkhdr; 11028f8db14SBruce Evans int max_protohdr; 11128f8db14SBruce Evans int max_hdr; 11228f8db14SBruce Evans int max_datalen; 11351710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 11455e9f80dSMike Silbersack int m_defragpackets; 11555e9f80dSMike Silbersack int m_defragbytes; 11655e9f80dSMike Silbersack int m_defraguseless; 11755e9f80dSMike Silbersack int m_defragfailure; 118352d050eSMike Silbersack int m_defragrandomfailures; 119352d050eSMike Silbersack #endif 1207d032714SBosko Milekic 1217d032714SBosko Milekic /* 1227d032714SBosko Milekic * sysctl(8) exported objects 1237d032714SBosko Milekic */ 12480444f88SAndre Oppermann SYSCTL_INT(_kern_ipc, KIPC_MAX_LINKHDR, max_linkhdr, CTLFLAG_RD, 12580444f88SAndre Oppermann &max_linkhdr, 0, "Size of largest link layer header"); 12680444f88SAndre Oppermann SYSCTL_INT(_kern_ipc, KIPC_MAX_PROTOHDR, max_protohdr, CTLFLAG_RD, 12780444f88SAndre Oppermann &max_protohdr, 0, "Size of largest protocol layer header"); 12880444f88SAndre Oppermann SYSCTL_INT(_kern_ipc, KIPC_MAX_HDR, max_hdr, CTLFLAG_RD, 12980444f88SAndre Oppermann &max_hdr, 0, "Size of largest link plus protocol header"); 13080444f88SAndre Oppermann SYSCTL_INT(_kern_ipc, KIPC_MAX_DATALEN, max_datalen, CTLFLAG_RD, 13180444f88SAndre Oppermann &max_datalen, 0, "Minimum space left in mbuf after max_hdr"); 13251710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 13355e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragpackets, CTLFLAG_RD, 13455e9f80dSMike Silbersack &m_defragpackets, 0, ""); 13555e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragbytes, CTLFLAG_RD, 13655e9f80dSMike Silbersack &m_defragbytes, 0, ""); 13755e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defraguseless, CTLFLAG_RD, 13855e9f80dSMike Silbersack &m_defraguseless, 0, ""); 13955e9f80dSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragfailure, CTLFLAG_RD, 14055e9f80dSMike Silbersack &m_defragfailure, 0, ""); 141352d050eSMike Silbersack SYSCTL_INT(_kern_ipc, OID_AUTO, m_defragrandomfailures, CTLFLAG_RW, 142352d050eSMike Silbersack &m_defragrandomfailures, 0, ""); 143352d050eSMike Silbersack #endif 144df8bae1dSRodney W. Grimes 145df8bae1dSRodney W. Grimes /* 146f729ede6SAndre Oppermann * Ensure the correct size of various mbuf parameters. It could be off due 147f729ede6SAndre Oppermann * to compiler-induced padding and alignment artifacts. 148f729ede6SAndre Oppermann */ 149f729ede6SAndre Oppermann CTASSERT(MSIZE - offsetof(struct mbuf, m_dat) == MLEN); 150f729ede6SAndre Oppermann CTASSERT(MSIZE - offsetof(struct mbuf, m_pktdat) == MHLEN); 151f729ede6SAndre Oppermann 152f729ede6SAndre Oppermann /* 1533d1a9ed3SRobert Watson * mbuf data storage should be 64-bit aligned regardless of architectural 1543d1a9ed3SRobert Watson * pointer size; check this is the case with and without a packet header. 1553d1a9ed3SRobert Watson */ 1563d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_dat) % 8 == 0); 1573d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_pktdat) % 8 == 0); 1583d1a9ed3SRobert Watson 1593d1a9ed3SRobert Watson /* 1603d1a9ed3SRobert Watson * While the specific values here don't matter too much (i.e., +/- a few 1613d1a9ed3SRobert Watson * words), we do want to ensure that changes to these values are carefully 1623d1a9ed3SRobert Watson * reasoned about and properly documented. This is especially the case as 1633d1a9ed3SRobert Watson * network-protocol and device-driver modules encode these layouts, and must 1643d1a9ed3SRobert Watson * be recompiled if the structures change. Check these values at compile time 1653d1a9ed3SRobert Watson * against the ones documented in comments in mbuf.h. 1663d1a9ed3SRobert Watson * 1673d1a9ed3SRobert Watson * NB: Possibly they should be documented there via #define's and not just 1683d1a9ed3SRobert Watson * comments. 1693d1a9ed3SRobert Watson */ 1703d1a9ed3SRobert Watson #if defined(__LP64__) 1713d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_dat) == 32); 1723d1a9ed3SRobert Watson CTASSERT(sizeof(struct pkthdr) == 56); 1730c103266SGleb Smirnoff CTASSERT(sizeof(struct m_ext) == 160); 1743d1a9ed3SRobert Watson #else 1753d1a9ed3SRobert Watson CTASSERT(offsetof(struct mbuf, m_dat) == 24); 1763d1a9ed3SRobert Watson CTASSERT(sizeof(struct pkthdr) == 48); 177f57d153eSBrandon Bergren #if defined(__powerpc__) && defined(BOOKE) 178f57d153eSBrandon Bergren /* PowerPC booke has 64-bit physical pointers. */ 179f57d153eSBrandon Bergren CTASSERT(sizeof(struct m_ext) == 184); 180f57d153eSBrandon Bergren #else 1810c103266SGleb Smirnoff CTASSERT(sizeof(struct m_ext) == 180); 1823d1a9ed3SRobert Watson #endif 183f57d153eSBrandon Bergren #endif 1843d1a9ed3SRobert Watson 1853d1a9ed3SRobert Watson /* 186ec9d83ddSGleb Smirnoff * Assert that the queue(3) macros produce code of the same size as an old 187ec9d83ddSGleb Smirnoff * plain pointer does. 188ec9d83ddSGleb Smirnoff */ 189ec9d83ddSGleb Smirnoff #ifdef INVARIANTS 190d53d6fa9SMark Johnston static struct mbuf __used m_assertbuf; 191ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_slist) == sizeof(m_assertbuf.m_next)); 192ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_stailq) == sizeof(m_assertbuf.m_next)); 193ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_slistpkt) == sizeof(m_assertbuf.m_nextpkt)); 194ec9d83ddSGleb Smirnoff CTASSERT(sizeof(m_assertbuf.m_stailqpkt) == sizeof(m_assertbuf.m_nextpkt)); 195ec9d83ddSGleb Smirnoff #endif 196ec9d83ddSGleb Smirnoff 197ec9d83ddSGleb Smirnoff /* 1986bccea7cSRebecca Cran * Attach the cluster from *m to *n, set up m_ext in *n 19956a4e45aSAndre Oppermann * and bump the refcount of the cluster. 20056a4e45aSAndre Oppermann */ 20186a996e6SHiren Panchasara void 20256a5f52eSGleb Smirnoff mb_dupcl(struct mbuf *n, struct mbuf *m) 20356a4e45aSAndre Oppermann { 20456a5f52eSGleb Smirnoff volatile u_int *refcnt; 20556a4e45aSAndre Oppermann 20661664ee7SGleb Smirnoff KASSERT(m->m_flags & (M_EXT|M_EXTPG), 20761664ee7SGleb Smirnoff ("%s: M_EXT|M_EXTPG not set on %p", __func__, m)); 20861664ee7SGleb Smirnoff KASSERT(!(n->m_flags & (M_EXT|M_EXTPG)), 20961664ee7SGleb Smirnoff ("%s: M_EXT|M_EXTPG set on %p", __func__, n)); 2101fbe6a82SGleb Smirnoff 21107e87a1dSGleb Smirnoff /* 2120c103266SGleb Smirnoff * Cache access optimization. 2130c103266SGleb Smirnoff * 2140c103266SGleb Smirnoff * o Regular M_EXT storage doesn't need full copy of m_ext, since 2150c103266SGleb Smirnoff * the holder of the 'ext_count' is responsible to carry the free 2160c103266SGleb Smirnoff * routine and its arguments. 21761664ee7SGleb Smirnoff * o M_EXTPG data is split between main part of mbuf and m_ext, the 2180c103266SGleb Smirnoff * main part is copied in full, the m_ext part is similar to M_EXT. 2190c103266SGleb Smirnoff * o EXT_EXTREF, where 'ext_cnt' doesn't point into mbuf at all, is 2200c103266SGleb Smirnoff * special - it needs full copy of m_ext into each mbuf, since any 2210c103266SGleb Smirnoff * copy could end up as the last to free. 22207e87a1dSGleb Smirnoff */ 22361664ee7SGleb Smirnoff if (m->m_flags & M_EXTPG) { 2247b6c99d0SGleb Smirnoff bcopy(&m->m_epg_startcopy, &n->m_epg_startcopy, 2257b6c99d0SGleb Smirnoff __rangeof(struct mbuf, m_epg_startcopy, m_epg_endcopy)); 2267b6c99d0SGleb Smirnoff bcopy(&m->m_ext, &n->m_ext, m_epg_ext_copylen); 22761664ee7SGleb Smirnoff } else if (m->m_ext.ext_type == EXT_EXTREF) 2280c103266SGleb Smirnoff bcopy(&m->m_ext, &n->m_ext, sizeof(struct m_ext)); 22961664ee7SGleb Smirnoff else 23007e87a1dSGleb Smirnoff bcopy(&m->m_ext, &n->m_ext, m_ext_copylen); 2310c103266SGleb Smirnoff 23261664ee7SGleb Smirnoff n->m_flags |= m->m_flags & (M_RDONLY | M_EXT | M_EXTPG); 23356a5f52eSGleb Smirnoff 23456a5f52eSGleb Smirnoff /* See if this is the mbuf that holds the embedded refcount. */ 23556a5f52eSGleb Smirnoff if (m->m_ext.ext_flags & EXT_FLAG_EMBREF) { 23656a5f52eSGleb Smirnoff refcnt = n->m_ext.ext_cnt = &m->m_ext.ext_count; 23756a5f52eSGleb Smirnoff n->m_ext.ext_flags &= ~EXT_FLAG_EMBREF; 23856a5f52eSGleb Smirnoff } else { 23956a5f52eSGleb Smirnoff KASSERT(m->m_ext.ext_cnt != NULL, 24056a5f52eSGleb Smirnoff ("%s: no refcounting pointer on %p", __func__, m)); 24156a5f52eSGleb Smirnoff refcnt = m->m_ext.ext_cnt; 24256a5f52eSGleb Smirnoff } 24356a5f52eSGleb Smirnoff 24456a5f52eSGleb Smirnoff if (*refcnt == 1) 24556a5f52eSGleb Smirnoff *refcnt += 1; 24656a5f52eSGleb Smirnoff else 24756a5f52eSGleb Smirnoff atomic_add_int(refcnt, 1); 24856a4e45aSAndre Oppermann } 24956a4e45aSAndre Oppermann 2509523d1bfSNavdeep Parhar void 2519523d1bfSNavdeep Parhar m_demote_pkthdr(struct mbuf *m) 2529523d1bfSNavdeep Parhar { 2539523d1bfSNavdeep Parhar 2549523d1bfSNavdeep Parhar M_ASSERTPKTHDR(m); 2559523d1bfSNavdeep Parhar 2569523d1bfSNavdeep Parhar m_tag_delete_chain(m, NULL); 2579523d1bfSNavdeep Parhar m->m_flags &= ~M_PKTHDR; 2589523d1bfSNavdeep Parhar bzero(&m->m_pkthdr, sizeof(struct pkthdr)); 2599523d1bfSNavdeep Parhar } 2609523d1bfSNavdeep Parhar 26156a4e45aSAndre Oppermann /* 262ed111688SAndre Oppermann * Clean up mbuf (chain) from any tags and packet headers. 263e0068c3aSAndre Oppermann * If "all" is set then the first mbuf in the chain will be 264e0068c3aSAndre Oppermann * cleaned too. 265ed111688SAndre Oppermann */ 266ed111688SAndre Oppermann void 267651e4e6aSGleb Smirnoff m_demote(struct mbuf *m0, int all, int flags) 268ed111688SAndre Oppermann { 269ed111688SAndre Oppermann struct mbuf *m; 270ed111688SAndre Oppermann 271ed111688SAndre Oppermann for (m = all ? m0 : m0->m_next; m != NULL; m = m->m_next) { 2727ee2d058SGleb Smirnoff KASSERT(m->m_nextpkt == NULL, ("%s: m_nextpkt in m %p, m0 %p", 2737ee2d058SGleb Smirnoff __func__, m, m0)); 2749523d1bfSNavdeep Parhar if (m->m_flags & M_PKTHDR) 2759523d1bfSNavdeep Parhar m_demote_pkthdr(m); 27682334850SJohn Baldwin m->m_flags = m->m_flags & (M_EXT | M_RDONLY | M_NOFREE | 2776edfd179SGleb Smirnoff M_EXTPG | flags); 278ed111688SAndre Oppermann } 279ed111688SAndre Oppermann } 280ed111688SAndre Oppermann 281ed111688SAndre Oppermann /* 282fdcc028dSAndre Oppermann * Sanity checks on mbuf (chain) for use in KASSERT() and general 283fdcc028dSAndre Oppermann * debugging. 284fdcc028dSAndre Oppermann * Returns 0 or panics when bad and 1 on all tests passed. 285fdcc028dSAndre Oppermann * Sanitize, 0 to run M_SANITY_ACTION, 1 to garble things so they 286fdcc028dSAndre Oppermann * blow up later. 287a048affbSAndre Oppermann */ 288a048affbSAndre Oppermann int 289a048affbSAndre Oppermann m_sanity(struct mbuf *m0, int sanitize) 290a048affbSAndre Oppermann { 291a048affbSAndre Oppermann struct mbuf *m; 292a048affbSAndre Oppermann caddr_t a, b; 293a048affbSAndre Oppermann int pktlen = 0; 294a048affbSAndre Oppermann 29521ee3e7aSKip Macy #ifdef INVARIANTS 29621ee3e7aSKip Macy #define M_SANITY_ACTION(s) panic("mbuf %p: " s, m) 29721ee3e7aSKip Macy #else 29821ee3e7aSKip Macy #define M_SANITY_ACTION(s) printf("mbuf %p: " s, m) 29921ee3e7aSKip Macy #endif 300a048affbSAndre Oppermann 301fdcc028dSAndre Oppermann for (m = m0; m != NULL; m = m->m_next) { 302a048affbSAndre Oppermann /* 303a048affbSAndre Oppermann * Basic pointer checks. If any of these fails then some 304a048affbSAndre Oppermann * unrelated kernel memory before or after us is trashed. 305a048affbSAndre Oppermann * No way to recover from that. 306a048affbSAndre Oppermann */ 307b66f2a48SRobert Watson a = M_START(m); 308b66f2a48SRobert Watson b = a + M_SIZE(m); 309a048affbSAndre Oppermann if ((caddr_t)m->m_data < a) 310a048affbSAndre Oppermann M_SANITY_ACTION("m_data outside mbuf data range left"); 311a048affbSAndre Oppermann if ((caddr_t)m->m_data > b) 312a048affbSAndre Oppermann M_SANITY_ACTION("m_data outside mbuf data range right"); 313a048affbSAndre Oppermann if ((caddr_t)m->m_data + m->m_len > b) 314a048affbSAndre Oppermann M_SANITY_ACTION("m_data + m_len exeeds mbuf space"); 315a048affbSAndre Oppermann 316a048affbSAndre Oppermann /* m->m_nextpkt may only be set on first mbuf in chain. */ 317fdcc028dSAndre Oppermann if (m != m0 && m->m_nextpkt != NULL) { 318a048affbSAndre Oppermann if (sanitize) { 319a048affbSAndre Oppermann m_freem(m->m_nextpkt); 320a048affbSAndre Oppermann m->m_nextpkt = (struct mbuf *)0xDEADC0DE; 321a048affbSAndre Oppermann } else 322a048affbSAndre Oppermann M_SANITY_ACTION("m->m_nextpkt on in-chain mbuf"); 323a048affbSAndre Oppermann } 324a048affbSAndre Oppermann 325a048affbSAndre Oppermann /* packet length (not mbuf length!) calculation */ 326a048affbSAndre Oppermann if (m0->m_flags & M_PKTHDR) 327a048affbSAndre Oppermann pktlen += m->m_len; 328a048affbSAndre Oppermann 329a048affbSAndre Oppermann /* m_tags may only be attached to first mbuf in chain. */ 330a048affbSAndre Oppermann if (m != m0 && m->m_flags & M_PKTHDR && 331a048affbSAndre Oppermann !SLIST_EMPTY(&m->m_pkthdr.tags)) { 332a048affbSAndre Oppermann if (sanitize) { 333a048affbSAndre Oppermann m_tag_delete_chain(m, NULL); 334a048affbSAndre Oppermann /* put in 0xDEADC0DE perhaps? */ 335fdcc028dSAndre Oppermann } else 336a048affbSAndre Oppermann M_SANITY_ACTION("m_tags on in-chain mbuf"); 337a048affbSAndre Oppermann } 338a048affbSAndre Oppermann 339a048affbSAndre Oppermann /* M_PKTHDR may only be set on first mbuf in chain */ 340a048affbSAndre Oppermann if (m != m0 && m->m_flags & M_PKTHDR) { 341a048affbSAndre Oppermann if (sanitize) { 342a048affbSAndre Oppermann bzero(&m->m_pkthdr, sizeof(m->m_pkthdr)); 343a048affbSAndre Oppermann m->m_flags &= ~M_PKTHDR; 344a048affbSAndre Oppermann /* put in 0xDEADCODE and leave hdr flag in */ 345a048affbSAndre Oppermann } else 346a048affbSAndre Oppermann M_SANITY_ACTION("M_PKTHDR on in-chain mbuf"); 347a048affbSAndre Oppermann } 348a048affbSAndre Oppermann } 349fdcc028dSAndre Oppermann m = m0; 350fdcc028dSAndre Oppermann if (pktlen && pktlen != m->m_pkthdr.len) { 351a048affbSAndre Oppermann if (sanitize) 352fdcc028dSAndre Oppermann m->m_pkthdr.len = 0; 353a048affbSAndre Oppermann else 354a048affbSAndre Oppermann M_SANITY_ACTION("m_pkthdr.len != mbuf chain length"); 355a048affbSAndre Oppermann } 356a048affbSAndre Oppermann return 1; 357fdcc028dSAndre Oppermann 358fdcc028dSAndre Oppermann #undef M_SANITY_ACTION 359a048affbSAndre Oppermann } 360a048affbSAndre Oppermann 3615e4bc63bSGleb Smirnoff /* 3625e4bc63bSGleb Smirnoff * Non-inlined part of m_init(). 3635e4bc63bSGleb Smirnoff */ 3645e4bc63bSGleb Smirnoff int 3655e4bc63bSGleb Smirnoff m_pkthdr_init(struct mbuf *m, int how) 3665e4bc63bSGleb Smirnoff { 3675e4bc63bSGleb Smirnoff #ifdef MAC 3685e4bc63bSGleb Smirnoff int error; 3695e4bc63bSGleb Smirnoff #endif 3705e4bc63bSGleb Smirnoff m->m_data = m->m_pktdat; 3715e4bc63bSGleb Smirnoff bzero(&m->m_pkthdr, sizeof(m->m_pkthdr)); 37250575ce1SAndrew Gallatin #ifdef NUMA 37350575ce1SAndrew Gallatin m->m_pkthdr.numa_domain = M_NODOM; 37450575ce1SAndrew Gallatin #endif 3755e4bc63bSGleb Smirnoff #ifdef MAC 3765e4bc63bSGleb Smirnoff /* If the label init fails, fail the alloc */ 3775e4bc63bSGleb Smirnoff error = mac_mbuf_init(m, how); 3785e4bc63bSGleb Smirnoff if (error) 3795e4bc63bSGleb Smirnoff return (error); 3805e4bc63bSGleb Smirnoff #endif 3815e4bc63bSGleb Smirnoff 3825e4bc63bSGleb Smirnoff return (0); 3835e4bc63bSGleb Smirnoff } 384a048affbSAndre Oppermann 385a048affbSAndre Oppermann /* 3869967cafcSSam Leffler * "Move" mbuf pkthdr from "from" to "to". 387e37b1fcdSRobert Watson * "from" must have M_PKTHDR set, and "to" must be empty. 388e37b1fcdSRobert Watson */ 389e37b1fcdSRobert Watson void 3909967cafcSSam Leffler m_move_pkthdr(struct mbuf *to, struct mbuf *from) 391e37b1fcdSRobert Watson { 392e37b1fcdSRobert Watson 393e37b1fcdSRobert Watson #if 0 3949967cafcSSam Leffler /* see below for why these are not enabled */ 395fe584538SDag-Erling Smørgrav M_ASSERTPKTHDR(to); 396225bff6fSRobert Watson /* Note: with MAC, this may not be a good assertion. */ 3979967cafcSSam Leffler KASSERT(SLIST_EMPTY(&to->m_pkthdr.tags), 3989967cafcSSam Leffler ("m_move_pkthdr: to has tags")); 399e37b1fcdSRobert Watson #endif 400e32a5b94SRobert Watson #ifdef MAC 401225bff6fSRobert Watson /* 402225bff6fSRobert Watson * XXXMAC: It could be this should also occur for non-MAC? 403225bff6fSRobert Watson */ 404e32a5b94SRobert Watson if (to->m_flags & M_PKTHDR) 405225bff6fSRobert Watson m_tag_delete_chain(to, NULL); 406e32a5b94SRobert Watson #endif 40782334850SJohn Baldwin to->m_flags = (from->m_flags & M_COPYFLAGS) | 4086edfd179SGleb Smirnoff (to->m_flags & (M_EXT | M_EXTPG)); 409a4e71429SSam Leffler if ((to->m_flags & M_EXT) == 0) 4109967cafcSSam Leffler to->m_data = to->m_pktdat; 4119967cafcSSam Leffler to->m_pkthdr = from->m_pkthdr; /* especially tags */ 4129967cafcSSam Leffler SLIST_INIT(&from->m_pkthdr.tags); /* purge tags from src */ 4139967cafcSSam Leffler from->m_flags &= ~M_PKTHDR; 414fb3bc596SJohn Baldwin if (from->m_pkthdr.csum_flags & CSUM_SND_TAG) { 415fb3bc596SJohn Baldwin from->m_pkthdr.csum_flags &= ~CSUM_SND_TAG; 416fb3bc596SJohn Baldwin from->m_pkthdr.snd_tag = NULL; 417fb3bc596SJohn Baldwin } 4189967cafcSSam Leffler } 4199967cafcSSam Leffler 4209967cafcSSam Leffler /* 4219967cafcSSam Leffler * Duplicate "from"'s mbuf pkthdr in "to". 4229967cafcSSam Leffler * "from" must have M_PKTHDR set, and "to" must be empty. 4239967cafcSSam Leffler * In particular, this does a deep copy of the packet tags. 4249967cafcSSam Leffler */ 4259967cafcSSam Leffler int 4260cbefd30SAlexander V. Chernikov m_dup_pkthdr(struct mbuf *to, const struct mbuf *from, int how) 4279967cafcSSam Leffler { 4289967cafcSSam Leffler 4299967cafcSSam Leffler #if 0 4309967cafcSSam Leffler /* 4319967cafcSSam Leffler * The mbuf allocator only initializes the pkthdr 432c95be8b5SGleb Smirnoff * when the mbuf is allocated with m_gethdr(). Many users 433c95be8b5SGleb Smirnoff * (e.g. m_copy*, m_prepend) use m_get() and then 4349967cafcSSam Leffler * smash the pkthdr as needed causing these 4359967cafcSSam Leffler * assertions to trip. For now just disable them. 4369967cafcSSam Leffler */ 437fe584538SDag-Erling Smørgrav M_ASSERTPKTHDR(to); 438225bff6fSRobert Watson /* Note: with MAC, this may not be a good assertion. */ 4399967cafcSSam Leffler KASSERT(SLIST_EMPTY(&to->m_pkthdr.tags), ("m_dup_pkthdr: to has tags")); 4409967cafcSSam Leffler #endif 441063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 4429967cafcSSam Leffler #ifdef MAC 4439967cafcSSam Leffler if (to->m_flags & M_PKTHDR) 444225bff6fSRobert Watson m_tag_delete_chain(to, NULL); 4459967cafcSSam Leffler #endif 44682334850SJohn Baldwin to->m_flags = (from->m_flags & M_COPYFLAGS) | 4476edfd179SGleb Smirnoff (to->m_flags & (M_EXT | M_EXTPG)); 448df8c7fc9SMike Silbersack if ((to->m_flags & M_EXT) == 0) 4499967cafcSSam Leffler to->m_data = to->m_pktdat; 450e37b1fcdSRobert Watson to->m_pkthdr = from->m_pkthdr; 451fb3bc596SJohn Baldwin if (from->m_pkthdr.csum_flags & CSUM_SND_TAG) 452fb3bc596SJohn Baldwin m_snd_tag_ref(from->m_pkthdr.snd_tag); 4539967cafcSSam Leffler SLIST_INIT(&to->m_pkthdr.tags); 45494985f74SGleb Smirnoff return (m_tag_copy_chain(to, from, how)); 455e37b1fcdSRobert Watson } 456e37b1fcdSRobert Watson 457e37b1fcdSRobert Watson /* 458df8bae1dSRodney W. Grimes * Lesser-used path for M_PREPEND: 459df8bae1dSRodney W. Grimes * allocate new mbuf to prepend to chain, 460df8bae1dSRodney W. Grimes * copy junk along. 461df8bae1dSRodney W. Grimes */ 462df8bae1dSRodney W. Grimes struct mbuf * 463122a814aSBosko Milekic m_prepend(struct mbuf *m, int len, int how) 464df8bae1dSRodney W. Grimes { 465df8bae1dSRodney W. Grimes struct mbuf *mn; 466df8bae1dSRodney W. Grimes 467f8bf8e39SMike Silbersack if (m->m_flags & M_PKTHDR) 468c95be8b5SGleb Smirnoff mn = m_gethdr(how, m->m_type); 469f8bf8e39SMike Silbersack else 470c95be8b5SGleb Smirnoff mn = m_get(how, m->m_type); 471122a814aSBosko Milekic if (mn == NULL) { 472df8bae1dSRodney W. Grimes m_freem(m); 473122a814aSBosko Milekic return (NULL); 474df8bae1dSRodney W. Grimes } 475225bff6fSRobert Watson if (m->m_flags & M_PKTHDR) 476c95be8b5SGleb Smirnoff m_move_pkthdr(mn, m); 477df8bae1dSRodney W. Grimes mn->m_next = m; 478df8bae1dSRodney W. Grimes m = mn; 479ed6a66caSRobert Watson if (len < M_SIZE(m)) 4805288989fSRandall Stewart M_ALIGN(m, len); 481df8bae1dSRodney W. Grimes m->m_len = len; 482df8bae1dSRodney W. Grimes return (m); 483df8bae1dSRodney W. Grimes } 484df8bae1dSRodney W. Grimes 485df8bae1dSRodney W. Grimes /* 486df8bae1dSRodney W. Grimes * Make a copy of an mbuf chain starting "off0" bytes from the beginning, 487df8bae1dSRodney W. Grimes * continuing for "len" bytes. If len is M_COPYALL, copy to end of mbuf. 488eb1b1807SGleb Smirnoff * The wait parameter is a choice of M_WAITOK/M_NOWAIT from caller. 4891c38f2eaSArchie Cobbs * Note that the copy is read-only, because clusters are not copied, 4901c38f2eaSArchie Cobbs * only their reference counts are incremented. 491df8bae1dSRodney W. Grimes */ 492df8bae1dSRodney W. Grimes struct mbuf * 49356a5f52eSGleb Smirnoff m_copym(struct mbuf *m, int off0, int len, int wait) 494df8bae1dSRodney W. Grimes { 495122a814aSBosko Milekic struct mbuf *n, **np; 496122a814aSBosko Milekic int off = off0; 497df8bae1dSRodney W. Grimes struct mbuf *top; 498df8bae1dSRodney W. Grimes int copyhdr = 0; 499df8bae1dSRodney W. Grimes 500e0a653ddSAlfred Perlstein KASSERT(off >= 0, ("m_copym, negative off %d", off)); 501e0a653ddSAlfred Perlstein KASSERT(len >= 0, ("m_copym, negative len %d", len)); 502063d8114SAlfred Perlstein MBUF_CHECKSLEEP(wait); 503df8bae1dSRodney W. Grimes if (off == 0 && m->m_flags & M_PKTHDR) 504df8bae1dSRodney W. Grimes copyhdr = 1; 505df8bae1dSRodney W. Grimes while (off > 0) { 506e0a653ddSAlfred Perlstein KASSERT(m != NULL, ("m_copym, offset > size of mbuf chain")); 507df8bae1dSRodney W. Grimes if (off < m->m_len) 508df8bae1dSRodney W. Grimes break; 509df8bae1dSRodney W. Grimes off -= m->m_len; 510df8bae1dSRodney W. Grimes m = m->m_next; 511df8bae1dSRodney W. Grimes } 512df8bae1dSRodney W. Grimes np = ⊤ 513b85f65afSPedro F. Giffuni top = NULL; 514df8bae1dSRodney W. Grimes while (len > 0) { 515122a814aSBosko Milekic if (m == NULL) { 516e0a653ddSAlfred Perlstein KASSERT(len == M_COPYALL, 517e0a653ddSAlfred Perlstein ("m_copym, length > size of mbuf chain")); 518df8bae1dSRodney W. Grimes break; 519df8bae1dSRodney W. Grimes } 520f8bf8e39SMike Silbersack if (copyhdr) 521c95be8b5SGleb Smirnoff n = m_gethdr(wait, m->m_type); 522f8bf8e39SMike Silbersack else 523c95be8b5SGleb Smirnoff n = m_get(wait, m->m_type); 524df8bae1dSRodney W. Grimes *np = n; 525122a814aSBosko Milekic if (n == NULL) 526df8bae1dSRodney W. Grimes goto nospace; 527df8bae1dSRodney W. Grimes if (copyhdr) { 5289967cafcSSam Leffler if (!m_dup_pkthdr(n, m, wait)) 5299967cafcSSam Leffler goto nospace; 530df8bae1dSRodney W. Grimes if (len == M_COPYALL) 531df8bae1dSRodney W. Grimes n->m_pkthdr.len -= off0; 532df8bae1dSRodney W. Grimes else 533df8bae1dSRodney W. Grimes n->m_pkthdr.len = len; 534df8bae1dSRodney W. Grimes copyhdr = 0; 535df8bae1dSRodney W. Grimes } 536df8bae1dSRodney W. Grimes n->m_len = min(len, m->m_len - off); 53761664ee7SGleb Smirnoff if (m->m_flags & (M_EXT|M_EXTPG)) { 538df8bae1dSRodney W. Grimes n->m_data = m->m_data + off; 53956a4e45aSAndre Oppermann mb_dupcl(n, m); 540df8bae1dSRodney W. Grimes } else 541df8bae1dSRodney W. Grimes bcopy(mtod(m, caddr_t)+off, mtod(n, caddr_t), 542bd395ae8SBosko Milekic (u_int)n->m_len); 543df8bae1dSRodney W. Grimes if (len != M_COPYALL) 544df8bae1dSRodney W. Grimes len -= n->m_len; 545df8bae1dSRodney W. Grimes off = 0; 546df8bae1dSRodney W. Grimes m = m->m_next; 547df8bae1dSRodney W. Grimes np = &n->m_next; 548df8bae1dSRodney W. Grimes } 54908442f8aSBosko Milekic 550df8bae1dSRodney W. Grimes return (top); 551df8bae1dSRodney W. Grimes nospace: 552df8bae1dSRodney W. Grimes m_freem(top); 553122a814aSBosko Milekic return (NULL); 554df8bae1dSRodney W. Grimes } 555df8bae1dSRodney W. Grimes 556df8bae1dSRodney W. Grimes /* 5576a06dea0SGarrett Wollman * Copy an entire packet, including header (which must be present). 5586a06dea0SGarrett Wollman * An optimization of the common case `m_copym(m, 0, M_COPYALL, how)'. 5591c38f2eaSArchie Cobbs * Note that the copy is read-only, because clusters are not copied, 5601c38f2eaSArchie Cobbs * only their reference counts are incremented. 5615fe86675SLuigi Rizzo * Preserve alignment of the first mbuf so if the creator has left 5625fe86675SLuigi Rizzo * some room at the beginning (e.g. for inserting protocol headers) 5635fe86675SLuigi Rizzo * the copies still have the room available. 5646a06dea0SGarrett Wollman */ 5656a06dea0SGarrett Wollman struct mbuf * 566122a814aSBosko Milekic m_copypacket(struct mbuf *m, int how) 5676a06dea0SGarrett Wollman { 5686a06dea0SGarrett Wollman struct mbuf *top, *n, *o; 5696a06dea0SGarrett Wollman 570063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 571c95be8b5SGleb Smirnoff n = m_get(how, m->m_type); 5726a06dea0SGarrett Wollman top = n; 573122a814aSBosko Milekic if (n == NULL) 5746a06dea0SGarrett Wollman goto nospace; 5756a06dea0SGarrett Wollman 5769967cafcSSam Leffler if (!m_dup_pkthdr(n, m, how)) 5779967cafcSSam Leffler goto nospace; 5786a06dea0SGarrett Wollman n->m_len = m->m_len; 57961664ee7SGleb Smirnoff if (m->m_flags & (M_EXT|M_EXTPG)) { 5806a06dea0SGarrett Wollman n->m_data = m->m_data; 58156a4e45aSAndre Oppermann mb_dupcl(n, m); 5826a06dea0SGarrett Wollman } else { 5835fe86675SLuigi Rizzo n->m_data = n->m_pktdat + (m->m_data - m->m_pktdat ); 5846a06dea0SGarrett Wollman bcopy(mtod(m, char *), mtod(n, char *), n->m_len); 5856a06dea0SGarrett Wollman } 5866a06dea0SGarrett Wollman 5876a06dea0SGarrett Wollman m = m->m_next; 5886a06dea0SGarrett Wollman while (m) { 589c95be8b5SGleb Smirnoff o = m_get(how, m->m_type); 590122a814aSBosko Milekic if (o == NULL) 5916a06dea0SGarrett Wollman goto nospace; 5926a06dea0SGarrett Wollman 5936a06dea0SGarrett Wollman n->m_next = o; 5946a06dea0SGarrett Wollman n = n->m_next; 5956a06dea0SGarrett Wollman 5966a06dea0SGarrett Wollman n->m_len = m->m_len; 59761664ee7SGleb Smirnoff if (m->m_flags & (M_EXT|M_EXTPG)) { 5986a06dea0SGarrett Wollman n->m_data = m->m_data; 59956a4e45aSAndre Oppermann mb_dupcl(n, m); 6006a06dea0SGarrett Wollman } else { 6016a06dea0SGarrett Wollman bcopy(mtod(m, char *), mtod(n, char *), n->m_len); 6026a06dea0SGarrett Wollman } 6036a06dea0SGarrett Wollman 6046a06dea0SGarrett Wollman m = m->m_next; 6056a06dea0SGarrett Wollman } 6066a06dea0SGarrett Wollman return top; 6076a06dea0SGarrett Wollman nospace: 6086a06dea0SGarrett Wollman m_freem(top); 609122a814aSBosko Milekic return (NULL); 6106a06dea0SGarrett Wollman } 6116a06dea0SGarrett Wollman 61282334850SJohn Baldwin static void 61382334850SJohn Baldwin m_copyfromunmapped(const struct mbuf *m, int off, int len, caddr_t cp) 61482334850SJohn Baldwin { 61582334850SJohn Baldwin struct iovec iov; 61682334850SJohn Baldwin struct uio uio; 61782334850SJohn Baldwin int error; 61882334850SJohn Baldwin 61982334850SJohn Baldwin KASSERT(off >= 0, ("m_copyfromunmapped: negative off %d", off)); 62082334850SJohn Baldwin KASSERT(len >= 0, ("m_copyfromunmapped: negative len %d", len)); 62182334850SJohn Baldwin KASSERT(off < m->m_len, 62282334850SJohn Baldwin ("m_copyfromunmapped: len exceeds mbuf length")); 62382334850SJohn Baldwin iov.iov_base = cp; 62482334850SJohn Baldwin iov.iov_len = len; 62582334850SJohn Baldwin uio.uio_resid = len; 62682334850SJohn Baldwin uio.uio_iov = &iov; 62782334850SJohn Baldwin uio.uio_segflg = UIO_SYSSPACE; 62882334850SJohn Baldwin uio.uio_iovcnt = 1; 62982334850SJohn Baldwin uio.uio_offset = 0; 63082334850SJohn Baldwin uio.uio_rw = UIO_READ; 63182334850SJohn Baldwin error = m_unmappedtouio(m, off, &uio, len); 63282334850SJohn Baldwin KASSERT(error == 0, ("m_unmappedtouio failed: off %d, len %d", off, 63382334850SJohn Baldwin len)); 63482334850SJohn Baldwin } 63582334850SJohn Baldwin 6366a06dea0SGarrett Wollman /* 637df8bae1dSRodney W. Grimes * Copy data from an mbuf chain starting "off" bytes from the beginning, 638df8bae1dSRodney W. Grimes * continuing for "len" bytes, into the indicated buffer. 639df8bae1dSRodney W. Grimes */ 64026f9a767SRodney W. Grimes void 641a8cfc0eeSJulian Elischer m_copydata(const struct mbuf *m, int off, int len, caddr_t cp) 642df8bae1dSRodney W. Grimes { 643bd395ae8SBosko Milekic u_int count; 644df8bae1dSRodney W. Grimes 645e0a653ddSAlfred Perlstein KASSERT(off >= 0, ("m_copydata, negative off %d", off)); 646e0a653ddSAlfred Perlstein KASSERT(len >= 0, ("m_copydata, negative len %d", len)); 647df8bae1dSRodney W. Grimes while (off > 0) { 648e0a653ddSAlfred Perlstein KASSERT(m != NULL, ("m_copydata, offset > size of mbuf chain")); 649df8bae1dSRodney W. Grimes if (off < m->m_len) 650df8bae1dSRodney W. Grimes break; 651df8bae1dSRodney W. Grimes off -= m->m_len; 652df8bae1dSRodney W. Grimes m = m->m_next; 653df8bae1dSRodney W. Grimes } 654df8bae1dSRodney W. Grimes while (len > 0) { 655e0a653ddSAlfred Perlstein KASSERT(m != NULL, ("m_copydata, length > size of mbuf chain")); 656df8bae1dSRodney W. Grimes count = min(m->m_len - off, len); 6576edfd179SGleb Smirnoff if ((m->m_flags & M_EXTPG) != 0) 65882334850SJohn Baldwin m_copyfromunmapped(m, off, count, cp); 65982334850SJohn Baldwin else 660df8bae1dSRodney W. Grimes bcopy(mtod(m, caddr_t) + off, cp, count); 661df8bae1dSRodney W. Grimes len -= count; 662df8bae1dSRodney W. Grimes cp += count; 663df8bae1dSRodney W. Grimes off = 0; 664df8bae1dSRodney W. Grimes m = m->m_next; 665df8bae1dSRodney W. Grimes } 666df8bae1dSRodney W. Grimes } 667df8bae1dSRodney W. Grimes 668df8bae1dSRodney W. Grimes /* 6691c38f2eaSArchie Cobbs * Copy a packet header mbuf chain into a completely new chain, including 6701c38f2eaSArchie Cobbs * copying any mbuf clusters. Use this instead of m_copypacket() when 6711c38f2eaSArchie Cobbs * you need a writable copy of an mbuf chain. 6721c38f2eaSArchie Cobbs */ 6731c38f2eaSArchie Cobbs struct mbuf * 6740cbefd30SAlexander V. Chernikov m_dup(const struct mbuf *m, int how) 6751c38f2eaSArchie Cobbs { 6761c38f2eaSArchie Cobbs struct mbuf **p, *top = NULL; 6771c38f2eaSArchie Cobbs int remain, moff, nsize; 6781c38f2eaSArchie Cobbs 679063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 6801c38f2eaSArchie Cobbs /* Sanity check */ 6811c38f2eaSArchie Cobbs if (m == NULL) 682122a814aSBosko Milekic return (NULL); 683fe584538SDag-Erling Smørgrav M_ASSERTPKTHDR(m); 6841c38f2eaSArchie Cobbs 6851c38f2eaSArchie Cobbs /* While there's more data, get a new mbuf, tack it on, and fill it */ 6861c38f2eaSArchie Cobbs remain = m->m_pkthdr.len; 6871c38f2eaSArchie Cobbs moff = 0; 6881c38f2eaSArchie Cobbs p = ⊤ 6891c38f2eaSArchie Cobbs while (remain > 0 || top == NULL) { /* allow m->m_pkthdr.len == 0 */ 6901c38f2eaSArchie Cobbs struct mbuf *n; 6911c38f2eaSArchie Cobbs 6921c38f2eaSArchie Cobbs /* Get the next new mbuf */ 693099a0e58SBosko Milekic if (remain >= MINCLSIZE) { 694099a0e58SBosko Milekic n = m_getcl(how, m->m_type, 0); 695099a0e58SBosko Milekic nsize = MCLBYTES; 696099a0e58SBosko Milekic } else { 697099a0e58SBosko Milekic n = m_get(how, m->m_type); 698099a0e58SBosko Milekic nsize = MLEN; 699099a0e58SBosko Milekic } 7001c38f2eaSArchie Cobbs if (n == NULL) 7011c38f2eaSArchie Cobbs goto nospace; 702099a0e58SBosko Milekic 703099a0e58SBosko Milekic if (top == NULL) { /* First one, must be PKTHDR */ 704099a0e58SBosko Milekic if (!m_dup_pkthdr(n, m, how)) { 705099a0e58SBosko Milekic m_free(n); 7061c38f2eaSArchie Cobbs goto nospace; 7071c38f2eaSArchie Cobbs } 70863e6f390SEd Maste if ((n->m_flags & M_EXT) == 0) 709099a0e58SBosko Milekic nsize = MHLEN; 710089bb672SAndrey V. Elsukov n->m_flags &= ~M_RDONLY; 7111c38f2eaSArchie Cobbs } 7121c38f2eaSArchie Cobbs n->m_len = 0; 7131c38f2eaSArchie Cobbs 7141c38f2eaSArchie Cobbs /* Link it into the new chain */ 7151c38f2eaSArchie Cobbs *p = n; 7161c38f2eaSArchie Cobbs p = &n->m_next; 7171c38f2eaSArchie Cobbs 7181c38f2eaSArchie Cobbs /* Copy data from original mbuf(s) into new mbuf */ 7191c38f2eaSArchie Cobbs while (n->m_len < nsize && m != NULL) { 7201c38f2eaSArchie Cobbs int chunk = min(nsize - n->m_len, m->m_len - moff); 7211c38f2eaSArchie Cobbs 7221c38f2eaSArchie Cobbs bcopy(m->m_data + moff, n->m_data + n->m_len, chunk); 7231c38f2eaSArchie Cobbs moff += chunk; 7241c38f2eaSArchie Cobbs n->m_len += chunk; 7251c38f2eaSArchie Cobbs remain -= chunk; 7261c38f2eaSArchie Cobbs if (moff == m->m_len) { 7271c38f2eaSArchie Cobbs m = m->m_next; 7281c38f2eaSArchie Cobbs moff = 0; 7291c38f2eaSArchie Cobbs } 7301c38f2eaSArchie Cobbs } 7311c38f2eaSArchie Cobbs 7321c38f2eaSArchie Cobbs /* Check correct total mbuf length */ 7331c38f2eaSArchie Cobbs KASSERT((remain > 0 && m != NULL) || (remain == 0 && m == NULL), 734a48740b6SDavid E. O'Brien ("%s: bogus m_pkthdr.len", __func__)); 7351c38f2eaSArchie Cobbs } 7361c38f2eaSArchie Cobbs return (top); 7371c38f2eaSArchie Cobbs 7381c38f2eaSArchie Cobbs nospace: 7391c38f2eaSArchie Cobbs m_freem(top); 740122a814aSBosko Milekic return (NULL); 7411c38f2eaSArchie Cobbs } 7421c38f2eaSArchie Cobbs 7431c38f2eaSArchie Cobbs /* 744df8bae1dSRodney W. Grimes * Concatenate mbuf chain n to m. 745df8bae1dSRodney W. Grimes * Both chains must be of the same type (e.g. MT_DATA). 746df8bae1dSRodney W. Grimes * Any m_pkthdr is not updated. 747df8bae1dSRodney W. Grimes */ 74826f9a767SRodney W. Grimes void 749122a814aSBosko Milekic m_cat(struct mbuf *m, struct mbuf *n) 750df8bae1dSRodney W. Grimes { 751df8bae1dSRodney W. Grimes while (m->m_next) 752df8bae1dSRodney W. Grimes m = m->m_next; 753df8bae1dSRodney W. Grimes while (n) { 75414d7c5b1SAndre Oppermann if (!M_WRITABLE(m) || 7556edfd179SGleb Smirnoff (n->m_flags & M_EXTPG) != 0 || 75614d7c5b1SAndre Oppermann M_TRAILINGSPACE(m) < n->m_len) { 757df8bae1dSRodney W. Grimes /* just join the two chains */ 758df8bae1dSRodney W. Grimes m->m_next = n; 759df8bae1dSRodney W. Grimes return; 760df8bae1dSRodney W. Grimes } 761df8bae1dSRodney W. Grimes /* splat the data from one into the other */ 762df8bae1dSRodney W. Grimes bcopy(mtod(n, caddr_t), mtod(m, caddr_t) + m->m_len, 763df8bae1dSRodney W. Grimes (u_int)n->m_len); 764df8bae1dSRodney W. Grimes m->m_len += n->m_len; 765df8bae1dSRodney W. Grimes n = m_free(n); 766df8bae1dSRodney W. Grimes } 767df8bae1dSRodney W. Grimes } 768df8bae1dSRodney W. Grimes 7691967edbaSGleb Smirnoff /* 7701967edbaSGleb Smirnoff * Concatenate two pkthdr mbuf chains. 7711967edbaSGleb Smirnoff */ 7721967edbaSGleb Smirnoff void 7731967edbaSGleb Smirnoff m_catpkt(struct mbuf *m, struct mbuf *n) 7741967edbaSGleb Smirnoff { 7751967edbaSGleb Smirnoff 7761967edbaSGleb Smirnoff M_ASSERTPKTHDR(m); 7771967edbaSGleb Smirnoff M_ASSERTPKTHDR(n); 7781967edbaSGleb Smirnoff 7791967edbaSGleb Smirnoff m->m_pkthdr.len += n->m_pkthdr.len; 780651e4e6aSGleb Smirnoff m_demote(n, 1, 0); 7811967edbaSGleb Smirnoff 7821967edbaSGleb Smirnoff m_cat(m, n); 7831967edbaSGleb Smirnoff } 7841967edbaSGleb Smirnoff 78526f9a767SRodney W. Grimes void 786122a814aSBosko Milekic m_adj(struct mbuf *mp, int req_len) 787df8bae1dSRodney W. Grimes { 788122a814aSBosko Milekic int len = req_len; 789122a814aSBosko Milekic struct mbuf *m; 790122a814aSBosko Milekic int count; 791df8bae1dSRodney W. Grimes 792df8bae1dSRodney W. Grimes if ((m = mp) == NULL) 793df8bae1dSRodney W. Grimes return; 794df8bae1dSRodney W. Grimes if (len >= 0) { 795df8bae1dSRodney W. Grimes /* 796df8bae1dSRodney W. Grimes * Trim from head. 797df8bae1dSRodney W. Grimes */ 798df8bae1dSRodney W. Grimes while (m != NULL && len > 0) { 799df8bae1dSRodney W. Grimes if (m->m_len <= len) { 800df8bae1dSRodney W. Grimes len -= m->m_len; 801df8bae1dSRodney W. Grimes m->m_len = 0; 802df8bae1dSRodney W. Grimes m = m->m_next; 803df8bae1dSRodney W. Grimes } else { 804df8bae1dSRodney W. Grimes m->m_len -= len; 805df8bae1dSRodney W. Grimes m->m_data += len; 806df8bae1dSRodney W. Grimes len = 0; 807df8bae1dSRodney W. Grimes } 808df8bae1dSRodney W. Grimes } 809df8bae1dSRodney W. Grimes if (mp->m_flags & M_PKTHDR) 810a83baab6SMarko Zec mp->m_pkthdr.len -= (req_len - len); 811df8bae1dSRodney W. Grimes } else { 812df8bae1dSRodney W. Grimes /* 813df8bae1dSRodney W. Grimes * Trim from tail. Scan the mbuf chain, 814df8bae1dSRodney W. Grimes * calculating its length and finding the last mbuf. 815df8bae1dSRodney W. Grimes * If the adjustment only affects this mbuf, then just 816df8bae1dSRodney W. Grimes * adjust and return. Otherwise, rescan and truncate 817df8bae1dSRodney W. Grimes * after the remaining size. 818df8bae1dSRodney W. Grimes */ 819df8bae1dSRodney W. Grimes len = -len; 820df8bae1dSRodney W. Grimes count = 0; 821df8bae1dSRodney W. Grimes for (;;) { 822df8bae1dSRodney W. Grimes count += m->m_len; 823df8bae1dSRodney W. Grimes if (m->m_next == (struct mbuf *)0) 824df8bae1dSRodney W. Grimes break; 825df8bae1dSRodney W. Grimes m = m->m_next; 826df8bae1dSRodney W. Grimes } 827df8bae1dSRodney W. Grimes if (m->m_len >= len) { 828df8bae1dSRodney W. Grimes m->m_len -= len; 829df8bae1dSRodney W. Grimes if (mp->m_flags & M_PKTHDR) 830df8bae1dSRodney W. Grimes mp->m_pkthdr.len -= len; 831df8bae1dSRodney W. Grimes return; 832df8bae1dSRodney W. Grimes } 833df8bae1dSRodney W. Grimes count -= len; 834df8bae1dSRodney W. Grimes if (count < 0) 835df8bae1dSRodney W. Grimes count = 0; 836df8bae1dSRodney W. Grimes /* 837df8bae1dSRodney W. Grimes * Correct length for chain is "count". 838df8bae1dSRodney W. Grimes * Find the mbuf with last data, adjust its length, 839df8bae1dSRodney W. Grimes * and toss data from remaining mbufs on chain. 840df8bae1dSRodney W. Grimes */ 841df8bae1dSRodney W. Grimes m = mp; 842df8bae1dSRodney W. Grimes if (m->m_flags & M_PKTHDR) 843df8bae1dSRodney W. Grimes m->m_pkthdr.len = count; 844df8bae1dSRodney W. Grimes for (; m; m = m->m_next) { 845df8bae1dSRodney W. Grimes if (m->m_len >= count) { 846df8bae1dSRodney W. Grimes m->m_len = count; 84759d8b310SSam Leffler if (m->m_next != NULL) { 84859d8b310SSam Leffler m_freem(m->m_next); 84959d8b310SSam Leffler m->m_next = NULL; 85059d8b310SSam Leffler } 851df8bae1dSRodney W. Grimes break; 852df8bae1dSRodney W. Grimes } 853df8bae1dSRodney W. Grimes count -= m->m_len; 854df8bae1dSRodney W. Grimes } 855df8bae1dSRodney W. Grimes } 856df8bae1dSRodney W. Grimes } 857df8bae1dSRodney W. Grimes 858e243367bSKonstantin Belousov void 859e243367bSKonstantin Belousov m_adj_decap(struct mbuf *mp, int len) 860e243367bSKonstantin Belousov { 861e243367bSKonstantin Belousov uint8_t rsstype; 862e243367bSKonstantin Belousov 863e243367bSKonstantin Belousov m_adj(mp, len); 864e243367bSKonstantin Belousov if ((mp->m_flags & M_PKTHDR) != 0) { 865e243367bSKonstantin Belousov /* 866e243367bSKonstantin Belousov * If flowid was calculated by card from the inner 867e243367bSKonstantin Belousov * headers, move flowid to the decapsulated mbuf 868e243367bSKonstantin Belousov * chain, otherwise clear. This depends on the 869e243367bSKonstantin Belousov * internals of m_adj, which keeps pkthdr as is, in 870e243367bSKonstantin Belousov * particular not changing rsstype and flowid. 871e243367bSKonstantin Belousov */ 872e243367bSKonstantin Belousov rsstype = mp->m_pkthdr.rsstype; 873e243367bSKonstantin Belousov if ((rsstype & M_HASHTYPE_INNER) != 0) { 874e243367bSKonstantin Belousov M_HASHTYPE_SET(mp, rsstype & ~M_HASHTYPE_INNER); 875e243367bSKonstantin Belousov } else { 876e243367bSKonstantin Belousov M_HASHTYPE_CLEAR(mp); 877e243367bSKonstantin Belousov } 878e243367bSKonstantin Belousov } 879e243367bSKonstantin Belousov } 880e243367bSKonstantin Belousov 881df8bae1dSRodney W. Grimes /* 882df8bae1dSRodney W. Grimes * Rearange an mbuf chain so that len bytes are contiguous 883a2c36a02SKevin Lo * and in the data area of an mbuf (so that mtod will work 884a2c36a02SKevin Lo * for a structure of size len). Returns the resulting 885df8bae1dSRodney W. Grimes * mbuf chain on success, frees it and returns null on failure. 886df8bae1dSRodney W. Grimes * If there is room, it will add up to max_protohdr-len extra bytes to the 887df8bae1dSRodney W. Grimes * contiguous region in an attempt to avoid being called next time. 888df8bae1dSRodney W. Grimes */ 889df8bae1dSRodney W. Grimes struct mbuf * 890122a814aSBosko Milekic m_pullup(struct mbuf *n, int len) 891df8bae1dSRodney W. Grimes { 892122a814aSBosko Milekic struct mbuf *m; 893122a814aSBosko Milekic int count; 894df8bae1dSRodney W. Grimes int space; 895df8bae1dSRodney W. Grimes 8966edfd179SGleb Smirnoff KASSERT((n->m_flags & M_EXTPG) == 0, 89782334850SJohn Baldwin ("%s: unmapped mbuf %p", __func__, n)); 89882334850SJohn Baldwin 899df8bae1dSRodney W. Grimes /* 900df8bae1dSRodney W. Grimes * If first mbuf has no cluster, and has room for len bytes 901df8bae1dSRodney W. Grimes * without shifting current data, pullup into it, 902df8bae1dSRodney W. Grimes * otherwise allocate a new mbuf to prepend to the chain. 903df8bae1dSRodney W. Grimes */ 904df8bae1dSRodney W. Grimes if ((n->m_flags & M_EXT) == 0 && 905df8bae1dSRodney W. Grimes n->m_data + len < &n->m_dat[MLEN] && n->m_next) { 906df8bae1dSRodney W. Grimes if (n->m_len >= len) 907df8bae1dSRodney W. Grimes return (n); 908df8bae1dSRodney W. Grimes m = n; 909df8bae1dSRodney W. Grimes n = n->m_next; 910df8bae1dSRodney W. Grimes len -= m->m_len; 911df8bae1dSRodney W. Grimes } else { 912df8bae1dSRodney W. Grimes if (len > MHLEN) 913df8bae1dSRodney W. Grimes goto bad; 914c95be8b5SGleb Smirnoff m = m_get(M_NOWAIT, n->m_type); 915122a814aSBosko Milekic if (m == NULL) 916df8bae1dSRodney W. Grimes goto bad; 9179967cafcSSam Leffler if (n->m_flags & M_PKTHDR) 918c95be8b5SGleb Smirnoff m_move_pkthdr(m, n); 919df8bae1dSRodney W. Grimes } 920df8bae1dSRodney W. Grimes space = &m->m_dat[MLEN] - (m->m_data + m->m_len); 921df8bae1dSRodney W. Grimes do { 922df8bae1dSRodney W. Grimes count = min(min(max(len, max_protohdr), space), n->m_len); 923df8bae1dSRodney W. Grimes bcopy(mtod(n, caddr_t), mtod(m, caddr_t) + m->m_len, 924bd395ae8SBosko Milekic (u_int)count); 925df8bae1dSRodney W. Grimes len -= count; 926df8bae1dSRodney W. Grimes m->m_len += count; 927df8bae1dSRodney W. Grimes n->m_len -= count; 928df8bae1dSRodney W. Grimes space -= count; 929df8bae1dSRodney W. Grimes if (n->m_len) 930df8bae1dSRodney W. Grimes n->m_data += count; 931df8bae1dSRodney W. Grimes else 932df8bae1dSRodney W. Grimes n = m_free(n); 933df8bae1dSRodney W. Grimes } while (len > 0 && n); 934df8bae1dSRodney W. Grimes if (len > 0) { 935df8bae1dSRodney W. Grimes (void) m_free(m); 936df8bae1dSRodney W. Grimes goto bad; 937df8bae1dSRodney W. Grimes } 938df8bae1dSRodney W. Grimes m->m_next = n; 939df8bae1dSRodney W. Grimes return (m); 940df8bae1dSRodney W. Grimes bad: 941df8bae1dSRodney W. Grimes m_freem(n); 942122a814aSBosko Milekic return (NULL); 943df8bae1dSRodney W. Grimes } 944df8bae1dSRodney W. Grimes 945df8bae1dSRodney W. Grimes /* 9467ac139a9SJohn-Mark Gurney * Like m_pullup(), except a new mbuf is always allocated, and we allow 9477ac139a9SJohn-Mark Gurney * the amount of empty space before the data in the new mbuf to be specified 9487ac139a9SJohn-Mark Gurney * (in the event that the caller expects to prepend later). 9497ac139a9SJohn-Mark Gurney */ 9507ac139a9SJohn-Mark Gurney struct mbuf * 9517ac139a9SJohn-Mark Gurney m_copyup(struct mbuf *n, int len, int dstoff) 9527ac139a9SJohn-Mark Gurney { 9537ac139a9SJohn-Mark Gurney struct mbuf *m; 9547ac139a9SJohn-Mark Gurney int count, space; 9557ac139a9SJohn-Mark Gurney 9567ac139a9SJohn-Mark Gurney if (len > (MHLEN - dstoff)) 9577ac139a9SJohn-Mark Gurney goto bad; 958c95be8b5SGleb Smirnoff m = m_get(M_NOWAIT, n->m_type); 9597ac139a9SJohn-Mark Gurney if (m == NULL) 9607ac139a9SJohn-Mark Gurney goto bad; 9617ac139a9SJohn-Mark Gurney if (n->m_flags & M_PKTHDR) 962c95be8b5SGleb Smirnoff m_move_pkthdr(m, n); 9637ac139a9SJohn-Mark Gurney m->m_data += dstoff; 9647ac139a9SJohn-Mark Gurney space = &m->m_dat[MLEN] - (m->m_data + m->m_len); 9657ac139a9SJohn-Mark Gurney do { 9667ac139a9SJohn-Mark Gurney count = min(min(max(len, max_protohdr), space), n->m_len); 9677ac139a9SJohn-Mark Gurney memcpy(mtod(m, caddr_t) + m->m_len, mtod(n, caddr_t), 9687ac139a9SJohn-Mark Gurney (unsigned)count); 9697ac139a9SJohn-Mark Gurney len -= count; 9707ac139a9SJohn-Mark Gurney m->m_len += count; 9717ac139a9SJohn-Mark Gurney n->m_len -= count; 9727ac139a9SJohn-Mark Gurney space -= count; 9737ac139a9SJohn-Mark Gurney if (n->m_len) 9747ac139a9SJohn-Mark Gurney n->m_data += count; 9757ac139a9SJohn-Mark Gurney else 9767ac139a9SJohn-Mark Gurney n = m_free(n); 9777ac139a9SJohn-Mark Gurney } while (len > 0 && n); 9787ac139a9SJohn-Mark Gurney if (len > 0) { 9797ac139a9SJohn-Mark Gurney (void) m_free(m); 9807ac139a9SJohn-Mark Gurney goto bad; 9817ac139a9SJohn-Mark Gurney } 9827ac139a9SJohn-Mark Gurney m->m_next = n; 9837ac139a9SJohn-Mark Gurney return (m); 9847ac139a9SJohn-Mark Gurney bad: 9857ac139a9SJohn-Mark Gurney m_freem(n); 9867ac139a9SJohn-Mark Gurney return (NULL); 9877ac139a9SJohn-Mark Gurney } 9887ac139a9SJohn-Mark Gurney 9897ac139a9SJohn-Mark Gurney /* 990df8bae1dSRodney W. Grimes * Partition an mbuf chain in two pieces, returning the tail -- 991df8bae1dSRodney W. Grimes * all but the first len0 bytes. In case of failure, it returns NULL and 992df8bae1dSRodney W. Grimes * attempts to restore the chain to its original state. 99348d183faSArchie Cobbs * 99448d183faSArchie Cobbs * Note that the resulting mbufs might be read-only, because the new 99548d183faSArchie Cobbs * mbuf can end up sharing an mbuf cluster with the original mbuf if 99648d183faSArchie Cobbs * the "breaking point" happens to lie within a cluster mbuf. Use the 99748d183faSArchie Cobbs * M_WRITABLE() macro to check for this case. 998df8bae1dSRodney W. Grimes */ 999df8bae1dSRodney W. Grimes struct mbuf * 1000122a814aSBosko Milekic m_split(struct mbuf *m0, int len0, int wait) 1001df8bae1dSRodney W. Grimes { 1002122a814aSBosko Milekic struct mbuf *m, *n; 1003bd395ae8SBosko Milekic u_int len = len0, remain; 1004df8bae1dSRodney W. Grimes 1005063d8114SAlfred Perlstein MBUF_CHECKSLEEP(wait); 1006df8bae1dSRodney W. Grimes for (m = m0; m && len > m->m_len; m = m->m_next) 1007df8bae1dSRodney W. Grimes len -= m->m_len; 1008122a814aSBosko Milekic if (m == NULL) 1009122a814aSBosko Milekic return (NULL); 1010df8bae1dSRodney W. Grimes remain = m->m_len - len; 101121f39848SGleb Smirnoff if (m0->m_flags & M_PKTHDR && remain == 0) { 101221f39848SGleb Smirnoff n = m_gethdr(wait, m0->m_type); 101377badb18SGleb Smirnoff if (n == NULL) 101421f39848SGleb Smirnoff return (NULL); 101521f39848SGleb Smirnoff n->m_next = m->m_next; 101621f39848SGleb Smirnoff m->m_next = NULL; 1017fb3bc596SJohn Baldwin if (m0->m_pkthdr.csum_flags & CSUM_SND_TAG) { 1018fb3bc596SJohn Baldwin n->m_pkthdr.snd_tag = 1019fb3bc596SJohn Baldwin m_snd_tag_ref(m0->m_pkthdr.snd_tag); 1020fb3bc596SJohn Baldwin n->m_pkthdr.csum_flags |= CSUM_SND_TAG; 1021fb3bc596SJohn Baldwin } else 102221f39848SGleb Smirnoff n->m_pkthdr.rcvif = m0->m_pkthdr.rcvif; 102321f39848SGleb Smirnoff n->m_pkthdr.len = m0->m_pkthdr.len - len0; 102421f39848SGleb Smirnoff m0->m_pkthdr.len = len0; 102521f39848SGleb Smirnoff return (n); 102621f39848SGleb Smirnoff } else if (m0->m_flags & M_PKTHDR) { 1027c95be8b5SGleb Smirnoff n = m_gethdr(wait, m0->m_type); 1028122a814aSBosko Milekic if (n == NULL) 1029122a814aSBosko Milekic return (NULL); 1030fb3bc596SJohn Baldwin if (m0->m_pkthdr.csum_flags & CSUM_SND_TAG) { 1031fb3bc596SJohn Baldwin n->m_pkthdr.snd_tag = 1032fb3bc596SJohn Baldwin m_snd_tag_ref(m0->m_pkthdr.snd_tag); 1033fb3bc596SJohn Baldwin n->m_pkthdr.csum_flags |= CSUM_SND_TAG; 1034fb3bc596SJohn Baldwin } else 1035df8bae1dSRodney W. Grimes n->m_pkthdr.rcvif = m0->m_pkthdr.rcvif; 1036df8bae1dSRodney W. Grimes n->m_pkthdr.len = m0->m_pkthdr.len - len0; 1037df8bae1dSRodney W. Grimes m0->m_pkthdr.len = len0; 103861664ee7SGleb Smirnoff if (m->m_flags & (M_EXT|M_EXTPG)) 1039df8bae1dSRodney W. Grimes goto extpacket; 1040df8bae1dSRodney W. Grimes if (remain > MHLEN) { 1041df8bae1dSRodney W. Grimes /* m can't be the lead packet */ 1042ed6a66caSRobert Watson M_ALIGN(n, 0); 1043df8bae1dSRodney W. Grimes n->m_next = m_split(m, len, wait); 1044122a814aSBosko Milekic if (n->m_next == NULL) { 1045df8bae1dSRodney W. Grimes (void) m_free(n); 1046122a814aSBosko Milekic return (NULL); 104740376987SJeffrey Hsu } else { 104840376987SJeffrey Hsu n->m_len = 0; 1049df8bae1dSRodney W. Grimes return (n); 105040376987SJeffrey Hsu } 1051df8bae1dSRodney W. Grimes } else 1052ed6a66caSRobert Watson M_ALIGN(n, remain); 1053df8bae1dSRodney W. Grimes } else if (remain == 0) { 1054df8bae1dSRodney W. Grimes n = m->m_next; 1055122a814aSBosko Milekic m->m_next = NULL; 1056df8bae1dSRodney W. Grimes return (n); 1057df8bae1dSRodney W. Grimes } else { 1058c95be8b5SGleb Smirnoff n = m_get(wait, m->m_type); 1059122a814aSBosko Milekic if (n == NULL) 1060122a814aSBosko Milekic return (NULL); 1061df8bae1dSRodney W. Grimes M_ALIGN(n, remain); 1062df8bae1dSRodney W. Grimes } 1063df8bae1dSRodney W. Grimes extpacket: 106461664ee7SGleb Smirnoff if (m->m_flags & (M_EXT|M_EXTPG)) { 1065df8bae1dSRodney W. Grimes n->m_data = m->m_data + len; 106656a4e45aSAndre Oppermann mb_dupcl(n, m); 1067df8bae1dSRodney W. Grimes } else { 1068df8bae1dSRodney W. Grimes bcopy(mtod(m, caddr_t) + len, mtod(n, caddr_t), remain); 1069df8bae1dSRodney W. Grimes } 1070df8bae1dSRodney W. Grimes n->m_len = remain; 1071df8bae1dSRodney W. Grimes m->m_len = len; 1072df8bae1dSRodney W. Grimes n->m_next = m->m_next; 1073122a814aSBosko Milekic m->m_next = NULL; 1074df8bae1dSRodney W. Grimes return (n); 1075df8bae1dSRodney W. Grimes } 1076df8bae1dSRodney W. Grimes /* 1077df8bae1dSRodney W. Grimes * Routine to copy from device local memory into mbufs. 1078f5eece3fSBosko Milekic * Note that `off' argument is offset into first mbuf of target chain from 1079f5eece3fSBosko Milekic * which to begin copying the data to. 1080df8bae1dSRodney W. Grimes */ 1081df8bae1dSRodney W. Grimes struct mbuf * 1082f5eece3fSBosko Milekic m_devget(char *buf, int totlen, int off, struct ifnet *ifp, 1083122a814aSBosko Milekic void (*copy)(char *from, caddr_t to, u_int len)) 1084df8bae1dSRodney W. Grimes { 1085122a814aSBosko Milekic struct mbuf *m; 1086099a0e58SBosko Milekic struct mbuf *top = NULL, **mp = ⊤ 1087f5eece3fSBosko Milekic int len; 1088df8bae1dSRodney W. Grimes 1089f5eece3fSBosko Milekic if (off < 0 || off > MHLEN) 1090f5eece3fSBosko Milekic return (NULL); 1091f5eece3fSBosko Milekic 1092df8bae1dSRodney W. Grimes while (totlen > 0) { 1093099a0e58SBosko Milekic if (top == NULL) { /* First one, must be PKTHDR */ 1094f5eece3fSBosko Milekic if (totlen + off >= MINCLSIZE) { 1095eb1b1807SGleb Smirnoff m = m_getcl(M_NOWAIT, MT_DATA, M_PKTHDR); 1096f5eece3fSBosko Milekic len = MCLBYTES; 1097df8bae1dSRodney W. Grimes } else { 1098eb1b1807SGleb Smirnoff m = m_gethdr(M_NOWAIT, MT_DATA); 1099099a0e58SBosko Milekic len = MHLEN; 1100099a0e58SBosko Milekic 1101099a0e58SBosko Milekic /* Place initial small packet/header at end of mbuf */ 1102cee4a056SKevin Lo if (m && totlen + off + max_linkhdr <= MHLEN) { 1103df8bae1dSRodney W. Grimes m->m_data += max_linkhdr; 1104f5eece3fSBosko Milekic len -= max_linkhdr; 1105df8bae1dSRodney W. Grimes } 1106f5eece3fSBosko Milekic } 1107099a0e58SBosko Milekic if (m == NULL) 1108099a0e58SBosko Milekic return NULL; 1109099a0e58SBosko Milekic m->m_pkthdr.rcvif = ifp; 1110099a0e58SBosko Milekic m->m_pkthdr.len = totlen; 1111099a0e58SBosko Milekic } else { 1112099a0e58SBosko Milekic if (totlen + off >= MINCLSIZE) { 1113eb1b1807SGleb Smirnoff m = m_getcl(M_NOWAIT, MT_DATA, 0); 1114099a0e58SBosko Milekic len = MCLBYTES; 1115099a0e58SBosko Milekic } else { 1116eb1b1807SGleb Smirnoff m = m_get(M_NOWAIT, MT_DATA); 1117099a0e58SBosko Milekic len = MLEN; 1118099a0e58SBosko Milekic } 1119099a0e58SBosko Milekic if (m == NULL) { 1120099a0e58SBosko Milekic m_freem(top); 1121099a0e58SBosko Milekic return NULL; 1122099a0e58SBosko Milekic } 1123099a0e58SBosko Milekic } 1124f5eece3fSBosko Milekic if (off) { 1125f5eece3fSBosko Milekic m->m_data += off; 1126f5eece3fSBosko Milekic len -= off; 1127f5eece3fSBosko Milekic off = 0; 1128f5eece3fSBosko Milekic } 1129f5eece3fSBosko Milekic m->m_len = len = min(totlen, len); 1130df8bae1dSRodney W. Grimes if (copy) 1131bd395ae8SBosko Milekic copy(buf, mtod(m, caddr_t), (u_int)len); 1132df8bae1dSRodney W. Grimes else 1133bd395ae8SBosko Milekic bcopy(buf, mtod(m, caddr_t), (u_int)len); 1134f5eece3fSBosko Milekic buf += len; 1135df8bae1dSRodney W. Grimes *mp = m; 1136df8bae1dSRodney W. Grimes mp = &m->m_next; 1137df8bae1dSRodney W. Grimes totlen -= len; 1138df8bae1dSRodney W. Grimes } 1139df8bae1dSRodney W. Grimes return (top); 1140df8bae1dSRodney W. Grimes } 1141c5789ba3SPoul-Henning Kamp 1142*3f9dac85SJohn Baldwin static void 1143*3f9dac85SJohn Baldwin m_copytounmapped(const struct mbuf *m, int off, int len, c_caddr_t cp) 1144*3f9dac85SJohn Baldwin { 1145*3f9dac85SJohn Baldwin struct iovec iov; 1146*3f9dac85SJohn Baldwin struct uio uio; 1147*3f9dac85SJohn Baldwin int error; 1148*3f9dac85SJohn Baldwin 1149*3f9dac85SJohn Baldwin KASSERT(off >= 0, ("m_copytounmapped: negative off %d", off)); 1150*3f9dac85SJohn Baldwin KASSERT(len >= 0, ("m_copytounmapped: negative len %d", len)); 1151*3f9dac85SJohn Baldwin KASSERT(off < m->m_len, ("m_copytounmapped: len exceeds mbuf length")); 1152*3f9dac85SJohn Baldwin iov.iov_base = __DECONST(caddr_t, cp); 1153*3f9dac85SJohn Baldwin iov.iov_len = len; 1154*3f9dac85SJohn Baldwin uio.uio_resid = len; 1155*3f9dac85SJohn Baldwin uio.uio_iov = &iov; 1156*3f9dac85SJohn Baldwin uio.uio_segflg = UIO_SYSSPACE; 1157*3f9dac85SJohn Baldwin uio.uio_iovcnt = 1; 1158*3f9dac85SJohn Baldwin uio.uio_offset = 0; 1159*3f9dac85SJohn Baldwin uio.uio_rw = UIO_WRITE; 1160*3f9dac85SJohn Baldwin error = m_unmappedtouio(m, off, &uio, len); 1161*3f9dac85SJohn Baldwin KASSERT(error == 0, ("m_unmappedtouio failed: off %d, len %d", off, 1162*3f9dac85SJohn Baldwin len)); 1163*3f9dac85SJohn Baldwin } 1164*3f9dac85SJohn Baldwin 1165c5789ba3SPoul-Henning Kamp /* 1166c5789ba3SPoul-Henning Kamp * Copy data from a buffer back into the indicated mbuf chain, 1167c5789ba3SPoul-Henning Kamp * starting "off" bytes from the beginning, extending the mbuf 1168c5789ba3SPoul-Henning Kamp * chain if necessary. 1169c5789ba3SPoul-Henning Kamp */ 1170c5789ba3SPoul-Henning Kamp void 117124665342SLuigi Rizzo m_copyback(struct mbuf *m0, int off, int len, c_caddr_t cp) 1172c5789ba3SPoul-Henning Kamp { 1173122a814aSBosko Milekic int mlen; 1174122a814aSBosko Milekic struct mbuf *m = m0, *n; 1175c5789ba3SPoul-Henning Kamp int totlen = 0; 1176c5789ba3SPoul-Henning Kamp 1177122a814aSBosko Milekic if (m0 == NULL) 1178c5789ba3SPoul-Henning Kamp return; 1179c5789ba3SPoul-Henning Kamp while (off > (mlen = m->m_len)) { 1180c5789ba3SPoul-Henning Kamp off -= mlen; 1181c5789ba3SPoul-Henning Kamp totlen += mlen; 1182122a814aSBosko Milekic if (m->m_next == NULL) { 1183eb1b1807SGleb Smirnoff n = m_get(M_NOWAIT, m->m_type); 1184122a814aSBosko Milekic if (n == NULL) 1185c5789ba3SPoul-Henning Kamp goto out; 1186099a0e58SBosko Milekic bzero(mtod(n, caddr_t), MLEN); 1187c5789ba3SPoul-Henning Kamp n->m_len = min(MLEN, len + off); 1188c5789ba3SPoul-Henning Kamp m->m_next = n; 1189c5789ba3SPoul-Henning Kamp } 1190c5789ba3SPoul-Henning Kamp m = m->m_next; 1191c5789ba3SPoul-Henning Kamp } 1192c5789ba3SPoul-Henning Kamp while (len > 0) { 1193129c5c81SAlexander Motin if (m->m_next == NULL && (len > m->m_len - off)) { 1194129c5c81SAlexander Motin m->m_len += min(len - (m->m_len - off), 1195129c5c81SAlexander Motin M_TRAILINGSPACE(m)); 1196129c5c81SAlexander Motin } 1197c5789ba3SPoul-Henning Kamp mlen = min (m->m_len - off, len); 1198*3f9dac85SJohn Baldwin if ((m->m_flags & M_EXTPG) != 0) 1199*3f9dac85SJohn Baldwin m_copytounmapped(m, off, mlen, cp); 1200*3f9dac85SJohn Baldwin else 1201bd395ae8SBosko Milekic bcopy(cp, off + mtod(m, caddr_t), (u_int)mlen); 1202c5789ba3SPoul-Henning Kamp cp += mlen; 1203c5789ba3SPoul-Henning Kamp len -= mlen; 1204c5789ba3SPoul-Henning Kamp mlen += off; 1205c5789ba3SPoul-Henning Kamp off = 0; 1206c5789ba3SPoul-Henning Kamp totlen += mlen; 1207c5789ba3SPoul-Henning Kamp if (len == 0) 1208c5789ba3SPoul-Henning Kamp break; 1209122a814aSBosko Milekic if (m->m_next == NULL) { 1210eb1b1807SGleb Smirnoff n = m_get(M_NOWAIT, m->m_type); 1211122a814aSBosko Milekic if (n == NULL) 1212c5789ba3SPoul-Henning Kamp break; 1213c5789ba3SPoul-Henning Kamp n->m_len = min(MLEN, len); 1214c5789ba3SPoul-Henning Kamp m->m_next = n; 1215c5789ba3SPoul-Henning Kamp } 1216c5789ba3SPoul-Henning Kamp m = m->m_next; 1217c5789ba3SPoul-Henning Kamp } 1218c5789ba3SPoul-Henning Kamp out: if (((m = m0)->m_flags & M_PKTHDR) && (m->m_pkthdr.len < totlen)) 1219c5789ba3SPoul-Henning Kamp m->m_pkthdr.len = totlen; 1220c5789ba3SPoul-Henning Kamp } 1221ce4a64f7SPoul-Henning Kamp 122237621fd5SBruce M Simpson /* 12234873d175SSam Leffler * Append the specified data to the indicated mbuf chain, 12244873d175SSam Leffler * Extend the mbuf chain if the new data does not fit in 12254873d175SSam Leffler * existing space. 12264873d175SSam Leffler * 12274873d175SSam Leffler * Return 1 if able to complete the job; otherwise 0. 12284873d175SSam Leffler */ 12294873d175SSam Leffler int 12304873d175SSam Leffler m_append(struct mbuf *m0, int len, c_caddr_t cp) 12314873d175SSam Leffler { 12324873d175SSam Leffler struct mbuf *m, *n; 12334873d175SSam Leffler int remainder, space; 12344873d175SSam Leffler 12354873d175SSam Leffler for (m = m0; m->m_next != NULL; m = m->m_next) 12364873d175SSam Leffler ; 12374873d175SSam Leffler remainder = len; 12384873d175SSam Leffler space = M_TRAILINGSPACE(m); 12394873d175SSam Leffler if (space > 0) { 12404873d175SSam Leffler /* 12414873d175SSam Leffler * Copy into available space. 12424873d175SSam Leffler */ 12434873d175SSam Leffler if (space > remainder) 12444873d175SSam Leffler space = remainder; 12454873d175SSam Leffler bcopy(cp, mtod(m, caddr_t) + m->m_len, space); 12464873d175SSam Leffler m->m_len += space; 12474873d175SSam Leffler cp += space, remainder -= space; 12484873d175SSam Leffler } 12494873d175SSam Leffler while (remainder > 0) { 12504873d175SSam Leffler /* 12514873d175SSam Leffler * Allocate a new mbuf; could check space 12524873d175SSam Leffler * and allocate a cluster instead. 12534873d175SSam Leffler */ 1254eb1b1807SGleb Smirnoff n = m_get(M_NOWAIT, m->m_type); 12554873d175SSam Leffler if (n == NULL) 12564873d175SSam Leffler break; 12574873d175SSam Leffler n->m_len = min(MLEN, remainder); 1258a37c415eSSam Leffler bcopy(cp, mtod(n, caddr_t), n->m_len); 1259a37c415eSSam Leffler cp += n->m_len, remainder -= n->m_len; 12604873d175SSam Leffler m->m_next = n; 12614873d175SSam Leffler m = n; 12624873d175SSam Leffler } 12634873d175SSam Leffler if (m0->m_flags & M_PKTHDR) 12644873d175SSam Leffler m0->m_pkthdr.len += len - remainder; 12654873d175SSam Leffler return (remainder == 0); 12664873d175SSam Leffler } 12674873d175SSam Leffler 12683c7a01d7SJohn Baldwin static int 12693c7a01d7SJohn Baldwin m_apply_extpg_one(struct mbuf *m, int off, int len, 12703c7a01d7SJohn Baldwin int (*f)(void *, void *, u_int), void *arg) 12713c7a01d7SJohn Baldwin { 12723c7a01d7SJohn Baldwin void *p; 12733c7a01d7SJohn Baldwin u_int i, count, pgoff, pglen; 12743c7a01d7SJohn Baldwin int rval; 12753c7a01d7SJohn Baldwin 12763c7a01d7SJohn Baldwin KASSERT(PMAP_HAS_DMAP, 12773c7a01d7SJohn Baldwin ("m_apply_extpg_one does not support unmapped mbufs")); 12783c7a01d7SJohn Baldwin off += mtod(m, vm_offset_t); 12793c7a01d7SJohn Baldwin if (off < m->m_epg_hdrlen) { 12803c7a01d7SJohn Baldwin count = min(m->m_epg_hdrlen - off, len); 12813c7a01d7SJohn Baldwin rval = f(arg, m->m_epg_hdr + off, count); 12823c7a01d7SJohn Baldwin if (rval) 12833c7a01d7SJohn Baldwin return (rval); 12843c7a01d7SJohn Baldwin len -= count; 12853c7a01d7SJohn Baldwin off = 0; 12863c7a01d7SJohn Baldwin } else 12873c7a01d7SJohn Baldwin off -= m->m_epg_hdrlen; 12883c7a01d7SJohn Baldwin pgoff = m->m_epg_1st_off; 12893c7a01d7SJohn Baldwin for (i = 0; i < m->m_epg_npgs && len > 0; i++) { 12903c7a01d7SJohn Baldwin pglen = m_epg_pagelen(m, i, pgoff); 12913c7a01d7SJohn Baldwin if (off < pglen) { 12923c7a01d7SJohn Baldwin count = min(pglen - off, len); 12933c7a01d7SJohn Baldwin p = (void *)PHYS_TO_DMAP(m->m_epg_pa[i] + pgoff); 12943c7a01d7SJohn Baldwin rval = f(arg, p, count); 12953c7a01d7SJohn Baldwin if (rval) 12963c7a01d7SJohn Baldwin return (rval); 12973c7a01d7SJohn Baldwin len -= count; 12983c7a01d7SJohn Baldwin off = 0; 12993c7a01d7SJohn Baldwin } else 13003c7a01d7SJohn Baldwin off -= pglen; 13013c7a01d7SJohn Baldwin pgoff = 0; 13023c7a01d7SJohn Baldwin } 13033c7a01d7SJohn Baldwin if (len > 0) { 13043c7a01d7SJohn Baldwin KASSERT(off < m->m_epg_trllen, 13053c7a01d7SJohn Baldwin ("m_apply_extpg_one: offset beyond trailer")); 13063c7a01d7SJohn Baldwin KASSERT(len <= m->m_epg_trllen - off, 13073c7a01d7SJohn Baldwin ("m_apply_extpg_one: length beyond trailer")); 13083c7a01d7SJohn Baldwin return (f(arg, m->m_epg_trail + off, len)); 13093c7a01d7SJohn Baldwin } 13103c7a01d7SJohn Baldwin return (0); 13113c7a01d7SJohn Baldwin } 13123c7a01d7SJohn Baldwin 13133c7a01d7SJohn Baldwin /* Apply function f to the data in a single mbuf. */ 13143c7a01d7SJohn Baldwin static int 13153c7a01d7SJohn Baldwin m_apply_one(struct mbuf *m, int off, int len, 13163c7a01d7SJohn Baldwin int (*f)(void *, void *, u_int), void *arg) 13173c7a01d7SJohn Baldwin { 13183c7a01d7SJohn Baldwin if ((m->m_flags & M_EXTPG) != 0) 13193c7a01d7SJohn Baldwin return (m_apply_extpg_one(m, off, len, f, arg)); 13203c7a01d7SJohn Baldwin else 13213c7a01d7SJohn Baldwin return (f(arg, mtod(m, caddr_t) + off, len)); 13223c7a01d7SJohn Baldwin } 13233c7a01d7SJohn Baldwin 13244873d175SSam Leffler /* 132537621fd5SBruce M Simpson * Apply function f to the data in an mbuf chain starting "off" bytes from 132637621fd5SBruce M Simpson * the beginning, continuing for "len" bytes. 132737621fd5SBruce M Simpson */ 132837621fd5SBruce M Simpson int 132937621fd5SBruce M Simpson m_apply(struct mbuf *m, int off, int len, 133054065297SBruce M Simpson int (*f)(void *, void *, u_int), void *arg) 133137621fd5SBruce M Simpson { 133254065297SBruce M Simpson u_int count; 133337621fd5SBruce M Simpson int rval; 133437621fd5SBruce M Simpson 133537621fd5SBruce M Simpson KASSERT(off >= 0, ("m_apply, negative off %d", off)); 133637621fd5SBruce M Simpson KASSERT(len >= 0, ("m_apply, negative len %d", len)); 133737621fd5SBruce M Simpson while (off > 0) { 133837621fd5SBruce M Simpson KASSERT(m != NULL, ("m_apply, offset > size of mbuf chain")); 133937621fd5SBruce M Simpson if (off < m->m_len) 134037621fd5SBruce M Simpson break; 134137621fd5SBruce M Simpson off -= m->m_len; 134237621fd5SBruce M Simpson m = m->m_next; 134337621fd5SBruce M Simpson } 134437621fd5SBruce M Simpson while (len > 0) { 134537621fd5SBruce M Simpson KASSERT(m != NULL, ("m_apply, offset > size of mbuf chain")); 134637621fd5SBruce M Simpson count = min(m->m_len - off, len); 13473c7a01d7SJohn Baldwin rval = m_apply_one(m, off, count, f, arg); 134837621fd5SBruce M Simpson if (rval) 134937621fd5SBruce M Simpson return (rval); 135037621fd5SBruce M Simpson len -= count; 135137621fd5SBruce M Simpson off = 0; 135237621fd5SBruce M Simpson m = m->m_next; 135337621fd5SBruce M Simpson } 135437621fd5SBruce M Simpson return (0); 135537621fd5SBruce M Simpson } 135637621fd5SBruce M Simpson 135737621fd5SBruce M Simpson /* 135837621fd5SBruce M Simpson * Return a pointer to mbuf/offset of location in mbuf chain. 135937621fd5SBruce M Simpson */ 136037621fd5SBruce M Simpson struct mbuf * 136137621fd5SBruce M Simpson m_getptr(struct mbuf *m, int loc, int *off) 136237621fd5SBruce M Simpson { 136337621fd5SBruce M Simpson 136437621fd5SBruce M Simpson while (loc >= 0) { 136554065297SBruce M Simpson /* Normal end of search. */ 136637621fd5SBruce M Simpson if (m->m_len > loc) { 136737621fd5SBruce M Simpson *off = loc; 136837621fd5SBruce M Simpson return (m); 136937621fd5SBruce M Simpson } else { 137037621fd5SBruce M Simpson loc -= m->m_len; 137137621fd5SBruce M Simpson if (m->m_next == NULL) { 137237621fd5SBruce M Simpson if (loc == 0) { 137354065297SBruce M Simpson /* Point at the end of valid data. */ 137437621fd5SBruce M Simpson *off = m->m_len; 137537621fd5SBruce M Simpson return (m); 137654065297SBruce M Simpson } 137737621fd5SBruce M Simpson return (NULL); 137854065297SBruce M Simpson } 137937621fd5SBruce M Simpson m = m->m_next; 138037621fd5SBruce M Simpson } 138137621fd5SBruce M Simpson } 138237621fd5SBruce M Simpson return (NULL); 138337621fd5SBruce M Simpson } 138437621fd5SBruce M Simpson 1385ce4a64f7SPoul-Henning Kamp void 13867b125090SJohn-Mark Gurney m_print(const struct mbuf *m, int maxlen) 1387ce4a64f7SPoul-Henning Kamp { 1388ce4a64f7SPoul-Henning Kamp int len; 13897b125090SJohn-Mark Gurney int pdata; 13906357e7b5SEivind Eklund const struct mbuf *m2; 1391ce4a64f7SPoul-Henning Kamp 13927e949c46SKenneth D. Merry if (m == NULL) { 13937e949c46SKenneth D. Merry printf("mbuf: %p\n", m); 13947e949c46SKenneth D. Merry return; 13957e949c46SKenneth D. Merry } 13967e949c46SKenneth D. Merry 13977b125090SJohn-Mark Gurney if (m->m_flags & M_PKTHDR) 1398ce4a64f7SPoul-Henning Kamp len = m->m_pkthdr.len; 13997b125090SJohn-Mark Gurney else 14007b125090SJohn-Mark Gurney len = -1; 1401ce4a64f7SPoul-Henning Kamp m2 = m; 14027b125090SJohn-Mark Gurney while (m2 != NULL && (len == -1 || len)) { 14037b125090SJohn-Mark Gurney pdata = m2->m_len; 14047b125090SJohn-Mark Gurney if (maxlen != -1 && pdata > maxlen) 14057b125090SJohn-Mark Gurney pdata = maxlen; 14067b125090SJohn-Mark Gurney printf("mbuf: %p len: %d, next: %p, %b%s", m2, m2->m_len, 14077b125090SJohn-Mark Gurney m2->m_next, m2->m_flags, "\20\20freelist\17skipfw" 14087b125090SJohn-Mark Gurney "\11proto5\10proto4\7proto3\6proto2\5proto1\4rdonly" 14097b125090SJohn-Mark Gurney "\3eor\2pkthdr\1ext", pdata ? "" : "\n"); 14107b125090SJohn-Mark Gurney if (pdata) 141145e0d0aaSJohn-Mark Gurney printf(", %*D\n", pdata, (u_char *)m2->m_data, "-"); 14127b125090SJohn-Mark Gurney if (len != -1) 1413ce4a64f7SPoul-Henning Kamp len -= m2->m_len; 1414ce4a64f7SPoul-Henning Kamp m2 = m2->m_next; 1415ce4a64f7SPoul-Henning Kamp } 14167b125090SJohn-Mark Gurney if (len > 0) 14177b125090SJohn-Mark Gurney printf("%d bytes unaccounted for.\n", len); 1418ce4a64f7SPoul-Henning Kamp return; 1419ce4a64f7SPoul-Henning Kamp } 14203f2e06c5SPoul-Henning Kamp 1421bd395ae8SBosko Milekic u_int 14223f2e06c5SPoul-Henning Kamp m_fixhdr(struct mbuf *m0) 14233f2e06c5SPoul-Henning Kamp { 1424bd395ae8SBosko Milekic u_int len; 14253f2e06c5SPoul-Henning Kamp 1426ac6e585dSPoul-Henning Kamp len = m_length(m0, NULL); 14273f2e06c5SPoul-Henning Kamp m0->m_pkthdr.len = len; 1428ac6e585dSPoul-Henning Kamp return (len); 1429ac6e585dSPoul-Henning Kamp } 1430ac6e585dSPoul-Henning Kamp 1431bd395ae8SBosko Milekic u_int 1432ac6e585dSPoul-Henning Kamp m_length(struct mbuf *m0, struct mbuf **last) 1433ac6e585dSPoul-Henning Kamp { 1434ac6e585dSPoul-Henning Kamp struct mbuf *m; 1435bd395ae8SBosko Milekic u_int len; 1436ac6e585dSPoul-Henning Kamp 1437ac6e585dSPoul-Henning Kamp len = 0; 1438ac6e585dSPoul-Henning Kamp for (m = m0; m != NULL; m = m->m_next) { 1439ac6e585dSPoul-Henning Kamp len += m->m_len; 1440ac6e585dSPoul-Henning Kamp if (m->m_next == NULL) 1441ac6e585dSPoul-Henning Kamp break; 1442ac6e585dSPoul-Henning Kamp } 1443ac6e585dSPoul-Henning Kamp if (last != NULL) 1444ac6e585dSPoul-Henning Kamp *last = m; 1445ac6e585dSPoul-Henning Kamp return (len); 14463f2e06c5SPoul-Henning Kamp } 144755e9f80dSMike Silbersack 144855e9f80dSMike Silbersack /* 144955e9f80dSMike Silbersack * Defragment a mbuf chain, returning the shortest possible 145055e9f80dSMike Silbersack * chain of mbufs and clusters. If allocation fails and 145155e9f80dSMike Silbersack * this cannot be completed, NULL will be returned, but 145255e9f80dSMike Silbersack * the passed in chain will be unchanged. Upon success, 145355e9f80dSMike Silbersack * the original chain will be freed, and the new chain 145455e9f80dSMike Silbersack * will be returned. 145555e9f80dSMike Silbersack * 145655e9f80dSMike Silbersack * If a non-packet header is passed in, the original 145755e9f80dSMike Silbersack * mbuf (chain?) will be returned unharmed. 145855e9f80dSMike Silbersack */ 145955e9f80dSMike Silbersack struct mbuf * 146055e9f80dSMike Silbersack m_defrag(struct mbuf *m0, int how) 146155e9f80dSMike Silbersack { 146255e9f80dSMike Silbersack struct mbuf *m_new = NULL, *m_final = NULL; 146355e9f80dSMike Silbersack int progress = 0, length; 146455e9f80dSMike Silbersack 1465063d8114SAlfred Perlstein MBUF_CHECKSLEEP(how); 146655e9f80dSMike Silbersack if (!(m0->m_flags & M_PKTHDR)) 146755e9f80dSMike Silbersack return (m0); 146855e9f80dSMike Silbersack 1469f8bf8e39SMike Silbersack m_fixhdr(m0); /* Needed sanity check */ 1470f8bf8e39SMike Silbersack 1471352d050eSMike Silbersack #ifdef MBUF_STRESS_TEST 1472352d050eSMike Silbersack if (m_defragrandomfailures) { 1473352d050eSMike Silbersack int temp = arc4random() & 0xff; 1474352d050eSMike Silbersack if (temp == 0xba) 1475352d050eSMike Silbersack goto nospace; 1476352d050eSMike Silbersack } 1477352d050eSMike Silbersack #endif 147855e9f80dSMike Silbersack 147955e9f80dSMike Silbersack if (m0->m_pkthdr.len > MHLEN) 148055e9f80dSMike Silbersack m_final = m_getcl(how, MT_DATA, M_PKTHDR); 148155e9f80dSMike Silbersack else 148255e9f80dSMike Silbersack m_final = m_gethdr(how, MT_DATA); 148355e9f80dSMike Silbersack 148455e9f80dSMike Silbersack if (m_final == NULL) 148555e9f80dSMike Silbersack goto nospace; 148655e9f80dSMike Silbersack 1487a89ec05eSPeter Wemm if (m_dup_pkthdr(m_final, m0, how) == 0) 148855e9f80dSMike Silbersack goto nospace; 148955e9f80dSMike Silbersack 149055e9f80dSMike Silbersack m_new = m_final; 149155e9f80dSMike Silbersack 149255e9f80dSMike Silbersack while (progress < m0->m_pkthdr.len) { 149355e9f80dSMike Silbersack length = m0->m_pkthdr.len - progress; 149455e9f80dSMike Silbersack if (length > MCLBYTES) 149555e9f80dSMike Silbersack length = MCLBYTES; 149655e9f80dSMike Silbersack 149755e9f80dSMike Silbersack if (m_new == NULL) { 149855e9f80dSMike Silbersack if (length > MLEN) 149955e9f80dSMike Silbersack m_new = m_getcl(how, MT_DATA, 0); 150055e9f80dSMike Silbersack else 150155e9f80dSMike Silbersack m_new = m_get(how, MT_DATA); 150255e9f80dSMike Silbersack if (m_new == NULL) 150355e9f80dSMike Silbersack goto nospace; 150455e9f80dSMike Silbersack } 150555e9f80dSMike Silbersack 150655e9f80dSMike Silbersack m_copydata(m0, progress, length, mtod(m_new, caddr_t)); 150755e9f80dSMike Silbersack progress += length; 150855e9f80dSMike Silbersack m_new->m_len = length; 150955e9f80dSMike Silbersack if (m_new != m_final) 151055e9f80dSMike Silbersack m_cat(m_final, m_new); 151155e9f80dSMike Silbersack m_new = NULL; 151255e9f80dSMike Silbersack } 151351710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 151455e9f80dSMike Silbersack if (m0->m_next == NULL) 151555e9f80dSMike Silbersack m_defraguseless++; 151651710a45SMike Silbersack #endif 151755e9f80dSMike Silbersack m_freem(m0); 151855e9f80dSMike Silbersack m0 = m_final; 151951710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 152055e9f80dSMike Silbersack m_defragpackets++; 152155e9f80dSMike Silbersack m_defragbytes += m0->m_pkthdr.len; 152251710a45SMike Silbersack #endif 152355e9f80dSMike Silbersack return (m0); 152455e9f80dSMike Silbersack nospace: 152551710a45SMike Silbersack #ifdef MBUF_STRESS_TEST 152655e9f80dSMike Silbersack m_defragfailure++; 152751710a45SMike Silbersack #endif 152855e9f80dSMike Silbersack if (m_final) 152955e9f80dSMike Silbersack m_freem(m_final); 153055e9f80dSMike Silbersack return (NULL); 153155e9f80dSMike Silbersack } 15323390d476SMike Silbersack 1533eeb76a18SSam Leffler /* 153482334850SJohn Baldwin * Return the number of fragments an mbuf will use. This is usually 153582334850SJohn Baldwin * used as a proxy for the number of scatter/gather elements needed by 153682334850SJohn Baldwin * a DMA engine to access an mbuf. In general mapped mbufs are 153782334850SJohn Baldwin * assumed to be backed by physically contiguous buffers that only 153882334850SJohn Baldwin * need a single fragment. Unmapped mbufs, on the other hand, can 153982334850SJohn Baldwin * span disjoint physical pages. 154082334850SJohn Baldwin */ 154182334850SJohn Baldwin static int 154282334850SJohn Baldwin frags_per_mbuf(struct mbuf *m) 154382334850SJohn Baldwin { 154482334850SJohn Baldwin int frags; 154582334850SJohn Baldwin 15466edfd179SGleb Smirnoff if ((m->m_flags & M_EXTPG) == 0) 154782334850SJohn Baldwin return (1); 154882334850SJohn Baldwin 154982334850SJohn Baldwin /* 155082334850SJohn Baldwin * The header and trailer are counted as a single fragment 155182334850SJohn Baldwin * each when present. 155282334850SJohn Baldwin * 155382334850SJohn Baldwin * XXX: This overestimates the number of fragments by assuming 155482334850SJohn Baldwin * all the backing physical pages are disjoint. 155582334850SJohn Baldwin */ 155682334850SJohn Baldwin frags = 0; 15577b6c99d0SGleb Smirnoff if (m->m_epg_hdrlen != 0) 155882334850SJohn Baldwin frags++; 15597b6c99d0SGleb Smirnoff frags += m->m_epg_npgs; 15607b6c99d0SGleb Smirnoff if (m->m_epg_trllen != 0) 156182334850SJohn Baldwin frags++; 156282334850SJohn Baldwin 156382334850SJohn Baldwin return (frags); 156482334850SJohn Baldwin } 156582334850SJohn Baldwin 156682334850SJohn Baldwin /* 1567eeb76a18SSam Leffler * Defragment an mbuf chain, returning at most maxfrags separate 1568eeb76a18SSam Leffler * mbufs+clusters. If this is not possible NULL is returned and 156928323addSBryan Drewery * the original mbuf chain is left in its present (potentially 1570eeb76a18SSam Leffler * modified) state. We use two techniques: collapsing consecutive 1571eeb76a18SSam Leffler * mbufs and replacing consecutive mbufs by a cluster. 1572eeb76a18SSam Leffler * 1573eeb76a18SSam Leffler * NB: this should really be named m_defrag but that name is taken 1574eeb76a18SSam Leffler */ 1575eeb76a18SSam Leffler struct mbuf * 1576eeb76a18SSam Leffler m_collapse(struct mbuf *m0, int how, int maxfrags) 1577eeb76a18SSam Leffler { 1578eeb76a18SSam Leffler struct mbuf *m, *n, *n2, **prev; 1579eeb76a18SSam Leffler u_int curfrags; 1580eeb76a18SSam Leffler 1581eeb76a18SSam Leffler /* 1582eeb76a18SSam Leffler * Calculate the current number of frags. 1583eeb76a18SSam Leffler */ 1584eeb76a18SSam Leffler curfrags = 0; 1585eeb76a18SSam Leffler for (m = m0; m != NULL; m = m->m_next) 158682334850SJohn Baldwin curfrags += frags_per_mbuf(m); 1587eeb76a18SSam Leffler /* 1588eeb76a18SSam Leffler * First, try to collapse mbufs. Note that we always collapse 1589eeb76a18SSam Leffler * towards the front so we don't need to deal with moving the 1590eeb76a18SSam Leffler * pkthdr. This may be suboptimal if the first mbuf has much 1591eeb76a18SSam Leffler * less data than the following. 1592eeb76a18SSam Leffler */ 1593eeb76a18SSam Leffler m = m0; 1594eeb76a18SSam Leffler again: 1595eeb76a18SSam Leffler for (;;) { 1596eeb76a18SSam Leffler n = m->m_next; 1597eeb76a18SSam Leffler if (n == NULL) 1598eeb76a18SSam Leffler break; 159914d7c5b1SAndre Oppermann if (M_WRITABLE(m) && 1600eeb76a18SSam Leffler n->m_len < M_TRAILINGSPACE(m)) { 160182334850SJohn Baldwin m_copydata(n, 0, n->m_len, 160282334850SJohn Baldwin mtod(m, char *) + m->m_len); 1603eeb76a18SSam Leffler m->m_len += n->m_len; 1604eeb76a18SSam Leffler m->m_next = n->m_next; 160582334850SJohn Baldwin curfrags -= frags_per_mbuf(n); 1606eeb76a18SSam Leffler m_free(n); 160782334850SJohn Baldwin if (curfrags <= maxfrags) 1608eeb76a18SSam Leffler return m0; 1609eeb76a18SSam Leffler } else 1610eeb76a18SSam Leffler m = n; 1611eeb76a18SSam Leffler } 1612eeb76a18SSam Leffler KASSERT(maxfrags > 1, 1613eeb76a18SSam Leffler ("maxfrags %u, but normal collapse failed", maxfrags)); 1614eeb76a18SSam Leffler /* 1615eeb76a18SSam Leffler * Collapse consecutive mbufs to a cluster. 1616eeb76a18SSam Leffler */ 1617eeb76a18SSam Leffler prev = &m0->m_next; /* NB: not the first mbuf */ 1618eeb76a18SSam Leffler while ((n = *prev) != NULL) { 1619eeb76a18SSam Leffler if ((n2 = n->m_next) != NULL && 1620eeb76a18SSam Leffler n->m_len + n2->m_len < MCLBYTES) { 1621eeb76a18SSam Leffler m = m_getcl(how, MT_DATA, 0); 1622eeb76a18SSam Leffler if (m == NULL) 1623eeb76a18SSam Leffler goto bad; 162482334850SJohn Baldwin m_copydata(n, 0, n->m_len, mtod(m, char *)); 162582334850SJohn Baldwin m_copydata(n2, 0, n2->m_len, 162682334850SJohn Baldwin mtod(m, char *) + n->m_len); 1627eeb76a18SSam Leffler m->m_len = n->m_len + n2->m_len; 1628eeb76a18SSam Leffler m->m_next = n2->m_next; 1629eeb76a18SSam Leffler *prev = m; 163082334850SJohn Baldwin curfrags += 1; /* For the new cluster */ 163182334850SJohn Baldwin curfrags -= frags_per_mbuf(n); 163282334850SJohn Baldwin curfrags -= frags_per_mbuf(n2); 1633eeb76a18SSam Leffler m_free(n); 1634eeb76a18SSam Leffler m_free(n2); 163582334850SJohn Baldwin if (curfrags <= maxfrags) 1636eeb76a18SSam Leffler return m0; 1637eeb76a18SSam Leffler /* 1638eeb76a18SSam Leffler * Still not there, try the normal collapse 1639eeb76a18SSam Leffler * again before we allocate another cluster. 1640eeb76a18SSam Leffler */ 1641eeb76a18SSam Leffler goto again; 1642eeb76a18SSam Leffler } 1643eeb76a18SSam Leffler prev = &n->m_next; 1644eeb76a18SSam Leffler } 1645eeb76a18SSam Leffler /* 1646eeb76a18SSam Leffler * No place where we can collapse to a cluster; punt. 1647eeb76a18SSam Leffler * This can occur if, for example, you request 2 frags 1648eeb76a18SSam Leffler * but the packet requires that both be clusters (we 1649eeb76a18SSam Leffler * never reallocate the first mbuf to avoid moving the 1650eeb76a18SSam Leffler * packet header). 1651eeb76a18SSam Leffler */ 1652eeb76a18SSam Leffler bad: 1653eeb76a18SSam Leffler return NULL; 1654eeb76a18SSam Leffler } 1655eeb76a18SSam Leffler 16563390d476SMike Silbersack #ifdef MBUF_STRESS_TEST 16573390d476SMike Silbersack 16583390d476SMike Silbersack /* 16593390d476SMike Silbersack * Fragment an mbuf chain. There's no reason you'd ever want to do 16603390d476SMike Silbersack * this in normal usage, but it's great for stress testing various 16613390d476SMike Silbersack * mbuf consumers. 16623390d476SMike Silbersack * 16633390d476SMike Silbersack * If fragmentation is not possible, the original chain will be 16643390d476SMike Silbersack * returned. 16653390d476SMike Silbersack * 16663390d476SMike Silbersack * Possible length values: 16673390d476SMike Silbersack * 0 no fragmentation will occur 16683390d476SMike Silbersack * > 0 each fragment will be of the specified length 16693390d476SMike Silbersack * -1 each fragment will be the same random value in length 16703390d476SMike Silbersack * -2 each fragment's length will be entirely random 16713390d476SMike Silbersack * (Random values range from 1 to 256) 16723390d476SMike Silbersack */ 16733390d476SMike Silbersack struct mbuf * 16743390d476SMike Silbersack m_fragment(struct mbuf *m0, int how, int length) 16753390d476SMike Silbersack { 167666234298SAndriy Voskoboinyk struct mbuf *m_first, *m_last; 167766234298SAndriy Voskoboinyk int divisor = 255, progress = 0, fraglen; 16783390d476SMike Silbersack 16793390d476SMike Silbersack if (!(m0->m_flags & M_PKTHDR)) 16803390d476SMike Silbersack return (m0); 16813390d476SMike Silbersack 168266234298SAndriy Voskoboinyk if (length == 0 || length < -2) 16833390d476SMike Silbersack return (m0); 168466234298SAndriy Voskoboinyk if (length > MCLBYTES) 168566234298SAndriy Voskoboinyk length = MCLBYTES; 168666234298SAndriy Voskoboinyk if (length < 0 && divisor > MCLBYTES) 168766234298SAndriy Voskoboinyk divisor = MCLBYTES; 168866234298SAndriy Voskoboinyk if (length == -1) 168966234298SAndriy Voskoboinyk length = 1 + (arc4random() % divisor); 169066234298SAndriy Voskoboinyk if (length > 0) 169166234298SAndriy Voskoboinyk fraglen = length; 16923390d476SMike Silbersack 16933390d476SMike Silbersack m_fixhdr(m0); /* Needed sanity check */ 16943390d476SMike Silbersack 169566234298SAndriy Voskoboinyk m_first = m_getcl(how, MT_DATA, M_PKTHDR); 169666234298SAndriy Voskoboinyk if (m_first == NULL) 16973390d476SMike Silbersack goto nospace; 16983390d476SMike Silbersack 169966234298SAndriy Voskoboinyk if (m_dup_pkthdr(m_first, m0, how) == 0) 17003390d476SMike Silbersack goto nospace; 17013390d476SMike Silbersack 170266234298SAndriy Voskoboinyk m_last = m_first; 17033390d476SMike Silbersack 17043390d476SMike Silbersack while (progress < m0->m_pkthdr.len) { 170566234298SAndriy Voskoboinyk if (length == -2) 170666234298SAndriy Voskoboinyk fraglen = 1 + (arc4random() % divisor); 17073390d476SMike Silbersack if (fraglen > m0->m_pkthdr.len - progress) 17083390d476SMike Silbersack fraglen = m0->m_pkthdr.len - progress; 17093390d476SMike Silbersack 171066234298SAndriy Voskoboinyk if (progress != 0) { 171166234298SAndriy Voskoboinyk struct mbuf *m_new = m_getcl(how, MT_DATA, 0); 17123390d476SMike Silbersack if (m_new == NULL) 17133390d476SMike Silbersack goto nospace; 171466234298SAndriy Voskoboinyk 171566234298SAndriy Voskoboinyk m_last->m_next = m_new; 171666234298SAndriy Voskoboinyk m_last = m_new; 17173390d476SMike Silbersack } 17183390d476SMike Silbersack 171966234298SAndriy Voskoboinyk m_copydata(m0, progress, fraglen, mtod(m_last, caddr_t)); 17203390d476SMike Silbersack progress += fraglen; 172166234298SAndriy Voskoboinyk m_last->m_len = fraglen; 17223390d476SMike Silbersack } 17233390d476SMike Silbersack m_freem(m0); 172466234298SAndriy Voskoboinyk m0 = m_first; 17253390d476SMike Silbersack return (m0); 17263390d476SMike Silbersack nospace: 172766234298SAndriy Voskoboinyk if (m_first) 172866234298SAndriy Voskoboinyk m_freem(m_first); 17293390d476SMike Silbersack /* Return the original chain on failure */ 17303390d476SMike Silbersack return (m0); 17313390d476SMike Silbersack } 17323390d476SMike Silbersack 17333390d476SMike Silbersack #endif 1734beb699c7SMike Silbersack 17355e20f43dSAndre Oppermann /* 173682334850SJohn Baldwin * Free pages from mbuf_ext_pgs, assuming they were allocated via 173782334850SJohn Baldwin * vm_page_alloc() and aren't associated with any object. Complement 173882334850SJohn Baldwin * to allocator from m_uiotombuf_nomap(). 173982334850SJohn Baldwin */ 174082334850SJohn Baldwin void 174182334850SJohn Baldwin mb_free_mext_pgs(struct mbuf *m) 174282334850SJohn Baldwin { 174382334850SJohn Baldwin vm_page_t pg; 174482334850SJohn Baldwin 1745365e8da4SGleb Smirnoff M_ASSERTEXTPG(m); 17467b6c99d0SGleb Smirnoff for (int i = 0; i < m->m_epg_npgs; i++) { 17470c103266SGleb Smirnoff pg = PHYS_TO_VM_PAGE(m->m_epg_pa[i]); 17489fb7c918SMark Johnston vm_page_unwire_noq(pg); 17499fb7c918SMark Johnston vm_page_free(pg); 175082334850SJohn Baldwin } 175182334850SJohn Baldwin } 175282334850SJohn Baldwin 175382334850SJohn Baldwin static struct mbuf * 175482334850SJohn Baldwin m_uiotombuf_nomap(struct uio *uio, int how, int len, int maxseg, int flags) 175582334850SJohn Baldwin { 175682334850SJohn Baldwin struct mbuf *m, *mb, *prev; 175782334850SJohn Baldwin vm_page_t pg_array[MBUF_PEXT_MAX_PGS]; 17589fb7c918SMark Johnston int error, length, i, needed; 175982334850SJohn Baldwin ssize_t total; 17609fb7c918SMark Johnston int pflags = malloc2vm_flags(how) | VM_ALLOC_NOOBJ | VM_ALLOC_NODUMP | 17619fb7c918SMark Johnston VM_ALLOC_WIRED; 176282334850SJohn Baldwin 1763c2a8fd6fSJohn Baldwin MPASS((flags & M_PKTHDR) == 0); 1764608c44f9SMark Johnston MPASS((how & M_ZERO) == 0); 1765c2a8fd6fSJohn Baldwin 176682334850SJohn Baldwin /* 176782334850SJohn Baldwin * len can be zero or an arbitrary large value bound by 176882334850SJohn Baldwin * the total data supplied by the uio. 176982334850SJohn Baldwin */ 177082334850SJohn Baldwin if (len > 0) 177182334850SJohn Baldwin total = MIN(uio->uio_resid, len); 177282334850SJohn Baldwin else 177382334850SJohn Baldwin total = uio->uio_resid; 177482334850SJohn Baldwin 177582334850SJohn Baldwin if (maxseg == 0) 177682334850SJohn Baldwin maxseg = MBUF_PEXT_MAX_PGS * PAGE_SIZE; 177782334850SJohn Baldwin 177882334850SJohn Baldwin /* 1779c2a8fd6fSJohn Baldwin * If total is zero, return an empty mbuf. This can occur 1780c2a8fd6fSJohn Baldwin * for TLS 1.0 connections which send empty fragments as 1781c2a8fd6fSJohn Baldwin * a countermeasure against the known-IV weakness in CBC 1782c2a8fd6fSJohn Baldwin * ciphersuites. 1783c2a8fd6fSJohn Baldwin */ 1784c2a8fd6fSJohn Baldwin if (__predict_false(total == 0)) { 1785c2a8fd6fSJohn Baldwin mb = mb_alloc_ext_pgs(how, mb_free_mext_pgs); 1786c2a8fd6fSJohn Baldwin if (mb == NULL) 1787c2a8fd6fSJohn Baldwin return (NULL); 1788c2a8fd6fSJohn Baldwin mb->m_epg_flags = EPG_FLAG_ANON; 1789c2a8fd6fSJohn Baldwin return (mb); 1790c2a8fd6fSJohn Baldwin } 1791c2a8fd6fSJohn Baldwin 1792c2a8fd6fSJohn Baldwin /* 179382334850SJohn Baldwin * Allocate the pages 179482334850SJohn Baldwin */ 179582334850SJohn Baldwin m = NULL; 179682334850SJohn Baldwin while (total > 0) { 179723feb563SAndrew Gallatin mb = mb_alloc_ext_pgs(how, mb_free_mext_pgs); 179882334850SJohn Baldwin if (mb == NULL) 179982334850SJohn Baldwin goto failed; 180082334850SJohn Baldwin if (m == NULL) 180182334850SJohn Baldwin m = mb; 180282334850SJohn Baldwin else 180382334850SJohn Baldwin prev->m_next = mb; 180482334850SJohn Baldwin prev = mb; 18057b6c99d0SGleb Smirnoff mb->m_epg_flags = EPG_FLAG_ANON; 180682334850SJohn Baldwin needed = length = MIN(maxseg, total); 180782334850SJohn Baldwin for (i = 0; needed > 0; i++, needed -= PAGE_SIZE) { 180882334850SJohn Baldwin retry_page: 180982334850SJohn Baldwin pg_array[i] = vm_page_alloc(NULL, 0, pflags); 181082334850SJohn Baldwin if (pg_array[i] == NULL) { 181182334850SJohn Baldwin if (how & M_NOWAIT) { 181282334850SJohn Baldwin goto failed; 181382334850SJohn Baldwin } else { 181482334850SJohn Baldwin vm_wait(NULL); 181582334850SJohn Baldwin goto retry_page; 181682334850SJohn Baldwin } 181782334850SJohn Baldwin } 18180c103266SGleb Smirnoff mb->m_epg_pa[i] = VM_PAGE_TO_PHYS(pg_array[i]); 18197b6c99d0SGleb Smirnoff mb->m_epg_npgs++; 182082334850SJohn Baldwin } 18217b6c99d0SGleb Smirnoff mb->m_epg_last_len = length - PAGE_SIZE * (mb->m_epg_npgs - 1); 18220c103266SGleb Smirnoff MBUF_EXT_PGS_ASSERT_SANITY(mb); 182382334850SJohn Baldwin total -= length; 182482334850SJohn Baldwin error = uiomove_fromphys(pg_array, 0, length, uio); 182582334850SJohn Baldwin if (error != 0) 182682334850SJohn Baldwin goto failed; 182782334850SJohn Baldwin mb->m_len = length; 18287b6c99d0SGleb Smirnoff mb->m_ext.ext_size += PAGE_SIZE * mb->m_epg_npgs; 182982334850SJohn Baldwin if (flags & M_PKTHDR) 183082334850SJohn Baldwin m->m_pkthdr.len += length; 183182334850SJohn Baldwin } 183282334850SJohn Baldwin return (m); 183382334850SJohn Baldwin 183482334850SJohn Baldwin failed: 183582334850SJohn Baldwin m_freem(m); 183682334850SJohn Baldwin return (NULL); 183782334850SJohn Baldwin } 183882334850SJohn Baldwin 183982334850SJohn Baldwin /* 18405e20f43dSAndre Oppermann * Copy the contents of uio into a properly sized mbuf chain. 18415e20f43dSAndre Oppermann */ 1842beb699c7SMike Silbersack struct mbuf * 18435e20f43dSAndre Oppermann m_uiotombuf(struct uio *uio, int how, int len, int align, int flags) 1844beb699c7SMike Silbersack { 18455e20f43dSAndre Oppermann struct mbuf *m, *mb; 1846526d0bd5SKonstantin Belousov int error, length; 1847526d0bd5SKonstantin Belousov ssize_t total; 18485e20f43dSAndre Oppermann int progress = 0; 1849beb699c7SMike Silbersack 18506edfd179SGleb Smirnoff if (flags & M_EXTPG) 185182334850SJohn Baldwin return (m_uiotombuf_nomap(uio, how, len, align, flags)); 185282334850SJohn Baldwin 18535e20f43dSAndre Oppermann /* 18545e20f43dSAndre Oppermann * len can be zero or an arbitrary large value bound by 18555e20f43dSAndre Oppermann * the total data supplied by the uio. 18565e20f43dSAndre Oppermann */ 1857beb699c7SMike Silbersack if (len > 0) 1858f5b7359aSConrad Meyer total = (uio->uio_resid < len) ? uio->uio_resid : len; 1859beb699c7SMike Silbersack else 1860beb699c7SMike Silbersack total = uio->uio_resid; 18615e20f43dSAndre Oppermann 18625e20f43dSAndre Oppermann /* 18635e20f43dSAndre Oppermann * The smallest unit returned by m_getm2() is a single mbuf 18649128ec21SAndrew Thompson * with pkthdr. We can't align past it. 18655e20f43dSAndre Oppermann */ 186675ae2570SMaksim Yevmenkin if (align >= MHLEN) 1867beb699c7SMike Silbersack return (NULL); 18685e20f43dSAndre Oppermann 18697c32173bSAndre Oppermann /* 18707c32173bSAndre Oppermann * Give us the full allocation or nothing. 18717c32173bSAndre Oppermann * If len is zero return the smallest empty mbuf. 18727c32173bSAndre Oppermann */ 18737c32173bSAndre Oppermann m = m_getm2(NULL, max(total + align, 1), how, MT_DATA, flags); 18745e20f43dSAndre Oppermann if (m == NULL) 18755e20f43dSAndre Oppermann return (NULL); 18765e20f43dSAndre Oppermann m->m_data += align; 18775e20f43dSAndre Oppermann 18785e20f43dSAndre Oppermann /* Fill all mbufs with uio data and update header information. */ 18795e20f43dSAndre Oppermann for (mb = m; mb != NULL; mb = mb->m_next) { 18805e20f43dSAndre Oppermann length = min(M_TRAILINGSPACE(mb), total - progress); 18815e20f43dSAndre Oppermann 18825e20f43dSAndre Oppermann error = uiomove(mtod(mb, void *), length, uio); 18835e20f43dSAndre Oppermann if (error) { 18845e20f43dSAndre Oppermann m_freem(m); 18855e20f43dSAndre Oppermann return (NULL); 18865e20f43dSAndre Oppermann } 18875e20f43dSAndre Oppermann 18885e20f43dSAndre Oppermann mb->m_len = length; 18895e20f43dSAndre Oppermann progress += length; 18905e20f43dSAndre Oppermann if (flags & M_PKTHDR) 18915e20f43dSAndre Oppermann m->m_pkthdr.len += length; 18925e20f43dSAndre Oppermann } 18935e20f43dSAndre Oppermann KASSERT(progress == total, ("%s: progress != total", __func__)); 18945e20f43dSAndre Oppermann 18955e20f43dSAndre Oppermann return (m); 1896beb699c7SMike Silbersack } 1897ab8ab90cSSam Leffler 1898ab8ab90cSSam Leffler /* 1899*3f9dac85SJohn Baldwin * Copy data to/from an unmapped mbuf into a uio limited by len if set. 190082334850SJohn Baldwin */ 190182334850SJohn Baldwin int 190282334850SJohn Baldwin m_unmappedtouio(const struct mbuf *m, int m_off, struct uio *uio, int len) 190382334850SJohn Baldwin { 190482334850SJohn Baldwin vm_page_t pg; 190582334850SJohn Baldwin int error, i, off, pglen, pgoff, seglen, segoff; 190682334850SJohn Baldwin 1907365e8da4SGleb Smirnoff M_ASSERTEXTPG(m); 190882334850SJohn Baldwin error = 0; 190982334850SJohn Baldwin 191082334850SJohn Baldwin /* Skip over any data removed from the front. */ 191182334850SJohn Baldwin off = mtod(m, vm_offset_t); 191282334850SJohn Baldwin 191382334850SJohn Baldwin off += m_off; 19147b6c99d0SGleb Smirnoff if (m->m_epg_hdrlen != 0) { 19157b6c99d0SGleb Smirnoff if (off >= m->m_epg_hdrlen) { 19167b6c99d0SGleb Smirnoff off -= m->m_epg_hdrlen; 191782334850SJohn Baldwin } else { 19187b6c99d0SGleb Smirnoff seglen = m->m_epg_hdrlen - off; 191982334850SJohn Baldwin segoff = off; 192082334850SJohn Baldwin seglen = min(seglen, len); 192182334850SJohn Baldwin off = 0; 192282334850SJohn Baldwin len -= seglen; 19230c103266SGleb Smirnoff error = uiomove(__DECONST(void *, 19240c103266SGleb Smirnoff &m->m_epg_hdr[segoff]), seglen, uio); 192582334850SJohn Baldwin } 192682334850SJohn Baldwin } 19277b6c99d0SGleb Smirnoff pgoff = m->m_epg_1st_off; 19287b6c99d0SGleb Smirnoff for (i = 0; i < m->m_epg_npgs && error == 0 && len > 0; i++) { 1929c4ee38f8SGleb Smirnoff pglen = m_epg_pagelen(m, i, pgoff); 193082334850SJohn Baldwin if (off >= pglen) { 193182334850SJohn Baldwin off -= pglen; 193282334850SJohn Baldwin pgoff = 0; 193382334850SJohn Baldwin continue; 193482334850SJohn Baldwin } 193582334850SJohn Baldwin seglen = pglen - off; 193682334850SJohn Baldwin segoff = pgoff + off; 193782334850SJohn Baldwin off = 0; 193882334850SJohn Baldwin seglen = min(seglen, len); 193982334850SJohn Baldwin len -= seglen; 19400c103266SGleb Smirnoff pg = PHYS_TO_VM_PAGE(m->m_epg_pa[i]); 194182334850SJohn Baldwin error = uiomove_fromphys(&pg, segoff, seglen, uio); 194282334850SJohn Baldwin pgoff = 0; 194382334850SJohn Baldwin }; 194482334850SJohn Baldwin if (len != 0 && error == 0) { 19457b6c99d0SGleb Smirnoff KASSERT((off + len) <= m->m_epg_trllen, 194682334850SJohn Baldwin ("off + len > trail (%d + %d > %d, m_off = %d)", off, len, 19477b6c99d0SGleb Smirnoff m->m_epg_trllen, m_off)); 19480c103266SGleb Smirnoff error = uiomove(__DECONST(void *, &m->m_epg_trail[off]), 19490c103266SGleb Smirnoff len, uio); 195082334850SJohn Baldwin } 195182334850SJohn Baldwin return (error); 195282334850SJohn Baldwin } 195382334850SJohn Baldwin 195482334850SJohn Baldwin /* 1955bc05b2f6SAndre Oppermann * Copy an mbuf chain into a uio limited by len if set. 1956bc05b2f6SAndre Oppermann */ 1957bc05b2f6SAndre Oppermann int 195814984031SGleb Smirnoff m_mbuftouio(struct uio *uio, const struct mbuf *m, int len) 1959bc05b2f6SAndre Oppermann { 1960bc05b2f6SAndre Oppermann int error, length, total; 1961bc05b2f6SAndre Oppermann int progress = 0; 1962bc05b2f6SAndre Oppermann 1963bc05b2f6SAndre Oppermann if (len > 0) 1964bc05b2f6SAndre Oppermann total = min(uio->uio_resid, len); 1965bc05b2f6SAndre Oppermann else 1966bc05b2f6SAndre Oppermann total = uio->uio_resid; 1967bc05b2f6SAndre Oppermann 1968bc05b2f6SAndre Oppermann /* Fill the uio with data from the mbufs. */ 1969bc05b2f6SAndre Oppermann for (; m != NULL; m = m->m_next) { 1970bc05b2f6SAndre Oppermann length = min(m->m_len, total - progress); 1971bc05b2f6SAndre Oppermann 19726edfd179SGleb Smirnoff if ((m->m_flags & M_EXTPG) != 0) 197382334850SJohn Baldwin error = m_unmappedtouio(m, 0, uio, length); 197482334850SJohn Baldwin else 1975bc05b2f6SAndre Oppermann error = uiomove(mtod(m, void *), length, uio); 1976bc05b2f6SAndre Oppermann if (error) 1977bc05b2f6SAndre Oppermann return (error); 1978bc05b2f6SAndre Oppermann 1979bc05b2f6SAndre Oppermann progress += length; 1980bc05b2f6SAndre Oppermann } 1981bc05b2f6SAndre Oppermann 1982bc05b2f6SAndre Oppermann return (0); 1983bc05b2f6SAndre Oppermann } 1984bc05b2f6SAndre Oppermann 1985bc05b2f6SAndre Oppermann /* 198647e2996eSSam Leffler * Create a writable copy of the mbuf chain. While doing this 198747e2996eSSam Leffler * we compact the chain with a goal of producing a chain with 198847e2996eSSam Leffler * at most two mbufs. The second mbuf in this chain is likely 198947e2996eSSam Leffler * to be a cluster. The primary purpose of this work is to create 199047e2996eSSam Leffler * a writable packet for encryption, compression, etc. The 199147e2996eSSam Leffler * secondary goal is to linearize the data so the data can be 199247e2996eSSam Leffler * passed to crypto hardware in the most efficient manner possible. 199347e2996eSSam Leffler */ 199447e2996eSSam Leffler struct mbuf * 199547e2996eSSam Leffler m_unshare(struct mbuf *m0, int how) 199647e2996eSSam Leffler { 199747e2996eSSam Leffler struct mbuf *m, *mprev; 199847e2996eSSam Leffler struct mbuf *n, *mfirst, *mlast; 199947e2996eSSam Leffler int len, off; 200047e2996eSSam Leffler 200147e2996eSSam Leffler mprev = NULL; 200247e2996eSSam Leffler for (m = m0; m != NULL; m = mprev->m_next) { 200347e2996eSSam Leffler /* 200447e2996eSSam Leffler * Regular mbufs are ignored unless there's a cluster 200547e2996eSSam Leffler * in front of it that we can use to coalesce. We do 200647e2996eSSam Leffler * the latter mainly so later clusters can be coalesced 200747e2996eSSam Leffler * also w/o having to handle them specially (i.e. convert 200847e2996eSSam Leffler * mbuf+cluster -> cluster). This optimization is heavily 200947e2996eSSam Leffler * influenced by the assumption that we're running over 201047e2996eSSam Leffler * Ethernet where MCLBYTES is large enough that the max 201147e2996eSSam Leffler * packet size will permit lots of coalescing into a 201247e2996eSSam Leffler * single cluster. This in turn permits efficient 201347e2996eSSam Leffler * crypto operations, especially when using hardware. 201447e2996eSSam Leffler */ 201547e2996eSSam Leffler if ((m->m_flags & M_EXT) == 0) { 201647e2996eSSam Leffler if (mprev && (mprev->m_flags & M_EXT) && 201747e2996eSSam Leffler m->m_len <= M_TRAILINGSPACE(mprev)) { 201847e2996eSSam Leffler /* XXX: this ignores mbuf types */ 201947e2996eSSam Leffler memcpy(mtod(mprev, caddr_t) + mprev->m_len, 202047e2996eSSam Leffler mtod(m, caddr_t), m->m_len); 202147e2996eSSam Leffler mprev->m_len += m->m_len; 202247e2996eSSam Leffler mprev->m_next = m->m_next; /* unlink from chain */ 202347e2996eSSam Leffler m_free(m); /* reclaim mbuf */ 202447e2996eSSam Leffler } else { 202547e2996eSSam Leffler mprev = m; 202647e2996eSSam Leffler } 202747e2996eSSam Leffler continue; 202847e2996eSSam Leffler } 202947e2996eSSam Leffler /* 203047e2996eSSam Leffler * Writable mbufs are left alone (for now). 203147e2996eSSam Leffler */ 203247e2996eSSam Leffler if (M_WRITABLE(m)) { 203347e2996eSSam Leffler mprev = m; 203447e2996eSSam Leffler continue; 203547e2996eSSam Leffler } 203647e2996eSSam Leffler 203747e2996eSSam Leffler /* 203847e2996eSSam Leffler * Not writable, replace with a copy or coalesce with 203947e2996eSSam Leffler * the previous mbuf if possible (since we have to copy 204047e2996eSSam Leffler * it anyway, we try to reduce the number of mbufs and 204147e2996eSSam Leffler * clusters so that future work is easier). 204247e2996eSSam Leffler */ 204347e2996eSSam Leffler KASSERT(m->m_flags & M_EXT, ("m_flags 0x%x", m->m_flags)); 204447e2996eSSam Leffler /* NB: we only coalesce into a cluster or larger */ 204547e2996eSSam Leffler if (mprev != NULL && (mprev->m_flags & M_EXT) && 204647e2996eSSam Leffler m->m_len <= M_TRAILINGSPACE(mprev)) { 204747e2996eSSam Leffler /* XXX: this ignores mbuf types */ 204847e2996eSSam Leffler memcpy(mtod(mprev, caddr_t) + mprev->m_len, 204947e2996eSSam Leffler mtod(m, caddr_t), m->m_len); 205047e2996eSSam Leffler mprev->m_len += m->m_len; 205147e2996eSSam Leffler mprev->m_next = m->m_next; /* unlink from chain */ 205247e2996eSSam Leffler m_free(m); /* reclaim mbuf */ 205347e2996eSSam Leffler continue; 205447e2996eSSam Leffler } 205547e2996eSSam Leffler 205647e2996eSSam Leffler /* 20575368b81eSGleb Smirnoff * Allocate new space to hold the copy and copy the data. 20585368b81eSGleb Smirnoff * We deal with jumbo mbufs (i.e. m_len > MCLBYTES) by 20595368b81eSGleb Smirnoff * splitting them into clusters. We could just malloc a 20605368b81eSGleb Smirnoff * buffer and make it external but too many device drivers 20615368b81eSGleb Smirnoff * don't know how to break up the non-contiguous memory when 20625368b81eSGleb Smirnoff * doing DMA. 206347e2996eSSam Leffler */ 2064fddd4f62SNavdeep Parhar n = m_getcl(how, m->m_type, m->m_flags & M_COPYFLAGS); 206547e2996eSSam Leffler if (n == NULL) { 206647e2996eSSam Leffler m_freem(m0); 206747e2996eSSam Leffler return (NULL); 206847e2996eSSam Leffler } 2069e40e8705SGleb Smirnoff if (m->m_flags & M_PKTHDR) { 2070e40e8705SGleb Smirnoff KASSERT(mprev == NULL, ("%s: m0 %p, m %p has M_PKTHDR", 2071e40e8705SGleb Smirnoff __func__, m0, m)); 2072e40e8705SGleb Smirnoff m_move_pkthdr(n, m); 2073e40e8705SGleb Smirnoff } 207447e2996eSSam Leffler len = m->m_len; 207547e2996eSSam Leffler off = 0; 207647e2996eSSam Leffler mfirst = n; 207747e2996eSSam Leffler mlast = NULL; 207847e2996eSSam Leffler for (;;) { 207947e2996eSSam Leffler int cc = min(len, MCLBYTES); 208047e2996eSSam Leffler memcpy(mtod(n, caddr_t), mtod(m, caddr_t) + off, cc); 208147e2996eSSam Leffler n->m_len = cc; 208247e2996eSSam Leffler if (mlast != NULL) 208347e2996eSSam Leffler mlast->m_next = n; 208447e2996eSSam Leffler mlast = n; 208547e2996eSSam Leffler #if 0 208647e2996eSSam Leffler newipsecstat.ips_clcopied++; 208747e2996eSSam Leffler #endif 208847e2996eSSam Leffler 208947e2996eSSam Leffler len -= cc; 209047e2996eSSam Leffler if (len <= 0) 209147e2996eSSam Leffler break; 209247e2996eSSam Leffler off += cc; 209347e2996eSSam Leffler 2094fddd4f62SNavdeep Parhar n = m_getcl(how, m->m_type, m->m_flags & M_COPYFLAGS); 209547e2996eSSam Leffler if (n == NULL) { 209647e2996eSSam Leffler m_freem(mfirst); 209747e2996eSSam Leffler m_freem(m0); 209847e2996eSSam Leffler return (NULL); 209947e2996eSSam Leffler } 210047e2996eSSam Leffler } 210147e2996eSSam Leffler n->m_next = m->m_next; 210247e2996eSSam Leffler if (mprev == NULL) 210347e2996eSSam Leffler m0 = mfirst; /* new head of chain */ 210447e2996eSSam Leffler else 210547e2996eSSam Leffler mprev->m_next = mfirst; /* replace old mbuf */ 210647e2996eSSam Leffler m_free(m); /* release old mbuf */ 210747e2996eSSam Leffler mprev = mfirst; 210847e2996eSSam Leffler } 210947e2996eSSam Leffler return (m0); 211047e2996eSSam Leffler } 21116eeac1d9SJulian Elischer 21126eeac1d9SJulian Elischer #ifdef MBUF_PROFILING 21136eeac1d9SJulian Elischer 21146eeac1d9SJulian Elischer #define MP_BUCKETS 32 /* don't just change this as things may overflow.*/ 21156eeac1d9SJulian Elischer struct mbufprofile { 21162182c0cfSJulian Elischer uintmax_t wasted[MP_BUCKETS]; 21172182c0cfSJulian Elischer uintmax_t used[MP_BUCKETS]; 21182182c0cfSJulian Elischer uintmax_t segments[MP_BUCKETS]; 21196eeac1d9SJulian Elischer } mbprof; 21206eeac1d9SJulian Elischer 21216eeac1d9SJulian Elischer #define MP_MAXDIGITS 21 /* strlen("16,000,000,000,000,000,000") == 21 */ 21226eeac1d9SJulian Elischer #define MP_NUMLINES 6 21236eeac1d9SJulian Elischer #define MP_NUMSPERLINE 16 21246eeac1d9SJulian Elischer #define MP_EXTRABYTES 64 /* > strlen("used:\nwasted:\nsegments:\n") */ 21256eeac1d9SJulian Elischer /* work out max space needed and add a bit of spare space too */ 21266eeac1d9SJulian Elischer #define MP_MAXLINE ((MP_MAXDIGITS+1) * MP_NUMSPERLINE) 21276eeac1d9SJulian Elischer #define MP_BUFSIZE ((MP_MAXLINE * MP_NUMLINES) + 1 + MP_EXTRABYTES) 21286eeac1d9SJulian Elischer 21296eeac1d9SJulian Elischer char mbprofbuf[MP_BUFSIZE]; 21306eeac1d9SJulian Elischer 21316eeac1d9SJulian Elischer void 21326eeac1d9SJulian Elischer m_profile(struct mbuf *m) 21336eeac1d9SJulian Elischer { 21346eeac1d9SJulian Elischer int segments = 0; 21356eeac1d9SJulian Elischer int used = 0; 21366eeac1d9SJulian Elischer int wasted = 0; 21376eeac1d9SJulian Elischer 21386eeac1d9SJulian Elischer while (m) { 21396eeac1d9SJulian Elischer segments++; 21406eeac1d9SJulian Elischer used += m->m_len; 21416eeac1d9SJulian Elischer if (m->m_flags & M_EXT) { 21426eeac1d9SJulian Elischer wasted += MHLEN - sizeof(m->m_ext) + 21436eeac1d9SJulian Elischer m->m_ext.ext_size - m->m_len; 21446eeac1d9SJulian Elischer } else { 21456eeac1d9SJulian Elischer if (m->m_flags & M_PKTHDR) 21466eeac1d9SJulian Elischer wasted += MHLEN - m->m_len; 21476eeac1d9SJulian Elischer else 21486eeac1d9SJulian Elischer wasted += MLEN - m->m_len; 21496eeac1d9SJulian Elischer } 21506eeac1d9SJulian Elischer m = m->m_next; 21516eeac1d9SJulian Elischer } 21526eeac1d9SJulian Elischer /* be paranoid.. it helps */ 21536eeac1d9SJulian Elischer if (segments > MP_BUCKETS - 1) 21546eeac1d9SJulian Elischer segments = MP_BUCKETS - 1; 21556eeac1d9SJulian Elischer if (used > 100000) 21566eeac1d9SJulian Elischer used = 100000; 21576eeac1d9SJulian Elischer if (wasted > 100000) 21586eeac1d9SJulian Elischer wasted = 100000; 21596eeac1d9SJulian Elischer /* store in the appropriate bucket */ 21606eeac1d9SJulian Elischer /* don't bother locking. if it's slightly off, so what? */ 21616eeac1d9SJulian Elischer mbprof.segments[segments]++; 21626eeac1d9SJulian Elischer mbprof.used[fls(used)]++; 21636eeac1d9SJulian Elischer mbprof.wasted[fls(wasted)]++; 21646eeac1d9SJulian Elischer } 21656eeac1d9SJulian Elischer 21666eeac1d9SJulian Elischer static void 21676eeac1d9SJulian Elischer mbprof_textify(void) 21686eeac1d9SJulian Elischer { 21696eeac1d9SJulian Elischer int offset; 21706eeac1d9SJulian Elischer char *c; 217160ae52f7SEd Schouten uint64_t *p; 21726eeac1d9SJulian Elischer 21736eeac1d9SJulian Elischer p = &mbprof.wasted[0]; 21746eeac1d9SJulian Elischer c = mbprofbuf; 21756eeac1d9SJulian Elischer offset = snprintf(c, MP_MAXLINE + 10, 21766eeac1d9SJulian Elischer "wasted:\n" 21772182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 21782182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 21796eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 21806eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 21816eeac1d9SJulian Elischer #ifdef BIG_ARRAY 21826eeac1d9SJulian Elischer p = &mbprof.wasted[16]; 21836eeac1d9SJulian Elischer c += offset; 21846eeac1d9SJulian Elischer offset = snprintf(c, MP_MAXLINE, 21852182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 21862182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 21876eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 21886eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 21896eeac1d9SJulian Elischer #endif 21906eeac1d9SJulian Elischer p = &mbprof.used[0]; 21916eeac1d9SJulian Elischer c += offset; 21926eeac1d9SJulian Elischer offset = snprintf(c, MP_MAXLINE + 10, 21936eeac1d9SJulian Elischer "used:\n" 21942182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 21952182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 21966eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 21976eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 21986eeac1d9SJulian Elischer #ifdef BIG_ARRAY 21996eeac1d9SJulian Elischer p = &mbprof.used[16]; 22006eeac1d9SJulian Elischer c += offset; 22016eeac1d9SJulian Elischer offset = snprintf(c, MP_MAXLINE, 22022182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 22032182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 22046eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 22056eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 22066eeac1d9SJulian Elischer #endif 22076eeac1d9SJulian Elischer p = &mbprof.segments[0]; 22086eeac1d9SJulian Elischer c += offset; 22096eeac1d9SJulian Elischer offset = snprintf(c, MP_MAXLINE + 10, 22106eeac1d9SJulian Elischer "segments:\n" 22112182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 22122182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju\n", 22136eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 22146eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 22156eeac1d9SJulian Elischer #ifdef BIG_ARRAY 22166eeac1d9SJulian Elischer p = &mbprof.segments[16]; 22176eeac1d9SJulian Elischer c += offset; 22186eeac1d9SJulian Elischer offset = snprintf(c, MP_MAXLINE, 22192182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %ju " 22202182c0cfSJulian Elischer "%ju %ju %ju %ju %ju %ju %ju %jju", 22216eeac1d9SJulian Elischer p[0], p[1], p[2], p[3], p[4], p[5], p[6], p[7], 22226eeac1d9SJulian Elischer p[8], p[9], p[10], p[11], p[12], p[13], p[14], p[15]); 22236eeac1d9SJulian Elischer #endif 22246eeac1d9SJulian Elischer } 22256eeac1d9SJulian Elischer 22266eeac1d9SJulian Elischer static int 22276eeac1d9SJulian Elischer mbprof_handler(SYSCTL_HANDLER_ARGS) 22286eeac1d9SJulian Elischer { 22296eeac1d9SJulian Elischer int error; 22306eeac1d9SJulian Elischer 22316eeac1d9SJulian Elischer mbprof_textify(); 22326eeac1d9SJulian Elischer error = SYSCTL_OUT(req, mbprofbuf, strlen(mbprofbuf) + 1); 22336eeac1d9SJulian Elischer return (error); 22346eeac1d9SJulian Elischer } 22356eeac1d9SJulian Elischer 22366eeac1d9SJulian Elischer static int 22376eeac1d9SJulian Elischer mbprof_clr_handler(SYSCTL_HANDLER_ARGS) 22386eeac1d9SJulian Elischer { 22396eeac1d9SJulian Elischer int clear, error; 22406eeac1d9SJulian Elischer 22416eeac1d9SJulian Elischer clear = 0; 22426eeac1d9SJulian Elischer error = sysctl_handle_int(oidp, &clear, 0, req); 22436eeac1d9SJulian Elischer if (error || !req->newptr) 22446eeac1d9SJulian Elischer return (error); 22456eeac1d9SJulian Elischer 22466eeac1d9SJulian Elischer if (clear) { 22476eeac1d9SJulian Elischer bzero(&mbprof, sizeof(mbprof)); 22486eeac1d9SJulian Elischer } 22496eeac1d9SJulian Elischer 22506eeac1d9SJulian Elischer return (error); 22516eeac1d9SJulian Elischer } 22526eeac1d9SJulian Elischer 22537029da5cSPawel Biernacki SYSCTL_PROC(_kern_ipc, OID_AUTO, mbufprofile, 22547029da5cSPawel Biernacki CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_NEEDGIANT, NULL, 0, 22557029da5cSPawel Biernacki mbprof_handler, "A", 22567029da5cSPawel Biernacki "mbuf profiling statistics"); 22576eeac1d9SJulian Elischer 22587029da5cSPawel Biernacki SYSCTL_PROC(_kern_ipc, OID_AUTO, mbufprofileclr, 22597029da5cSPawel Biernacki CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, NULL, 0, 22607029da5cSPawel Biernacki mbprof_clr_handler, "I", 22617029da5cSPawel Biernacki "clear mbuf profiling statistics"); 22626eeac1d9SJulian Elischer #endif 2263