xref: /freebsd/usr.sbin/ppp/lqr.c (revision 1e413cf93298b5b97441a21d9a50fdcd0ee9945e)
1 /*-
2  * Copyright (c) 1996 - 2001 Brian Somers <brian@Awfulhak.org>
3  *          based on work by Toshiharu OHNO <tony-o@iij.ad.jp>
4  *                           Internet Initiative Japan, Inc (IIJ)
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  *
28  * $FreeBSD$
29  */
30 
31 #include <sys/param.h>
32 
33 #ifdef __FreeBSD__
34 #include <netinet/in.h>
35 #endif
36 #include <sys/un.h>
37 
38 #include <string.h>
39 #include <termios.h>
40 
41 #include "layer.h"
42 #include "mbuf.h"
43 #include "log.h"
44 #include "defs.h"
45 #include "timer.h"
46 #include "fsm.h"
47 #include "acf.h"
48 #include "proto.h"
49 #include "lqr.h"
50 #include "hdlc.h"
51 #include "lcp.h"
52 #include "async.h"
53 #include "throughput.h"
54 #include "ccp.h"
55 #include "link.h"
56 #include "descriptor.h"
57 #include "physical.h"
58 #include "mp.h"
59 #include "chat.h"
60 #include "auth.h"
61 #include "chap.h"
62 #include "command.h"
63 #include "cbcp.h"
64 #include "datalink.h"
65 
66 struct echolqr {
67   u_int32_t magic;
68   u_int32_t signature;
69   u_int32_t sequence;
70 };
71 
72 #define	SIGNATURE  0x594e4f54
73 
74 static void
75 SendEchoReq(struct lcp *lcp)
76 {
77   struct hdlc *hdlc = &link2physical(lcp->fsm.link)->hdlc;
78   struct echolqr echo;
79 
80   echo.magic = htonl(lcp->want_magic);
81   echo.signature = htonl(SIGNATURE);
82   echo.sequence = htonl(hdlc->lqm.echo.seq_sent);
83   fsm_Output(&lcp->fsm, CODE_ECHOREQ, hdlc->lqm.echo.seq_sent++,
84             (u_char *)&echo, sizeof echo, MB_ECHOOUT);
85 }
86 
87 struct mbuf *
88 lqr_RecvEcho(struct fsm *fp, struct mbuf *bp)
89 {
90   struct hdlc *hdlc = &link2physical(fp->link)->hdlc;
91   struct lcp *lcp = fsm2lcp(fp);
92   struct echolqr lqr;
93 
94   if (m_length(bp) >= sizeof lqr) {
95     m_freem(mbuf_Read(bp, &lqr, sizeof lqr));
96     bp = NULL;
97     lqr.magic = ntohl(lqr.magic);
98     lqr.signature = ntohl(lqr.signature);
99     lqr.sequence = ntohl(lqr.sequence);
100 
101     /* Tolerate echo replies with either magic number */
102     if (lqr.magic != 0 && lqr.magic != lcp->his_magic &&
103         lqr.magic != lcp->want_magic) {
104       log_Printf(LogWARN, "%s: lqr_RecvEcho: Bad magic: expected 0x%08x,"
105                  " got 0x%08x\n", fp->link->name, lcp->his_magic, lqr.magic);
106       /*
107        * XXX: We should send a terminate request. But poor implementations may
108        *      die as a result.
109        */
110     }
111     if (lqr.signature == SIGNATURE) {
112       /* careful not to update lqm.echo.seq_recv with older values */
113       if ((hdlc->lqm.echo.seq_recv > (u_int32_t)0 - 5 && lqr.sequence < 5) ||
114           (hdlc->lqm.echo.seq_recv <= (u_int32_t)0 - 5 &&
115            lqr.sequence > hdlc->lqm.echo.seq_recv))
116         hdlc->lqm.echo.seq_recv = lqr.sequence;
117     } else
118       log_Printf(LogWARN, "lqr_RecvEcho: Got sig 0x%08lx, not 0x%08lx !\n",
119                 (u_long)lqr.signature, (u_long)SIGNATURE);
120   } else
121     log_Printf(LogWARN, "lqr_RecvEcho: Got packet size %zd, expecting %ld !\n",
122               m_length(bp), (long)sizeof(struct echolqr));
123   return bp;
124 }
125 
126 void
127 lqr_ChangeOrder(struct lqrdata *src, struct lqrdata *dst)
128 {
129   u_int32_t *sp, *dp;
130   unsigned n;
131 
132   sp = (u_int32_t *) src;
133   dp = (u_int32_t *) dst;
134   for (n = 0; n < sizeof(struct lqrdata) / sizeof(u_int32_t); n++, sp++, dp++)
135     *dp = ntohl(*sp);
136 }
137 
138 static void
139 SendLqrData(struct lcp *lcp)
140 {
141   struct mbuf *bp;
142   int extra;
143 
144   extra = proto_WrapperOctets(lcp, PROTO_LQR) +
145           acf_WrapperOctets(lcp, PROTO_LQR);
146   bp = m_get(sizeof(struct lqrdata) + extra, MB_LQROUT);
147   bp->m_len -= extra;
148   bp->m_offset += extra;
149 
150   /*
151    * Send on the highest priority queue.  We send garbage - the real data
152    * is written by lqr_LayerPush() where we know how to fill in all the
153    * fields.  Note, lqr_LayerPush() ``knows'' that we're pushing onto the
154    * highest priority queue, and factors out packet & octet values from
155    * other queues!
156    */
157   link_PushPacket(lcp->fsm.link, bp, lcp->fsm.bundle,
158                   LINK_QUEUES(lcp->fsm.link) - 1, PROTO_LQR);
159 }
160 
161 static void
162 SendLqrReport(void *v)
163 {
164   struct lcp *lcp = (struct lcp *)v;
165   struct physical *p = link2physical(lcp->fsm.link);
166 
167   timer_Stop(&p->hdlc.lqm.timer);
168 
169   if (p->hdlc.lqm.method & LQM_LQR) {
170     if (p->hdlc.lqm.lqr.resent > 5) {
171       /* XXX: Should implement LQM strategy */
172       log_Printf(LogPHASE, "%s: ** Too many LQR packets lost **\n",
173                 lcp->fsm.link->name);
174       log_Printf(LogLQM, "%s: Too many LQR packets lost\n",
175                 lcp->fsm.link->name);
176       p->hdlc.lqm.method = 0;
177       datalink_Down(p->dl, CLOSE_NORMAL);
178     } else {
179       SendLqrData(lcp);
180       p->hdlc.lqm.lqr.resent++;
181     }
182   } else if (p->hdlc.lqm.method & LQM_ECHO) {
183     if ((p->hdlc.lqm.echo.seq_sent > 5 &&
184          p->hdlc.lqm.echo.seq_sent - 5 > p->hdlc.lqm.echo.seq_recv) ||
185         (p->hdlc.lqm.echo.seq_sent <= 5 &&
186          p->hdlc.lqm.echo.seq_sent > p->hdlc.lqm.echo.seq_recv + 5)) {
187       log_Printf(LogPHASE, "%s: ** Too many LCP ECHO packets lost **\n",
188                 lcp->fsm.link->name);
189       log_Printf(LogLQM, "%s: Too many LCP ECHO packets lost\n",
190                 lcp->fsm.link->name);
191       p->hdlc.lqm.method = 0;
192       datalink_Down(p->dl, CLOSE_NORMAL);
193     } else
194       SendEchoReq(lcp);
195   }
196   if (p->hdlc.lqm.method && p->hdlc.lqm.timer.load)
197     timer_Start(&p->hdlc.lqm.timer);
198 }
199 
200 struct mbuf *
201 lqr_Input(struct bundle *bundle __unused, struct link *l, struct mbuf *bp)
202 {
203   struct physical *p = link2physical(l);
204   struct lcp *lcp = p->hdlc.lqm.owner;
205   int len;
206 
207   if (p == NULL) {
208     log_Printf(LogERROR, "lqr_Input: Not a physical link - dropped\n");
209     m_freem(bp);
210     return NULL;
211   }
212 
213   len = m_length(bp);
214   if (len != sizeof(struct lqrdata))
215     log_Printf(LogWARN, "lqr_Input: Got packet size %d, expecting %ld !\n",
216               len, (long)sizeof(struct lqrdata));
217   else if (!IsAccepted(l->lcp.cfg.lqr) && !(p->hdlc.lqm.method & LQM_LQR)) {
218     bp = m_pullup(proto_Prepend(bp, PROTO_LQR, 0, 0));
219     lcp_SendProtoRej(lcp, MBUF_CTOP(bp), bp->m_len);
220   } else {
221     struct lqrdata *lqr;
222 
223     bp = m_pullup(bp);
224     lqr = (struct lqrdata *)MBUF_CTOP(bp);
225     if (ntohl(lqr->MagicNumber) != lcp->his_magic)
226       log_Printf(LogWARN, "lqr_Input: magic 0x%08lx is wrong,"
227                  " expecting 0x%08lx\n",
228 		 (u_long)ntohl(lqr->MagicNumber), (u_long)lcp->his_magic);
229     else {
230       struct lqrdata lastlqr;
231 
232       memcpy(&lastlqr, &p->hdlc.lqm.lqr.peer, sizeof lastlqr);
233       lqr_ChangeOrder(lqr, &p->hdlc.lqm.lqr.peer);
234       lqr_Dump(l->name, "Input", &p->hdlc.lqm.lqr.peer);
235       /* we have received an LQR from our peer */
236       p->hdlc.lqm.lqr.resent = 0;
237 
238       /* Snapshot our state when the LQR packet was received */
239       memcpy(&p->hdlc.lqm.lqr.prevSave, &p->hdlc.lqm.lqr.Save,
240              sizeof p->hdlc.lqm.lqr.prevSave);
241       p->hdlc.lqm.lqr.Save.InLQRs = ++p->hdlc.lqm.lqr.InLQRs;
242       p->hdlc.lqm.lqr.Save.InPackets = p->hdlc.lqm.ifInUniPackets;
243       p->hdlc.lqm.lqr.Save.InDiscards = p->hdlc.lqm.ifInDiscards;
244       p->hdlc.lqm.lqr.Save.InErrors = p->hdlc.lqm.ifInErrors;
245       p->hdlc.lqm.lqr.Save.InOctets = p->hdlc.lqm.lqr.InGoodOctets;
246 
247       lqr_Analyse(&p->hdlc, &lastlqr, &p->hdlc.lqm.lqr.peer);
248 
249       /*
250        * Generate an LQR response if we're not running an LQR timer OR
251        * two successive LQR's PeerInLQRs are the same.
252        */
253       if (p->hdlc.lqm.timer.load == 0 || !(p->hdlc.lqm.method & LQM_LQR) ||
254           (lastlqr.PeerInLQRs &&
255            lastlqr.PeerInLQRs == p->hdlc.lqm.lqr.peer.PeerInLQRs))
256         SendLqrData(lcp);
257     }
258   }
259   m_freem(bp);
260   return NULL;
261 }
262 
263 /*
264  *  When LCP is reached to opened state, We'll start LQM activity.
265  */
266 static void
267 lqr_Setup(struct lcp *lcp)
268 {
269   struct physical *physical = link2physical(lcp->fsm.link);
270   int period;
271 
272   physical->hdlc.lqm.lqr.resent = 0;
273   physical->hdlc.lqm.echo.seq_sent = 0;
274   physical->hdlc.lqm.echo.seq_recv = 0;
275   memset(&physical->hdlc.lqm.lqr.peer, '\0',
276          sizeof physical->hdlc.lqm.lqr.peer);
277 
278   physical->hdlc.lqm.method = lcp->cfg.echo ? LQM_ECHO : 0;
279   if (IsEnabled(lcp->cfg.lqr) && !REJECTED(lcp, TY_QUALPROTO))
280     physical->hdlc.lqm.method |= LQM_LQR;
281   timer_Stop(&physical->hdlc.lqm.timer);
282 
283   physical->hdlc.lqm.lqr.peer_timeout = lcp->his_lqrperiod;
284   if (lcp->his_lqrperiod)
285     log_Printf(LogLQM, "%s: Expecting LQR every %d.%02d secs\n",
286               physical->link.name, lcp->his_lqrperiod / 100,
287               lcp->his_lqrperiod % 100);
288 
289   period = lcp->want_lqrperiod ?
290     lcp->want_lqrperiod : lcp->cfg.lqrperiod * 100;
291   physical->hdlc.lqm.timer.func = SendLqrReport;
292   physical->hdlc.lqm.timer.name = "lqm";
293   physical->hdlc.lqm.timer.arg = lcp;
294 
295   if (lcp->want_lqrperiod || physical->hdlc.lqm.method & LQM_ECHO) {
296     log_Printf(LogLQM, "%s: Will send %s every %d.%02d secs\n",
297               physical->link.name, lcp->want_lqrperiod ? "LQR" : "LCP ECHO",
298               period / 100, period % 100);
299     physical->hdlc.lqm.timer.load = period * SECTICKS / 100;
300   } else {
301     physical->hdlc.lqm.timer.load = 0;
302     if (!lcp->his_lqrperiod)
303       log_Printf(LogLQM, "%s: LQR/LCP ECHO not negotiated\n",
304                  physical->link.name);
305   }
306 }
307 
308 void
309 lqr_Start(struct lcp *lcp)
310 {
311   struct physical *p = link2physical(lcp->fsm.link);
312 
313   lqr_Setup(lcp);
314   if (p->hdlc.lqm.timer.load)
315     SendLqrReport(lcp);
316 }
317 
318 void
319 lqr_reStart(struct lcp *lcp)
320 {
321   struct physical *p = link2physical(lcp->fsm.link);
322 
323   lqr_Setup(lcp);
324   if (p->hdlc.lqm.timer.load)
325     timer_Start(&p->hdlc.lqm.timer);
326 }
327 
328 void
329 lqr_StopTimer(struct physical *physical)
330 {
331   timer_Stop(&physical->hdlc.lqm.timer);
332 }
333 
334 void
335 lqr_Stop(struct physical *physical, int method)
336 {
337   if (method == LQM_LQR)
338     log_Printf(LogLQM, "%s: Stop sending LQR, Use LCP ECHO instead.\n",
339                physical->link.name);
340   if (method == LQM_ECHO)
341     log_Printf(LogLQM, "%s: Stop sending LCP ECHO.\n",
342                physical->link.name);
343   physical->hdlc.lqm.method &= ~method;
344   if (physical->hdlc.lqm.method)
345     SendLqrReport(physical->hdlc.lqm.owner);
346   else
347     timer_Stop(&physical->hdlc.lqm.timer);
348 }
349 
350 void
351 lqr_Dump(const char *link, const char *message, const struct lqrdata *lqr)
352 {
353   if (log_IsKept(LogLQM)) {
354     log_Printf(LogLQM, "%s: %s:\n", link, message);
355     log_Printf(LogLQM, "  Magic:          %08x   LastOutLQRs:    %08x\n",
356 	      lqr->MagicNumber, lqr->LastOutLQRs);
357     log_Printf(LogLQM, "  LastOutPackets: %08x   LastOutOctets:  %08x\n",
358 	      lqr->LastOutPackets, lqr->LastOutOctets);
359     log_Printf(LogLQM, "  PeerInLQRs:     %08x   PeerInPackets:  %08x\n",
360 	      lqr->PeerInLQRs, lqr->PeerInPackets);
361     log_Printf(LogLQM, "  PeerInDiscards: %08x   PeerInErrors:   %08x\n",
362 	      lqr->PeerInDiscards, lqr->PeerInErrors);
363     log_Printf(LogLQM, "  PeerInOctets:   %08x   PeerOutLQRs:    %08x\n",
364 	      lqr->PeerInOctets, lqr->PeerOutLQRs);
365     log_Printf(LogLQM, "  PeerOutPackets: %08x   PeerOutOctets:  %08x\n",
366 	      lqr->PeerOutPackets, lqr->PeerOutOctets);
367   }
368 }
369 
370 void
371 lqr_Analyse(const struct hdlc *hdlc, const struct lqrdata *oldlqr,
372             const struct lqrdata *newlqr)
373 {
374   u_int32_t LQRs, transitLQRs, pkts, octets, disc, err;
375 
376   if (!newlqr->PeerInLQRs)	/* No analysis possible yet! */
377     return;
378 
379   log_Printf(LogLQM, "Analysis:\n");
380 
381   LQRs = (newlqr->LastOutLQRs - oldlqr->LastOutLQRs) -
382          (newlqr->PeerInLQRs - oldlqr->PeerInLQRs);
383   transitLQRs = hdlc->lqm.lqr.OutLQRs - newlqr->LastOutLQRs;
384   pkts = (newlqr->LastOutPackets - oldlqr->LastOutPackets) -
385          (newlqr->PeerInPackets - oldlqr->PeerInPackets);
386   octets = (newlqr->LastOutOctets - oldlqr->LastOutOctets) -
387            (newlqr->PeerInOctets - oldlqr->PeerInOctets);
388   log_Printf(LogLQM, "  Outbound lossage: %d LQR%s (%d en route), %d packet%s,"
389              " %d octet%s\n", (int)LQRs, LQRs == 1 ? "" : "s", (int)transitLQRs,
390 	     (int)pkts, pkts == 1 ? "" : "s",
391 	     (int)octets, octets == 1 ? "" : "s");
392 
393   pkts = (newlqr->PeerOutPackets - oldlqr->PeerOutPackets) -
394     (hdlc->lqm.lqr.Save.InPackets - hdlc->lqm.lqr.prevSave.InPackets);
395   octets = (newlqr->PeerOutOctets - oldlqr->PeerOutOctets) -
396     (hdlc->lqm.lqr.Save.InOctets - hdlc->lqm.lqr.prevSave.InOctets);
397   log_Printf(LogLQM, "  Inbound lossage: %d packet%s, %d octet%s\n",
398 	     (int)pkts, pkts == 1 ? "" : "s",
399 	     (int)octets, octets == 1 ? "" : "s");
400 
401   disc = newlqr->PeerInDiscards - oldlqr->PeerInDiscards;
402   err = newlqr->PeerInErrors - oldlqr->PeerInErrors;
403   if (disc && err)
404     log_Printf(LogLQM, "                   Likely due to both peer congestion"
405                " and physical errors\n");
406   else if (disc)
407     log_Printf(LogLQM, "                   Likely due to peer congestion\n");
408   else if (err)
409     log_Printf(LogLQM, "                   Likely due to physical errors\n");
410   else if (pkts)
411     log_Printf(LogLQM, "                   Likely due to transport "
412 	       "congestion\n");
413 }
414 
415 static struct mbuf *
416 lqr_LayerPush(struct bundle *b __unused, struct link *l, struct mbuf *bp,
417               int pri __unused, u_short *proto)
418 {
419   struct physical *p = link2physical(l);
420   int len, layer, extra_async_bytes;
421 
422   if (!p) {
423     /* Oops - can't happen :-] */
424     m_freem(bp);
425     return NULL;
426   }
427 
428   bp = m_pullup(bp);
429   len = m_length(bp);
430 
431   /*-
432    * From rfc1989:
433    *
434    *  All octets which are included in the FCS calculation MUST be counted,
435    *  including the packet header, the information field, and any padding.
436    *  The FCS octets MUST also be counted, and one flag octet per frame
437    *  MUST be counted.  All other octets (such as additional flag
438    *  sequences, and escape bits or octets) MUST NOT be counted.
439    *
440    * As we're stacked higher than the HDLC layer (otherwise HDLC wouldn't be
441    * able to calculate the FCS), we must not forget about these additional
442    * bytes when we're asynchronous.
443    *
444    * We're also expecting to be stacked *before* the likes of the proto and
445    * acf layers (to avoid alignment issues), so deal with this too.
446    */
447 
448   extra_async_bytes = 0;
449   p->hdlc.lqm.ifOutUniPackets++;
450   p->hdlc.lqm.ifOutOctets += len + 1;		/* plus 1 flag octet! */
451   for (layer = 0; layer < l->nlayers; layer++)
452     switch (l->layer[layer]->type) {
453       case LAYER_ACF:
454         p->hdlc.lqm.ifOutOctets += acf_WrapperOctets(&l->lcp, *proto);
455         break;
456       case LAYER_ASYNC:
457         /* Not included - see rfc1989 */
458         break;
459       case LAYER_HDLC:
460         p->hdlc.lqm.ifOutOctets += hdlc_WrapperOctets();
461         break;
462       case LAYER_LQR:
463         layer = l->nlayers;
464         break;
465       case LAYER_PROTO:
466         p->hdlc.lqm.ifOutOctets += proto_WrapperOctets(&l->lcp, *proto);
467         break;
468       case LAYER_SYNC:
469         /* Nothing to add on */
470         break;
471       default:
472         log_Printf(LogWARN, "Oops, don't know how to do octets for %s layer\n",
473                    l->layer[layer]->name);
474         break;
475     }
476 
477   if (*proto == PROTO_LQR) {
478     /* Overwrite the entire packet (created in SendLqrData()) */
479     struct lqrdata lqr;
480     size_t pending_pkts, pending_octets;
481 
482     p->hdlc.lqm.lqr.OutLQRs++;
483 
484     /*
485      * We need to compensate for the fact that we're pushing our data
486      * onto the highest priority queue by factoring out packet & octet
487      * values from other queues!
488      */
489     link_PendingLowPriorityData(l, &pending_pkts, &pending_octets);
490 
491     memset(&lqr, '\0', sizeof lqr);
492     lqr.MagicNumber = p->link.lcp.want_magic;
493     lqr.LastOutLQRs = p->hdlc.lqm.lqr.peer.PeerOutLQRs;
494     lqr.LastOutPackets = p->hdlc.lqm.lqr.peer.PeerOutPackets;
495     lqr.LastOutOctets = p->hdlc.lqm.lqr.peer.PeerOutOctets;
496     lqr.PeerInLQRs = p->hdlc.lqm.lqr.Save.InLQRs;
497     lqr.PeerInPackets = p->hdlc.lqm.lqr.Save.InPackets;
498     lqr.PeerInDiscards = p->hdlc.lqm.lqr.Save.InDiscards;
499     lqr.PeerInErrors = p->hdlc.lqm.lqr.Save.InErrors;
500     lqr.PeerInOctets = p->hdlc.lqm.lqr.Save.InOctets;
501     lqr.PeerOutLQRs = p->hdlc.lqm.lqr.OutLQRs;
502     lqr.PeerOutPackets = p->hdlc.lqm.ifOutUniPackets - pending_pkts;
503     /* Don't forget our ``flag'' octets.... */
504     lqr.PeerOutOctets = p->hdlc.lqm.ifOutOctets - pending_octets - pending_pkts;
505     lqr_Dump(l->name, "Output", &lqr);
506     lqr_ChangeOrder(&lqr, (struct lqrdata *)MBUF_CTOP(bp));
507   }
508 
509   return bp;
510 }
511 
512 static struct mbuf *
513 lqr_LayerPull(struct bundle *b __unused, struct link *l __unused,
514 	      struct mbuf *bp, u_short *proto)
515 {
516   /*
517    * This is the ``Rx'' process from rfc1989, although a part of it is
518    * actually performed by sync_LayerPull() & hdlc_LayerPull() so that
519    * our octet counts are correct.
520    */
521 
522   if (*proto == PROTO_LQR)
523     m_settype(bp, MB_LQRIN);
524   return bp;
525 }
526 
527 /*
528  * Statistics for pulled packets are recorded either in hdlc_PullPacket()
529  * or sync_PullPacket()
530  */
531 
532 struct layer lqrlayer = { LAYER_LQR, "lqr", lqr_LayerPush, lqr_LayerPull };
533