1 /* 2 * 3 * Author Karsten Keil <kkeil@novell.com> 4 * 5 * Copyright 2008 by Karsten Keil <kkeil@novell.com> 6 * 7 * This program is free software; you can redistribute it and/or modify 8 * it under the terms of the GNU General Public License version 2 as 9 * published by the Free Software Foundation. 10 * 11 * This program is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 * GNU General Public License for more details. 15 * 16 */ 17 18 #include "fsm.h" 19 #include "layer2.h" 20 21 static int *debug; 22 23 static 24 struct Fsm l2fsm = {NULL, 0, 0, NULL, NULL}; 25 26 static char *strL2State[] = 27 { 28 "ST_L2_1", 29 "ST_L2_2", 30 "ST_L2_3", 31 "ST_L2_4", 32 "ST_L2_5", 33 "ST_L2_6", 34 "ST_L2_7", 35 "ST_L2_8", 36 }; 37 38 enum { 39 EV_L2_UI, 40 EV_L2_SABME, 41 EV_L2_DISC, 42 EV_L2_DM, 43 EV_L2_UA, 44 EV_L2_FRMR, 45 EV_L2_SUPER, 46 EV_L2_I, 47 EV_L2_DL_DATA, 48 EV_L2_ACK_PULL, 49 EV_L2_DL_UNITDATA, 50 EV_L2_DL_ESTABLISH_REQ, 51 EV_L2_DL_RELEASE_REQ, 52 EV_L2_MDL_ASSIGN, 53 EV_L2_MDL_REMOVE, 54 EV_L2_MDL_ERROR, 55 EV_L1_DEACTIVATE, 56 EV_L2_T200, 57 EV_L2_T203, 58 EV_L2_SET_OWN_BUSY, 59 EV_L2_CLEAR_OWN_BUSY, 60 EV_L2_FRAME_ERROR, 61 }; 62 63 #define L2_EVENT_COUNT (EV_L2_FRAME_ERROR+1) 64 65 static char *strL2Event[] = 66 { 67 "EV_L2_UI", 68 "EV_L2_SABME", 69 "EV_L2_DISC", 70 "EV_L2_DM", 71 "EV_L2_UA", 72 "EV_L2_FRMR", 73 "EV_L2_SUPER", 74 "EV_L2_I", 75 "EV_L2_DL_DATA", 76 "EV_L2_ACK_PULL", 77 "EV_L2_DL_UNITDATA", 78 "EV_L2_DL_ESTABLISH_REQ", 79 "EV_L2_DL_RELEASE_REQ", 80 "EV_L2_MDL_ASSIGN", 81 "EV_L2_MDL_REMOVE", 82 "EV_L2_MDL_ERROR", 83 "EV_L1_DEACTIVATE", 84 "EV_L2_T200", 85 "EV_L2_T203", 86 "EV_L2_SET_OWN_BUSY", 87 "EV_L2_CLEAR_OWN_BUSY", 88 "EV_L2_FRAME_ERROR", 89 }; 90 91 static void 92 l2m_debug(struct FsmInst *fi, char *fmt, ...) 93 { 94 struct layer2 *l2 = fi->userdata; 95 va_list va; 96 97 if (!(*debug & DEBUG_L2_FSM)) 98 return; 99 va_start(va, fmt); 100 printk(KERN_DEBUG "l2 (tei %d): ", l2->tei); 101 vprintk(fmt, va); 102 printk("\n"); 103 va_end(va); 104 } 105 106 inline u_int 107 l2headersize(struct layer2 *l2, int ui) 108 { 109 return ((test_bit(FLG_MOD128, &l2->flag) && (!ui)) ? 2 : 1) + 110 (test_bit(FLG_LAPD, &l2->flag) ? 2 : 1); 111 } 112 113 inline u_int 114 l2addrsize(struct layer2 *l2) 115 { 116 return test_bit(FLG_LAPD, &l2->flag) ? 2 : 1; 117 } 118 119 static u_int 120 l2_newid(struct layer2 *l2) 121 { 122 u_int id; 123 124 id = l2->next_id++; 125 if (id == 0x7fff) 126 l2->next_id = 1; 127 id <<= 16; 128 id |= l2->tei << 8; 129 id |= l2->sapi; 130 return id; 131 } 132 133 static void 134 l2up(struct layer2 *l2, u_int prim, struct sk_buff *skb) 135 { 136 int err; 137 138 if (!l2->up) 139 return; 140 mISDN_HEAD_PRIM(skb) = prim; 141 mISDN_HEAD_ID(skb) = (l2->ch.nr << 16) | l2->ch.addr; 142 err = l2->up->send(l2->up, skb); 143 if (err) { 144 printk(KERN_WARNING "%s: err=%d\n", __func__, err); 145 dev_kfree_skb(skb); 146 } 147 } 148 149 static void 150 l2up_create(struct layer2 *l2, u_int prim, int len, void *arg) 151 { 152 struct sk_buff *skb; 153 struct mISDNhead *hh; 154 int err; 155 156 if (!l2->up) 157 return; 158 skb = mI_alloc_skb(len, GFP_ATOMIC); 159 if (!skb) 160 return; 161 hh = mISDN_HEAD_P(skb); 162 hh->prim = prim; 163 hh->id = (l2->ch.nr << 16) | l2->ch.addr; 164 if (len) 165 memcpy(skb_put(skb, len), arg, len); 166 err = l2->up->send(l2->up, skb); 167 if (err) { 168 printk(KERN_WARNING "%s: err=%d\n", __func__, err); 169 dev_kfree_skb(skb); 170 } 171 } 172 173 static int 174 l2down_skb(struct layer2 *l2, struct sk_buff *skb) { 175 int ret; 176 177 ret = l2->ch.recv(l2->ch.peer, skb); 178 if (ret && (*debug & DEBUG_L2_RECV)) 179 printk(KERN_DEBUG "l2down_skb: ret(%d)\n", ret); 180 return ret; 181 } 182 183 static int 184 l2down_raw(struct layer2 *l2, struct sk_buff *skb) 185 { 186 struct mISDNhead *hh = mISDN_HEAD_P(skb); 187 188 if (hh->prim == PH_DATA_REQ) { 189 if (test_and_set_bit(FLG_L1_NOTREADY, &l2->flag)) { 190 skb_queue_tail(&l2->down_queue, skb); 191 return 0; 192 } 193 l2->down_id = mISDN_HEAD_ID(skb); 194 } 195 return l2down_skb(l2, skb); 196 } 197 198 static int 199 l2down(struct layer2 *l2, u_int prim, u_int id, struct sk_buff *skb) 200 { 201 struct mISDNhead *hh = mISDN_HEAD_P(skb); 202 203 hh->prim = prim; 204 hh->id = id; 205 return l2down_raw(l2, skb); 206 } 207 208 static int 209 l2down_create(struct layer2 *l2, u_int prim, u_int id, int len, void *arg) 210 { 211 struct sk_buff *skb; 212 int err; 213 struct mISDNhead *hh; 214 215 skb = mI_alloc_skb(len, GFP_ATOMIC); 216 if (!skb) 217 return -ENOMEM; 218 hh = mISDN_HEAD_P(skb); 219 hh->prim = prim; 220 hh->id = id; 221 if (len) 222 memcpy(skb_put(skb, len), arg, len); 223 err = l2down_raw(l2, skb); 224 if (err) 225 dev_kfree_skb(skb); 226 return err; 227 } 228 229 static int 230 ph_data_confirm(struct layer2 *l2, struct mISDNhead *hh, struct sk_buff *skb) { 231 struct sk_buff *nskb = skb; 232 int ret = -EAGAIN; 233 234 if (test_bit(FLG_L1_NOTREADY, &l2->flag)) { 235 if (hh->id == l2->down_id) { 236 nskb = skb_dequeue(&l2->down_queue); 237 if (nskb) { 238 l2->down_id = mISDN_HEAD_ID(nskb); 239 if (l2down_skb(l2, nskb)) { 240 dev_kfree_skb(nskb); 241 l2->down_id = MISDN_ID_NONE; 242 } 243 } else 244 l2->down_id = MISDN_ID_NONE; 245 if (ret) { 246 dev_kfree_skb(skb); 247 ret = 0; 248 } 249 if (l2->down_id == MISDN_ID_NONE) { 250 test_and_clear_bit(FLG_L1_NOTREADY, &l2->flag); 251 mISDN_FsmEvent(&l2->l2m, EV_L2_ACK_PULL, NULL); 252 } 253 } 254 } 255 if (!test_and_set_bit(FLG_L1_NOTREADY, &l2->flag)) { 256 nskb = skb_dequeue(&l2->down_queue); 257 if (nskb) { 258 l2->down_id = mISDN_HEAD_ID(nskb); 259 if (l2down_skb(l2, nskb)) { 260 dev_kfree_skb(nskb); 261 l2->down_id = MISDN_ID_NONE; 262 test_and_clear_bit(FLG_L1_NOTREADY, &l2->flag); 263 } 264 } else 265 test_and_clear_bit(FLG_L1_NOTREADY, &l2->flag); 266 } 267 return ret; 268 } 269 270 static int 271 l2mgr(struct layer2 *l2, u_int prim, void *arg) { 272 long c = (long)arg; 273 274 printk(KERN_WARNING 275 "l2mgr: addr:%x prim %x %c\n", l2->id, prim, (char)c); 276 if (test_bit(FLG_LAPD, &l2->flag) && 277 !test_bit(FLG_FIXED_TEI, &l2->flag)) { 278 switch (c) { 279 case 'C': 280 case 'D': 281 case 'G': 282 case 'H': 283 l2_tei(l2, prim, (u_long)arg); 284 break; 285 } 286 } 287 return 0; 288 } 289 290 static void 291 set_peer_busy(struct layer2 *l2) { 292 test_and_set_bit(FLG_PEER_BUSY, &l2->flag); 293 if (skb_queue_len(&l2->i_queue) || skb_queue_len(&l2->ui_queue)) 294 test_and_set_bit(FLG_L2BLOCK, &l2->flag); 295 } 296 297 static void 298 clear_peer_busy(struct layer2 *l2) { 299 if (test_and_clear_bit(FLG_PEER_BUSY, &l2->flag)) 300 test_and_clear_bit(FLG_L2BLOCK, &l2->flag); 301 } 302 303 static void 304 InitWin(struct layer2 *l2) 305 { 306 int i; 307 308 for (i = 0; i < MAX_WINDOW; i++) 309 l2->windowar[i] = NULL; 310 } 311 312 static int 313 freewin(struct layer2 *l2) 314 { 315 int i, cnt = 0; 316 317 for (i = 0; i < MAX_WINDOW; i++) { 318 if (l2->windowar[i]) { 319 cnt++; 320 dev_kfree_skb(l2->windowar[i]); 321 l2->windowar[i] = NULL; 322 } 323 } 324 return cnt; 325 } 326 327 static void 328 ReleaseWin(struct layer2 *l2) 329 { 330 int cnt = freewin(l2); 331 332 if (cnt) 333 printk(KERN_WARNING 334 "isdnl2 freed %d skbuffs in release\n", cnt); 335 } 336 337 inline unsigned int 338 cansend(struct layer2 *l2) 339 { 340 unsigned int p1; 341 342 if (test_bit(FLG_MOD128, &l2->flag)) 343 p1 = (l2->vs - l2->va) % 128; 344 else 345 p1 = (l2->vs - l2->va) % 8; 346 return (p1 < l2->window) && !test_bit(FLG_PEER_BUSY, &l2->flag); 347 } 348 349 inline void 350 clear_exception(struct layer2 *l2) 351 { 352 test_and_clear_bit(FLG_ACK_PEND, &l2->flag); 353 test_and_clear_bit(FLG_REJEXC, &l2->flag); 354 test_and_clear_bit(FLG_OWN_BUSY, &l2->flag); 355 clear_peer_busy(l2); 356 } 357 358 static int 359 sethdraddr(struct layer2 *l2, u_char *header, int rsp) 360 { 361 u_char *ptr = header; 362 int crbit = rsp; 363 364 if (test_bit(FLG_LAPD, &l2->flag)) { 365 if (test_bit(FLG_LAPD_NET, &l2->flag)) 366 crbit = !crbit; 367 *ptr++ = (l2->sapi << 2) | (crbit ? 2 : 0); 368 *ptr++ = (l2->tei << 1) | 1; 369 return 2; 370 } else { 371 if (test_bit(FLG_ORIG, &l2->flag)) 372 crbit = !crbit; 373 if (crbit) 374 *ptr++ = l2->addr.B; 375 else 376 *ptr++ = l2->addr.A; 377 return 1; 378 } 379 } 380 381 static inline void 382 enqueue_super(struct layer2 *l2, struct sk_buff *skb) 383 { 384 if (l2down(l2, PH_DATA_REQ, l2_newid(l2), skb)) 385 dev_kfree_skb(skb); 386 } 387 388 static inline void 389 enqueue_ui(struct layer2 *l2, struct sk_buff *skb) 390 { 391 if (l2->tm) 392 l2_tei(l2, MDL_STATUS_UI_IND, 0); 393 if (l2down(l2, PH_DATA_REQ, l2_newid(l2), skb)) 394 dev_kfree_skb(skb); 395 } 396 397 inline int 398 IsUI(u_char *data) 399 { 400 return (data[0] & 0xef) == UI; 401 } 402 403 inline int 404 IsUA(u_char *data) 405 { 406 return (data[0] & 0xef) == UA; 407 } 408 409 inline int 410 IsDM(u_char *data) 411 { 412 return (data[0] & 0xef) == DM; 413 } 414 415 inline int 416 IsDISC(u_char *data) 417 { 418 return (data[0] & 0xef) == DISC; 419 } 420 421 inline int 422 IsRR(u_char *data, struct layer2 *l2) 423 { 424 if (test_bit(FLG_MOD128, &l2->flag)) 425 return data[0] == RR; 426 else 427 return (data[0] & 0xf) == 1; 428 } 429 430 inline int 431 IsSFrame(u_char *data, struct layer2 *l2) 432 { 433 register u_char d = *data; 434 435 if (!test_bit(FLG_MOD128, &l2->flag)) 436 d &= 0xf; 437 return ((d & 0xf3) == 1) && ((d & 0x0c) != 0x0c); 438 } 439 440 inline int 441 IsSABME(u_char *data, struct layer2 *l2) 442 { 443 u_char d = data[0] & ~0x10; 444 445 return test_bit(FLG_MOD128, &l2->flag) ? d == SABME : d == SABM; 446 } 447 448 inline int 449 IsREJ(u_char *data, struct layer2 *l2) 450 { 451 return test_bit(FLG_MOD128, &l2->flag) ? 452 data[0] == REJ : (data[0] & 0xf) == REJ; 453 } 454 455 inline int 456 IsFRMR(u_char *data) 457 { 458 return (data[0] & 0xef) == FRMR; 459 } 460 461 inline int 462 IsRNR(u_char *data, struct layer2 *l2) 463 { 464 return test_bit(FLG_MOD128, &l2->flag) ? 465 data[0] == RNR : (data[0] & 0xf) == RNR; 466 } 467 468 int 469 iframe_error(struct layer2 *l2, struct sk_buff *skb) 470 { 471 u_int i; 472 int rsp = *skb->data & 0x2; 473 474 i = l2addrsize(l2) + (test_bit(FLG_MOD128, &l2->flag) ? 2 : 1); 475 if (test_bit(FLG_ORIG, &l2->flag)) 476 rsp = !rsp; 477 if (rsp) 478 return 'L'; 479 if (skb->len < i) 480 return 'N'; 481 if ((skb->len - i) > l2->maxlen) 482 return 'O'; 483 return 0; 484 } 485 486 int 487 super_error(struct layer2 *l2, struct sk_buff *skb) 488 { 489 if (skb->len != l2addrsize(l2) + 490 (test_bit(FLG_MOD128, &l2->flag) ? 2 : 1)) 491 return 'N'; 492 return 0; 493 } 494 495 int 496 unnum_error(struct layer2 *l2, struct sk_buff *skb, int wantrsp) 497 { 498 int rsp = (*skb->data & 0x2) >> 1; 499 if (test_bit(FLG_ORIG, &l2->flag)) 500 rsp = !rsp; 501 if (rsp != wantrsp) 502 return 'L'; 503 if (skb->len != l2addrsize(l2) + 1) 504 return 'N'; 505 return 0; 506 } 507 508 int 509 UI_error(struct layer2 *l2, struct sk_buff *skb) 510 { 511 int rsp = *skb->data & 0x2; 512 if (test_bit(FLG_ORIG, &l2->flag)) 513 rsp = !rsp; 514 if (rsp) 515 return 'L'; 516 if (skb->len > l2->maxlen + l2addrsize(l2) + 1) 517 return 'O'; 518 return 0; 519 } 520 521 int 522 FRMR_error(struct layer2 *l2, struct sk_buff *skb) 523 { 524 u_int headers = l2addrsize(l2) + 1; 525 u_char *datap = skb->data + headers; 526 int rsp = *skb->data & 0x2; 527 528 if (test_bit(FLG_ORIG, &l2->flag)) 529 rsp = !rsp; 530 if (!rsp) 531 return 'L'; 532 if (test_bit(FLG_MOD128, &l2->flag)) { 533 if (skb->len < headers + 5) 534 return 'N'; 535 else if (*debug & DEBUG_L2) 536 l2m_debug(&l2->l2m, 537 "FRMR information %2x %2x %2x %2x %2x", 538 datap[0], datap[1], datap[2], datap[3], datap[4]); 539 } else { 540 if (skb->len < headers + 3) 541 return 'N'; 542 else if (*debug & DEBUG_L2) 543 l2m_debug(&l2->l2m, 544 "FRMR information %2x %2x %2x", 545 datap[0], datap[1], datap[2]); 546 } 547 return 0; 548 } 549 550 static unsigned int 551 legalnr(struct layer2 *l2, unsigned int nr) 552 { 553 if (test_bit(FLG_MOD128, &l2->flag)) 554 return ((nr - l2->va) % 128) <= ((l2->vs - l2->va) % 128); 555 else 556 return ((nr - l2->va) % 8) <= ((l2->vs - l2->va) % 8); 557 } 558 559 static void 560 setva(struct layer2 *l2, unsigned int nr) 561 { 562 struct sk_buff *skb; 563 564 while (l2->va != nr) { 565 l2->va++; 566 if (test_bit(FLG_MOD128, &l2->flag)) 567 l2->va %= 128; 568 else 569 l2->va %= 8; 570 if (l2->windowar[l2->sow]) { 571 skb_trim(l2->windowar[l2->sow], 0); 572 skb_queue_tail(&l2->tmp_queue, l2->windowar[l2->sow]); 573 l2->windowar[l2->sow] = NULL; 574 } 575 l2->sow = (l2->sow + 1) % l2->window; 576 } 577 skb = skb_dequeue(&l2->tmp_queue); 578 while (skb) { 579 dev_kfree_skb(skb); 580 skb = skb_dequeue(&l2->tmp_queue); 581 } 582 } 583 584 static void 585 send_uframe(struct layer2 *l2, struct sk_buff *skb, u_char cmd, u_char cr) 586 { 587 u_char tmp[MAX_L2HEADER_LEN]; 588 int i; 589 590 i = sethdraddr(l2, tmp, cr); 591 tmp[i++] = cmd; 592 if (skb) 593 skb_trim(skb, 0); 594 else { 595 skb = mI_alloc_skb(i, GFP_ATOMIC); 596 if (!skb) { 597 printk(KERN_WARNING "%s: can't alloc skbuff\n", 598 __func__); 599 return; 600 } 601 } 602 memcpy(skb_put(skb, i), tmp, i); 603 enqueue_super(l2, skb); 604 } 605 606 607 inline u_char 608 get_PollFlag(struct layer2 *l2, struct sk_buff *skb) 609 { 610 return skb->data[l2addrsize(l2)] & 0x10; 611 } 612 613 inline u_char 614 get_PollFlagFree(struct layer2 *l2, struct sk_buff *skb) 615 { 616 u_char PF; 617 618 PF = get_PollFlag(l2, skb); 619 dev_kfree_skb(skb); 620 return PF; 621 } 622 623 inline void 624 start_t200(struct layer2 *l2, int i) 625 { 626 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, i); 627 test_and_set_bit(FLG_T200_RUN, &l2->flag); 628 } 629 630 inline void 631 restart_t200(struct layer2 *l2, int i) 632 { 633 mISDN_FsmRestartTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, i); 634 test_and_set_bit(FLG_T200_RUN, &l2->flag); 635 } 636 637 inline void 638 stop_t200(struct layer2 *l2, int i) 639 { 640 if (test_and_clear_bit(FLG_T200_RUN, &l2->flag)) 641 mISDN_FsmDelTimer(&l2->t200, i); 642 } 643 644 inline void 645 st5_dl_release_l2l3(struct layer2 *l2) 646 { 647 int pr; 648 649 if (test_and_clear_bit(FLG_PEND_REL, &l2->flag)) 650 pr = DL_RELEASE_CNF; 651 else 652 pr = DL_RELEASE_IND; 653 l2up_create(l2, pr, 0, NULL); 654 } 655 656 inline void 657 lapb_dl_release_l2l3(struct layer2 *l2, int f) 658 { 659 if (test_bit(FLG_LAPB, &l2->flag)) 660 l2down_create(l2, PH_DEACTIVATE_REQ, l2_newid(l2), 0, NULL); 661 l2up_create(l2, f, 0, NULL); 662 } 663 664 static void 665 establishlink(struct FsmInst *fi) 666 { 667 struct layer2 *l2 = fi->userdata; 668 u_char cmd; 669 670 clear_exception(l2); 671 l2->rc = 0; 672 cmd = (test_bit(FLG_MOD128, &l2->flag) ? SABME : SABM) | 0x10; 673 send_uframe(l2, NULL, cmd, CMD); 674 mISDN_FsmDelTimer(&l2->t203, 1); 675 restart_t200(l2, 1); 676 test_and_clear_bit(FLG_PEND_REL, &l2->flag); 677 freewin(l2); 678 mISDN_FsmChangeState(fi, ST_L2_5); 679 } 680 681 static void 682 l2_mdl_error_ua(struct FsmInst *fi, int event, void *arg) 683 { 684 struct sk_buff *skb = arg; 685 struct layer2 *l2 = fi->userdata; 686 687 if (get_PollFlagFree(l2, skb)) 688 l2mgr(l2, MDL_ERROR_IND, (void *) 'C'); 689 else 690 l2mgr(l2, MDL_ERROR_IND, (void *) 'D'); 691 692 } 693 694 static void 695 l2_mdl_error_dm(struct FsmInst *fi, int event, void *arg) 696 { 697 struct sk_buff *skb = arg; 698 struct layer2 *l2 = fi->userdata; 699 700 if (get_PollFlagFree(l2, skb)) 701 l2mgr(l2, MDL_ERROR_IND, (void *) 'B'); 702 else { 703 l2mgr(l2, MDL_ERROR_IND, (void *) 'E'); 704 establishlink(fi); 705 test_and_clear_bit(FLG_L3_INIT, &l2->flag); 706 } 707 } 708 709 static void 710 l2_st8_mdl_error_dm(struct FsmInst *fi, int event, void *arg) 711 { 712 struct sk_buff *skb = arg; 713 struct layer2 *l2 = fi->userdata; 714 715 if (get_PollFlagFree(l2, skb)) 716 l2mgr(l2, MDL_ERROR_IND, (void *) 'B'); 717 else 718 l2mgr(l2, MDL_ERROR_IND, (void *) 'E'); 719 establishlink(fi); 720 test_and_clear_bit(FLG_L3_INIT, &l2->flag); 721 } 722 723 static void 724 l2_go_st3(struct FsmInst *fi, int event, void *arg) 725 { 726 dev_kfree_skb((struct sk_buff *)arg); 727 mISDN_FsmChangeState(fi, ST_L2_3); 728 } 729 730 static void 731 l2_mdl_assign(struct FsmInst *fi, int event, void *arg) 732 { 733 struct layer2 *l2 = fi->userdata; 734 735 mISDN_FsmChangeState(fi, ST_L2_3); 736 dev_kfree_skb((struct sk_buff *)arg); 737 l2_tei(l2, MDL_ASSIGN_IND, 0); 738 } 739 740 static void 741 l2_queue_ui_assign(struct FsmInst *fi, int event, void *arg) 742 { 743 struct layer2 *l2 = fi->userdata; 744 struct sk_buff *skb = arg; 745 746 skb_queue_tail(&l2->ui_queue, skb); 747 mISDN_FsmChangeState(fi, ST_L2_2); 748 l2_tei(l2, MDL_ASSIGN_IND, 0); 749 } 750 751 static void 752 l2_queue_ui(struct FsmInst *fi, int event, void *arg) 753 { 754 struct layer2 *l2 = fi->userdata; 755 struct sk_buff *skb = arg; 756 757 skb_queue_tail(&l2->ui_queue, skb); 758 } 759 760 static void 761 tx_ui(struct layer2 *l2) 762 { 763 struct sk_buff *skb; 764 u_char header[MAX_L2HEADER_LEN]; 765 int i; 766 767 i = sethdraddr(l2, header, CMD); 768 if (test_bit(FLG_LAPD_NET, &l2->flag)) 769 header[1] = 0xff; /* tei 127 */ 770 header[i++] = UI; 771 while ((skb = skb_dequeue(&l2->ui_queue))) { 772 memcpy(skb_push(skb, i), header, i); 773 enqueue_ui(l2, skb); 774 } 775 } 776 777 static void 778 l2_send_ui(struct FsmInst *fi, int event, void *arg) 779 { 780 struct layer2 *l2 = fi->userdata; 781 struct sk_buff *skb = arg; 782 783 skb_queue_tail(&l2->ui_queue, skb); 784 tx_ui(l2); 785 } 786 787 static void 788 l2_got_ui(struct FsmInst *fi, int event, void *arg) 789 { 790 struct layer2 *l2 = fi->userdata; 791 struct sk_buff *skb = arg; 792 793 skb_pull(skb, l2headersize(l2, 1)); 794 /* 795 * in states 1-3 for broadcast 796 */ 797 798 if (l2->tm) 799 l2_tei(l2, MDL_STATUS_UI_IND, 0); 800 l2up(l2, DL_UNITDATA_IND, skb); 801 } 802 803 static void 804 l2_establish(struct FsmInst *fi, int event, void *arg) 805 { 806 struct sk_buff *skb = arg; 807 struct layer2 *l2 = fi->userdata; 808 809 establishlink(fi); 810 test_and_set_bit(FLG_L3_INIT, &l2->flag); 811 dev_kfree_skb(skb); 812 } 813 814 static void 815 l2_discard_i_setl3(struct FsmInst *fi, int event, void *arg) 816 { 817 struct sk_buff *skb = arg; 818 struct layer2 *l2 = fi->userdata; 819 820 skb_queue_purge(&l2->i_queue); 821 test_and_set_bit(FLG_L3_INIT, &l2->flag); 822 test_and_clear_bit(FLG_PEND_REL, &l2->flag); 823 dev_kfree_skb(skb); 824 } 825 826 static void 827 l2_l3_reestablish(struct FsmInst *fi, int event, void *arg) 828 { 829 struct sk_buff *skb = arg; 830 struct layer2 *l2 = fi->userdata; 831 832 skb_queue_purge(&l2->i_queue); 833 establishlink(fi); 834 test_and_set_bit(FLG_L3_INIT, &l2->flag); 835 dev_kfree_skb(skb); 836 } 837 838 static void 839 l2_release(struct FsmInst *fi, int event, void *arg) 840 { 841 struct layer2 *l2 = fi->userdata; 842 struct sk_buff *skb = arg; 843 844 skb_trim(skb, 0); 845 l2up(l2, DL_RELEASE_CNF, skb); 846 } 847 848 static void 849 l2_pend_rel(struct FsmInst *fi, int event, void *arg) 850 { 851 struct sk_buff *skb = arg; 852 struct layer2 *l2 = fi->userdata; 853 854 test_and_set_bit(FLG_PEND_REL, &l2->flag); 855 dev_kfree_skb(skb); 856 } 857 858 static void 859 l2_disconnect(struct FsmInst *fi, int event, void *arg) 860 { 861 struct layer2 *l2 = fi->userdata; 862 struct sk_buff *skb = arg; 863 864 skb_queue_purge(&l2->i_queue); 865 freewin(l2); 866 mISDN_FsmChangeState(fi, ST_L2_6); 867 l2->rc = 0; 868 send_uframe(l2, NULL, DISC | 0x10, CMD); 869 mISDN_FsmDelTimer(&l2->t203, 1); 870 restart_t200(l2, 2); 871 if (skb) 872 dev_kfree_skb(skb); 873 } 874 875 static void 876 l2_start_multi(struct FsmInst *fi, int event, void *arg) 877 { 878 struct layer2 *l2 = fi->userdata; 879 struct sk_buff *skb = arg; 880 881 l2->vs = 0; 882 l2->va = 0; 883 l2->vr = 0; 884 l2->sow = 0; 885 clear_exception(l2); 886 send_uframe(l2, NULL, UA | get_PollFlag(l2, skb), RSP); 887 mISDN_FsmChangeState(fi, ST_L2_7); 888 mISDN_FsmAddTimer(&l2->t203, l2->T203, EV_L2_T203, NULL, 3); 889 skb_trim(skb, 0); 890 l2up(l2, DL_ESTABLISH_IND, skb); 891 if (l2->tm) 892 l2_tei(l2, MDL_STATUS_UP_IND, 0); 893 } 894 895 static void 896 l2_send_UA(struct FsmInst *fi, int event, void *arg) 897 { 898 struct layer2 *l2 = fi->userdata; 899 struct sk_buff *skb = arg; 900 901 send_uframe(l2, skb, UA | get_PollFlag(l2, skb), RSP); 902 } 903 904 static void 905 l2_send_DM(struct FsmInst *fi, int event, void *arg) 906 { 907 struct layer2 *l2 = fi->userdata; 908 struct sk_buff *skb = arg; 909 910 send_uframe(l2, skb, DM | get_PollFlag(l2, skb), RSP); 911 } 912 913 static void 914 l2_restart_multi(struct FsmInst *fi, int event, void *arg) 915 { 916 struct layer2 *l2 = fi->userdata; 917 struct sk_buff *skb = arg; 918 int est = 0; 919 920 send_uframe(l2, skb, UA | get_PollFlag(l2, skb), RSP); 921 922 l2mgr(l2, MDL_ERROR_IND, (void *) 'F'); 923 924 if (l2->vs != l2->va) { 925 skb_queue_purge(&l2->i_queue); 926 est = 1; 927 } 928 929 clear_exception(l2); 930 l2->vs = 0; 931 l2->va = 0; 932 l2->vr = 0; 933 l2->sow = 0; 934 mISDN_FsmChangeState(fi, ST_L2_7); 935 stop_t200(l2, 3); 936 mISDN_FsmRestartTimer(&l2->t203, l2->T203, EV_L2_T203, NULL, 3); 937 938 if (est) 939 l2up_create(l2, DL_ESTABLISH_IND, 0, NULL); 940 /* mISDN_queue_data(&l2->inst, l2->inst.id | MSG_BROADCAST, 941 * MGR_SHORTSTATUS | INDICATION, SSTATUS_L2_ESTABLISHED, 942 * 0, NULL, 0); 943 */ 944 if (skb_queue_len(&l2->i_queue) && cansend(l2)) 945 mISDN_FsmEvent(fi, EV_L2_ACK_PULL, NULL); 946 } 947 948 static void 949 l2_stop_multi(struct FsmInst *fi, int event, void *arg) 950 { 951 struct layer2 *l2 = fi->userdata; 952 struct sk_buff *skb = arg; 953 954 mISDN_FsmChangeState(fi, ST_L2_4); 955 mISDN_FsmDelTimer(&l2->t203, 3); 956 stop_t200(l2, 4); 957 958 send_uframe(l2, skb, UA | get_PollFlag(l2, skb), RSP); 959 skb_queue_purge(&l2->i_queue); 960 freewin(l2); 961 lapb_dl_release_l2l3(l2, DL_RELEASE_IND); 962 if (l2->tm) 963 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 964 } 965 966 static void 967 l2_connected(struct FsmInst *fi, int event, void *arg) 968 { 969 struct layer2 *l2 = fi->userdata; 970 struct sk_buff *skb = arg; 971 int pr = -1; 972 973 if (!get_PollFlag(l2, skb)) { 974 l2_mdl_error_ua(fi, event, arg); 975 return; 976 } 977 dev_kfree_skb(skb); 978 if (test_and_clear_bit(FLG_PEND_REL, &l2->flag)) 979 l2_disconnect(fi, event, NULL); 980 if (test_and_clear_bit(FLG_L3_INIT, &l2->flag)) { 981 pr = DL_ESTABLISH_CNF; 982 } else if (l2->vs != l2->va) { 983 skb_queue_purge(&l2->i_queue); 984 pr = DL_ESTABLISH_IND; 985 } 986 stop_t200(l2, 5); 987 l2->vr = 0; 988 l2->vs = 0; 989 l2->va = 0; 990 l2->sow = 0; 991 mISDN_FsmChangeState(fi, ST_L2_7); 992 mISDN_FsmAddTimer(&l2->t203, l2->T203, EV_L2_T203, NULL, 4); 993 if (pr != -1) 994 l2up_create(l2, pr, 0, NULL); 995 996 if (skb_queue_len(&l2->i_queue) && cansend(l2)) 997 mISDN_FsmEvent(fi, EV_L2_ACK_PULL, NULL); 998 999 if (l2->tm) 1000 l2_tei(l2, MDL_STATUS_UP_IND, 0); 1001 } 1002 1003 static void 1004 l2_released(struct FsmInst *fi, int event, void *arg) 1005 { 1006 struct layer2 *l2 = fi->userdata; 1007 struct sk_buff *skb = arg; 1008 1009 if (!get_PollFlag(l2, skb)) { 1010 l2_mdl_error_ua(fi, event, arg); 1011 return; 1012 } 1013 dev_kfree_skb(skb); 1014 stop_t200(l2, 6); 1015 lapb_dl_release_l2l3(l2, DL_RELEASE_CNF); 1016 mISDN_FsmChangeState(fi, ST_L2_4); 1017 if (l2->tm) 1018 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1019 } 1020 1021 static void 1022 l2_reestablish(struct FsmInst *fi, int event, void *arg) 1023 { 1024 struct layer2 *l2 = fi->userdata; 1025 struct sk_buff *skb = arg; 1026 1027 if (!get_PollFlagFree(l2, skb)) { 1028 establishlink(fi); 1029 test_and_set_bit(FLG_L3_INIT, &l2->flag); 1030 } 1031 } 1032 1033 static void 1034 l2_st5_dm_release(struct FsmInst *fi, int event, void *arg) 1035 { 1036 struct layer2 *l2 = fi->userdata; 1037 struct sk_buff *skb = arg; 1038 1039 if (get_PollFlagFree(l2, skb)) { 1040 stop_t200(l2, 7); 1041 if (!test_bit(FLG_L3_INIT, &l2->flag)) 1042 skb_queue_purge(&l2->i_queue); 1043 if (test_bit(FLG_LAPB, &l2->flag)) 1044 l2down_create(l2, PH_DEACTIVATE_REQ, 1045 l2_newid(l2), 0, NULL); 1046 st5_dl_release_l2l3(l2); 1047 mISDN_FsmChangeState(fi, ST_L2_4); 1048 if (l2->tm) 1049 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1050 } 1051 } 1052 1053 static void 1054 l2_st6_dm_release(struct FsmInst *fi, int event, void *arg) 1055 { 1056 struct layer2 *l2 = fi->userdata; 1057 struct sk_buff *skb = arg; 1058 1059 if (get_PollFlagFree(l2, skb)) { 1060 stop_t200(l2, 8); 1061 lapb_dl_release_l2l3(l2, DL_RELEASE_CNF); 1062 mISDN_FsmChangeState(fi, ST_L2_4); 1063 if (l2->tm) 1064 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1065 } 1066 } 1067 1068 void 1069 enquiry_cr(struct layer2 *l2, u_char typ, u_char cr, u_char pf) 1070 { 1071 struct sk_buff *skb; 1072 u_char tmp[MAX_L2HEADER_LEN]; 1073 int i; 1074 1075 i = sethdraddr(l2, tmp, cr); 1076 if (test_bit(FLG_MOD128, &l2->flag)) { 1077 tmp[i++] = typ; 1078 tmp[i++] = (l2->vr << 1) | (pf ? 1 : 0); 1079 } else 1080 tmp[i++] = (l2->vr << 5) | typ | (pf ? 0x10 : 0); 1081 skb = mI_alloc_skb(i, GFP_ATOMIC); 1082 if (!skb) { 1083 printk(KERN_WARNING 1084 "isdnl2 can't alloc sbbuff for enquiry_cr\n"); 1085 return; 1086 } 1087 memcpy(skb_put(skb, i), tmp, i); 1088 enqueue_super(l2, skb); 1089 } 1090 1091 inline void 1092 enquiry_response(struct layer2 *l2) 1093 { 1094 if (test_bit(FLG_OWN_BUSY, &l2->flag)) 1095 enquiry_cr(l2, RNR, RSP, 1); 1096 else 1097 enquiry_cr(l2, RR, RSP, 1); 1098 test_and_clear_bit(FLG_ACK_PEND, &l2->flag); 1099 } 1100 1101 inline void 1102 transmit_enquiry(struct layer2 *l2) 1103 { 1104 if (test_bit(FLG_OWN_BUSY, &l2->flag)) 1105 enquiry_cr(l2, RNR, CMD, 1); 1106 else 1107 enquiry_cr(l2, RR, CMD, 1); 1108 test_and_clear_bit(FLG_ACK_PEND, &l2->flag); 1109 start_t200(l2, 9); 1110 } 1111 1112 1113 static void 1114 nrerrorrecovery(struct FsmInst *fi) 1115 { 1116 struct layer2 *l2 = fi->userdata; 1117 1118 l2mgr(l2, MDL_ERROR_IND, (void *) 'J'); 1119 establishlink(fi); 1120 test_and_clear_bit(FLG_L3_INIT, &l2->flag); 1121 } 1122 1123 static void 1124 invoke_retransmission(struct layer2 *l2, unsigned int nr) 1125 { 1126 u_int p1; 1127 1128 if (l2->vs != nr) { 1129 while (l2->vs != nr) { 1130 (l2->vs)--; 1131 if (test_bit(FLG_MOD128, &l2->flag)) { 1132 l2->vs %= 128; 1133 p1 = (l2->vs - l2->va) % 128; 1134 } else { 1135 l2->vs %= 8; 1136 p1 = (l2->vs - l2->va) % 8; 1137 } 1138 p1 = (p1 + l2->sow) % l2->window; 1139 if (l2->windowar[p1]) 1140 skb_queue_head(&l2->i_queue, l2->windowar[p1]); 1141 else 1142 printk(KERN_WARNING 1143 "%s: windowar[%d] is NULL\n", 1144 __func__, p1); 1145 l2->windowar[p1] = NULL; 1146 } 1147 mISDN_FsmEvent(&l2->l2m, EV_L2_ACK_PULL, NULL); 1148 } 1149 } 1150 1151 static void 1152 l2_st7_got_super(struct FsmInst *fi, int event, void *arg) 1153 { 1154 struct layer2 *l2 = fi->userdata; 1155 struct sk_buff *skb = arg; 1156 int PollFlag, rsp, typ = RR; 1157 unsigned int nr; 1158 1159 rsp = *skb->data & 0x2; 1160 if (test_bit(FLG_ORIG, &l2->flag)) 1161 rsp = !rsp; 1162 1163 skb_pull(skb, l2addrsize(l2)); 1164 if (IsRNR(skb->data, l2)) { 1165 set_peer_busy(l2); 1166 typ = RNR; 1167 } else 1168 clear_peer_busy(l2); 1169 if (IsREJ(skb->data, l2)) 1170 typ = REJ; 1171 1172 if (test_bit(FLG_MOD128, &l2->flag)) { 1173 PollFlag = (skb->data[1] & 0x1) == 0x1; 1174 nr = skb->data[1] >> 1; 1175 } else { 1176 PollFlag = (skb->data[0] & 0x10); 1177 nr = (skb->data[0] >> 5) & 0x7; 1178 } 1179 dev_kfree_skb(skb); 1180 1181 if (PollFlag) { 1182 if (rsp) 1183 l2mgr(l2, MDL_ERROR_IND, (void *) 'A'); 1184 else 1185 enquiry_response(l2); 1186 } 1187 if (legalnr(l2, nr)) { 1188 if (typ == REJ) { 1189 setva(l2, nr); 1190 invoke_retransmission(l2, nr); 1191 stop_t200(l2, 10); 1192 if (mISDN_FsmAddTimer(&l2->t203, l2->T203, 1193 EV_L2_T203, NULL, 6)) 1194 l2m_debug(&l2->l2m, "Restart T203 ST7 REJ"); 1195 } else if ((nr == l2->vs) && (typ == RR)) { 1196 setva(l2, nr); 1197 stop_t200(l2, 11); 1198 mISDN_FsmRestartTimer(&l2->t203, l2->T203, 1199 EV_L2_T203, NULL, 7); 1200 } else if ((l2->va != nr) || (typ == RNR)) { 1201 setva(l2, nr); 1202 if (typ != RR) 1203 mISDN_FsmDelTimer(&l2->t203, 9); 1204 restart_t200(l2, 12); 1205 } 1206 if (skb_queue_len(&l2->i_queue) && (typ == RR)) 1207 mISDN_FsmEvent(fi, EV_L2_ACK_PULL, NULL); 1208 } else 1209 nrerrorrecovery(fi); 1210 } 1211 1212 static void 1213 l2_feed_i_if_reest(struct FsmInst *fi, int event, void *arg) 1214 { 1215 struct layer2 *l2 = fi->userdata; 1216 struct sk_buff *skb = arg; 1217 1218 if (!test_bit(FLG_L3_INIT, &l2->flag)) 1219 skb_queue_tail(&l2->i_queue, skb); 1220 else 1221 dev_kfree_skb(skb); 1222 } 1223 1224 static void 1225 l2_feed_i_pull(struct FsmInst *fi, int event, void *arg) 1226 { 1227 struct layer2 *l2 = fi->userdata; 1228 struct sk_buff *skb = arg; 1229 1230 skb_queue_tail(&l2->i_queue, skb); 1231 mISDN_FsmEvent(fi, EV_L2_ACK_PULL, NULL); 1232 } 1233 1234 static void 1235 l2_feed_iqueue(struct FsmInst *fi, int event, void *arg) 1236 { 1237 struct layer2 *l2 = fi->userdata; 1238 struct sk_buff *skb = arg; 1239 1240 skb_queue_tail(&l2->i_queue, skb); 1241 } 1242 1243 static void 1244 l2_got_iframe(struct FsmInst *fi, int event, void *arg) 1245 { 1246 struct layer2 *l2 = fi->userdata; 1247 struct sk_buff *skb = arg; 1248 int PollFlag, i; 1249 u_int ns, nr; 1250 1251 i = l2addrsize(l2); 1252 if (test_bit(FLG_MOD128, &l2->flag)) { 1253 PollFlag = ((skb->data[i + 1] & 0x1) == 0x1); 1254 ns = skb->data[i] >> 1; 1255 nr = (skb->data[i + 1] >> 1) & 0x7f; 1256 } else { 1257 PollFlag = (skb->data[i] & 0x10); 1258 ns = (skb->data[i] >> 1) & 0x7; 1259 nr = (skb->data[i] >> 5) & 0x7; 1260 } 1261 if (test_bit(FLG_OWN_BUSY, &l2->flag)) { 1262 dev_kfree_skb(skb); 1263 if (PollFlag) 1264 enquiry_response(l2); 1265 } else { 1266 if (l2->vr == ns) { 1267 l2->vr++; 1268 if (test_bit(FLG_MOD128, &l2->flag)) 1269 l2->vr %= 128; 1270 else 1271 l2->vr %= 8; 1272 test_and_clear_bit(FLG_REJEXC, &l2->flag); 1273 if (PollFlag) 1274 enquiry_response(l2); 1275 else 1276 test_and_set_bit(FLG_ACK_PEND, &l2->flag); 1277 skb_pull(skb, l2headersize(l2, 0)); 1278 l2up(l2, DL_DATA_IND, skb); 1279 } else { 1280 /* n(s)!=v(r) */ 1281 dev_kfree_skb(skb); 1282 if (test_and_set_bit(FLG_REJEXC, &l2->flag)) { 1283 if (PollFlag) 1284 enquiry_response(l2); 1285 } else { 1286 enquiry_cr(l2, REJ, RSP, PollFlag); 1287 test_and_clear_bit(FLG_ACK_PEND, &l2->flag); 1288 } 1289 } 1290 } 1291 if (legalnr(l2, nr)) { 1292 if (!test_bit(FLG_PEER_BUSY, &l2->flag) && 1293 (fi->state == ST_L2_7)) { 1294 if (nr == l2->vs) { 1295 stop_t200(l2, 13); 1296 mISDN_FsmRestartTimer(&l2->t203, l2->T203, 1297 EV_L2_T203, NULL, 7); 1298 } else if (nr != l2->va) 1299 restart_t200(l2, 14); 1300 } 1301 setva(l2, nr); 1302 } else { 1303 nrerrorrecovery(fi); 1304 return; 1305 } 1306 if (skb_queue_len(&l2->i_queue) && (fi->state == ST_L2_7)) 1307 mISDN_FsmEvent(fi, EV_L2_ACK_PULL, NULL); 1308 if (test_and_clear_bit(FLG_ACK_PEND, &l2->flag)) 1309 enquiry_cr(l2, RR, RSP, 0); 1310 } 1311 1312 static void 1313 l2_got_tei(struct FsmInst *fi, int event, void *arg) 1314 { 1315 struct layer2 *l2 = fi->userdata; 1316 u_int info; 1317 1318 l2->tei = (signed char)(long)arg; 1319 set_channel_address(&l2->ch, l2->sapi, l2->tei); 1320 info = DL_INFO_L2_CONNECT; 1321 l2up_create(l2, DL_INFORMATION_IND, sizeof(info), &info); 1322 if (fi->state == ST_L2_3) { 1323 establishlink(fi); 1324 test_and_set_bit(FLG_L3_INIT, &l2->flag); 1325 } else 1326 mISDN_FsmChangeState(fi, ST_L2_4); 1327 if (skb_queue_len(&l2->ui_queue)) 1328 tx_ui(l2); 1329 } 1330 1331 static void 1332 l2_st5_tout_200(struct FsmInst *fi, int event, void *arg) 1333 { 1334 struct layer2 *l2 = fi->userdata; 1335 1336 if (test_bit(FLG_LAPD, &l2->flag) && 1337 test_bit(FLG_DCHAN_BUSY, &l2->flag)) { 1338 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, 9); 1339 } else if (l2->rc == l2->N200) { 1340 mISDN_FsmChangeState(fi, ST_L2_4); 1341 test_and_clear_bit(FLG_T200_RUN, &l2->flag); 1342 skb_queue_purge(&l2->i_queue); 1343 l2mgr(l2, MDL_ERROR_IND, (void *) 'G'); 1344 if (test_bit(FLG_LAPB, &l2->flag)) 1345 l2down_create(l2, PH_DEACTIVATE_REQ, 1346 l2_newid(l2), 0, NULL); 1347 st5_dl_release_l2l3(l2); 1348 if (l2->tm) 1349 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1350 } else { 1351 l2->rc++; 1352 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, 9); 1353 send_uframe(l2, NULL, (test_bit(FLG_MOD128, &l2->flag) ? 1354 SABME : SABM) | 0x10, CMD); 1355 } 1356 } 1357 1358 static void 1359 l2_st6_tout_200(struct FsmInst *fi, int event, void *arg) 1360 { 1361 struct layer2 *l2 = fi->userdata; 1362 1363 if (test_bit(FLG_LAPD, &l2->flag) && 1364 test_bit(FLG_DCHAN_BUSY, &l2->flag)) { 1365 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, 9); 1366 } else if (l2->rc == l2->N200) { 1367 mISDN_FsmChangeState(fi, ST_L2_4); 1368 test_and_clear_bit(FLG_T200_RUN, &l2->flag); 1369 l2mgr(l2, MDL_ERROR_IND, (void *) 'H'); 1370 lapb_dl_release_l2l3(l2, DL_RELEASE_CNF); 1371 if (l2->tm) 1372 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1373 } else { 1374 l2->rc++; 1375 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, 1376 NULL, 9); 1377 send_uframe(l2, NULL, DISC | 0x10, CMD); 1378 } 1379 } 1380 1381 static void 1382 l2_st7_tout_200(struct FsmInst *fi, int event, void *arg) 1383 { 1384 struct layer2 *l2 = fi->userdata; 1385 1386 if (test_bit(FLG_LAPD, &l2->flag) && 1387 test_bit(FLG_DCHAN_BUSY, &l2->flag)) { 1388 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, 9); 1389 return; 1390 } 1391 test_and_clear_bit(FLG_T200_RUN, &l2->flag); 1392 l2->rc = 0; 1393 mISDN_FsmChangeState(fi, ST_L2_8); 1394 transmit_enquiry(l2); 1395 l2->rc++; 1396 } 1397 1398 static void 1399 l2_st8_tout_200(struct FsmInst *fi, int event, void *arg) 1400 { 1401 struct layer2 *l2 = fi->userdata; 1402 1403 if (test_bit(FLG_LAPD, &l2->flag) && 1404 test_bit(FLG_DCHAN_BUSY, &l2->flag)) { 1405 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, 9); 1406 return; 1407 } 1408 test_and_clear_bit(FLG_T200_RUN, &l2->flag); 1409 if (l2->rc == l2->N200) { 1410 l2mgr(l2, MDL_ERROR_IND, (void *) 'I'); 1411 establishlink(fi); 1412 test_and_clear_bit(FLG_L3_INIT, &l2->flag); 1413 } else { 1414 transmit_enquiry(l2); 1415 l2->rc++; 1416 } 1417 } 1418 1419 static void 1420 l2_st7_tout_203(struct FsmInst *fi, int event, void *arg) 1421 { 1422 struct layer2 *l2 = fi->userdata; 1423 1424 if (test_bit(FLG_LAPD, &l2->flag) && 1425 test_bit(FLG_DCHAN_BUSY, &l2->flag)) { 1426 mISDN_FsmAddTimer(&l2->t203, l2->T203, EV_L2_T203, NULL, 9); 1427 return; 1428 } 1429 mISDN_FsmChangeState(fi, ST_L2_8); 1430 transmit_enquiry(l2); 1431 l2->rc = 0; 1432 } 1433 1434 static void 1435 l2_pull_iqueue(struct FsmInst *fi, int event, void *arg) 1436 { 1437 struct layer2 *l2 = fi->userdata; 1438 struct sk_buff *skb, *nskb, *oskb; 1439 u_char header[MAX_L2HEADER_LEN]; 1440 u_int i, p1; 1441 1442 if (!cansend(l2)) 1443 return; 1444 1445 skb = skb_dequeue(&l2->i_queue); 1446 if (!skb) 1447 return; 1448 1449 if (test_bit(FLG_MOD128, &l2->flag)) 1450 p1 = (l2->vs - l2->va) % 128; 1451 else 1452 p1 = (l2->vs - l2->va) % 8; 1453 p1 = (p1 + l2->sow) % l2->window; 1454 if (l2->windowar[p1]) { 1455 printk(KERN_WARNING "isdnl2 try overwrite ack queue entry %d\n", 1456 p1); 1457 dev_kfree_skb(l2->windowar[p1]); 1458 } 1459 l2->windowar[p1] = skb; 1460 i = sethdraddr(l2, header, CMD); 1461 if (test_bit(FLG_MOD128, &l2->flag)) { 1462 header[i++] = l2->vs << 1; 1463 header[i++] = l2->vr << 1; 1464 l2->vs = (l2->vs + 1) % 128; 1465 } else { 1466 header[i++] = (l2->vr << 5) | (l2->vs << 1); 1467 l2->vs = (l2->vs + 1) % 8; 1468 } 1469 1470 nskb = skb_clone(skb, GFP_ATOMIC); 1471 p1 = skb_headroom(nskb); 1472 if (p1 >= i) 1473 memcpy(skb_push(nskb, i), header, i); 1474 else { 1475 printk(KERN_WARNING 1476 "isdnl2 pull_iqueue skb header(%d/%d) too short\n", i, p1); 1477 oskb = nskb; 1478 nskb = mI_alloc_skb(oskb->len + i, GFP_ATOMIC); 1479 if (!nskb) { 1480 dev_kfree_skb(oskb); 1481 printk(KERN_WARNING "%s: no skb mem\n", __func__); 1482 return; 1483 } 1484 memcpy(skb_put(nskb, i), header, i); 1485 memcpy(skb_put(nskb, oskb->len), oskb->data, oskb->len); 1486 dev_kfree_skb(oskb); 1487 } 1488 l2down(l2, PH_DATA_REQ, l2_newid(l2), nskb); 1489 test_and_clear_bit(FLG_ACK_PEND, &l2->flag); 1490 if (!test_and_set_bit(FLG_T200_RUN, &l2->flag)) { 1491 mISDN_FsmDelTimer(&l2->t203, 13); 1492 mISDN_FsmAddTimer(&l2->t200, l2->T200, EV_L2_T200, NULL, 11); 1493 } 1494 } 1495 1496 static void 1497 l2_st8_got_super(struct FsmInst *fi, int event, void *arg) 1498 { 1499 struct layer2 *l2 = fi->userdata; 1500 struct sk_buff *skb = arg; 1501 int PollFlag, rsp, rnr = 0; 1502 unsigned int nr; 1503 1504 rsp = *skb->data & 0x2; 1505 if (test_bit(FLG_ORIG, &l2->flag)) 1506 rsp = !rsp; 1507 1508 skb_pull(skb, l2addrsize(l2)); 1509 1510 if (IsRNR(skb->data, l2)) { 1511 set_peer_busy(l2); 1512 rnr = 1; 1513 } else 1514 clear_peer_busy(l2); 1515 1516 if (test_bit(FLG_MOD128, &l2->flag)) { 1517 PollFlag = (skb->data[1] & 0x1) == 0x1; 1518 nr = skb->data[1] >> 1; 1519 } else { 1520 PollFlag = (skb->data[0] & 0x10); 1521 nr = (skb->data[0] >> 5) & 0x7; 1522 } 1523 dev_kfree_skb(skb); 1524 if (rsp && PollFlag) { 1525 if (legalnr(l2, nr)) { 1526 if (rnr) { 1527 restart_t200(l2, 15); 1528 } else { 1529 stop_t200(l2, 16); 1530 mISDN_FsmAddTimer(&l2->t203, l2->T203, 1531 EV_L2_T203, NULL, 5); 1532 setva(l2, nr); 1533 } 1534 invoke_retransmission(l2, nr); 1535 mISDN_FsmChangeState(fi, ST_L2_7); 1536 if (skb_queue_len(&l2->i_queue) && cansend(l2)) 1537 mISDN_FsmEvent(fi, EV_L2_ACK_PULL, NULL); 1538 } else 1539 nrerrorrecovery(fi); 1540 } else { 1541 if (!rsp && PollFlag) 1542 enquiry_response(l2); 1543 if (legalnr(l2, nr)) 1544 setva(l2, nr); 1545 else 1546 nrerrorrecovery(fi); 1547 } 1548 } 1549 1550 static void 1551 l2_got_FRMR(struct FsmInst *fi, int event, void *arg) 1552 { 1553 struct layer2 *l2 = fi->userdata; 1554 struct sk_buff *skb = arg; 1555 1556 skb_pull(skb, l2addrsize(l2) + 1); 1557 1558 if (!(skb->data[0] & 1) || ((skb->data[0] & 3) == 1) || /* I or S */ 1559 (IsUA(skb->data) && (fi->state == ST_L2_7))) { 1560 l2mgr(l2, MDL_ERROR_IND, (void *) 'K'); 1561 establishlink(fi); 1562 test_and_clear_bit(FLG_L3_INIT, &l2->flag); 1563 } 1564 dev_kfree_skb(skb); 1565 } 1566 1567 static void 1568 l2_st24_tei_remove(struct FsmInst *fi, int event, void *arg) 1569 { 1570 struct layer2 *l2 = fi->userdata; 1571 1572 skb_queue_purge(&l2->ui_queue); 1573 l2->tei = GROUP_TEI; 1574 mISDN_FsmChangeState(fi, ST_L2_1); 1575 } 1576 1577 static void 1578 l2_st3_tei_remove(struct FsmInst *fi, int event, void *arg) 1579 { 1580 struct layer2 *l2 = fi->userdata; 1581 1582 skb_queue_purge(&l2->ui_queue); 1583 l2->tei = GROUP_TEI; 1584 l2up_create(l2, DL_RELEASE_IND, 0, NULL); 1585 mISDN_FsmChangeState(fi, ST_L2_1); 1586 } 1587 1588 static void 1589 l2_st5_tei_remove(struct FsmInst *fi, int event, void *arg) 1590 { 1591 struct layer2 *l2 = fi->userdata; 1592 1593 skb_queue_purge(&l2->i_queue); 1594 skb_queue_purge(&l2->ui_queue); 1595 freewin(l2); 1596 l2->tei = GROUP_TEI; 1597 stop_t200(l2, 17); 1598 st5_dl_release_l2l3(l2); 1599 mISDN_FsmChangeState(fi, ST_L2_1); 1600 } 1601 1602 static void 1603 l2_st6_tei_remove(struct FsmInst *fi, int event, void *arg) 1604 { 1605 struct layer2 *l2 = fi->userdata; 1606 1607 skb_queue_purge(&l2->ui_queue); 1608 l2->tei = GROUP_TEI; 1609 stop_t200(l2, 18); 1610 l2up_create(l2, DL_RELEASE_IND, 0, NULL); 1611 mISDN_FsmChangeState(fi, ST_L2_1); 1612 } 1613 1614 static void 1615 l2_tei_remove(struct FsmInst *fi, int event, void *arg) 1616 { 1617 struct layer2 *l2 = fi->userdata; 1618 1619 skb_queue_purge(&l2->i_queue); 1620 skb_queue_purge(&l2->ui_queue); 1621 freewin(l2); 1622 l2->tei = GROUP_TEI; 1623 stop_t200(l2, 17); 1624 mISDN_FsmDelTimer(&l2->t203, 19); 1625 l2up_create(l2, DL_RELEASE_IND, 0, NULL); 1626 /* mISDN_queue_data(&l2->inst, l2->inst.id | MSG_BROADCAST, 1627 * MGR_SHORTSTATUS_IND, SSTATUS_L2_RELEASED, 1628 * 0, NULL, 0); 1629 */ 1630 mISDN_FsmChangeState(fi, ST_L2_1); 1631 } 1632 1633 static void 1634 l2_st14_persistant_da(struct FsmInst *fi, int event, void *arg) 1635 { 1636 struct layer2 *l2 = fi->userdata; 1637 struct sk_buff *skb = arg; 1638 1639 skb_queue_purge(&l2->i_queue); 1640 skb_queue_purge(&l2->ui_queue); 1641 if (test_and_clear_bit(FLG_ESTAB_PEND, &l2->flag)) 1642 l2up(l2, DL_RELEASE_IND, skb); 1643 else 1644 dev_kfree_skb(skb); 1645 } 1646 1647 static void 1648 l2_st5_persistant_da(struct FsmInst *fi, int event, void *arg) 1649 { 1650 struct layer2 *l2 = fi->userdata; 1651 struct sk_buff *skb = arg; 1652 1653 skb_queue_purge(&l2->i_queue); 1654 skb_queue_purge(&l2->ui_queue); 1655 freewin(l2); 1656 stop_t200(l2, 19); 1657 st5_dl_release_l2l3(l2); 1658 mISDN_FsmChangeState(fi, ST_L2_4); 1659 if (l2->tm) 1660 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1661 dev_kfree_skb(skb); 1662 } 1663 1664 static void 1665 l2_st6_persistant_da(struct FsmInst *fi, int event, void *arg) 1666 { 1667 struct layer2 *l2 = fi->userdata; 1668 struct sk_buff *skb = arg; 1669 1670 skb_queue_purge(&l2->ui_queue); 1671 stop_t200(l2, 20); 1672 l2up(l2, DL_RELEASE_CNF, skb); 1673 mISDN_FsmChangeState(fi, ST_L2_4); 1674 if (l2->tm) 1675 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1676 } 1677 1678 static void 1679 l2_persistant_da(struct FsmInst *fi, int event, void *arg) 1680 { 1681 struct layer2 *l2 = fi->userdata; 1682 struct sk_buff *skb = arg; 1683 1684 skb_queue_purge(&l2->i_queue); 1685 skb_queue_purge(&l2->ui_queue); 1686 freewin(l2); 1687 stop_t200(l2, 19); 1688 mISDN_FsmDelTimer(&l2->t203, 19); 1689 l2up(l2, DL_RELEASE_IND, skb); 1690 mISDN_FsmChangeState(fi, ST_L2_4); 1691 if (l2->tm) 1692 l2_tei(l2, MDL_STATUS_DOWN_IND, 0); 1693 } 1694 1695 static void 1696 l2_set_own_busy(struct FsmInst *fi, int event, void *arg) 1697 { 1698 struct layer2 *l2 = fi->userdata; 1699 struct sk_buff *skb = arg; 1700 1701 if (!test_and_set_bit(FLG_OWN_BUSY, &l2->flag)) { 1702 enquiry_cr(l2, RNR, RSP, 0); 1703 test_and_clear_bit(FLG_ACK_PEND, &l2->flag); 1704 } 1705 if (skb) 1706 dev_kfree_skb(skb); 1707 } 1708 1709 static void 1710 l2_clear_own_busy(struct FsmInst *fi, int event, void *arg) 1711 { 1712 struct layer2 *l2 = fi->userdata; 1713 struct sk_buff *skb = arg; 1714 1715 if (!test_and_clear_bit(FLG_OWN_BUSY, &l2->flag)) { 1716 enquiry_cr(l2, RR, RSP, 0); 1717 test_and_clear_bit(FLG_ACK_PEND, &l2->flag); 1718 } 1719 if (skb) 1720 dev_kfree_skb(skb); 1721 } 1722 1723 static void 1724 l2_frame_error(struct FsmInst *fi, int event, void *arg) 1725 { 1726 struct layer2 *l2 = fi->userdata; 1727 1728 l2mgr(l2, MDL_ERROR_IND, arg); 1729 } 1730 1731 static void 1732 l2_frame_error_reest(struct FsmInst *fi, int event, void *arg) 1733 { 1734 struct layer2 *l2 = fi->userdata; 1735 1736 l2mgr(l2, MDL_ERROR_IND, arg); 1737 establishlink(fi); 1738 test_and_clear_bit(FLG_L3_INIT, &l2->flag); 1739 } 1740 1741 static struct FsmNode L2FnList[] = 1742 { 1743 {ST_L2_1, EV_L2_DL_ESTABLISH_REQ, l2_mdl_assign}, 1744 {ST_L2_2, EV_L2_DL_ESTABLISH_REQ, l2_go_st3}, 1745 {ST_L2_4, EV_L2_DL_ESTABLISH_REQ, l2_establish}, 1746 {ST_L2_5, EV_L2_DL_ESTABLISH_REQ, l2_discard_i_setl3}, 1747 {ST_L2_7, EV_L2_DL_ESTABLISH_REQ, l2_l3_reestablish}, 1748 {ST_L2_8, EV_L2_DL_ESTABLISH_REQ, l2_l3_reestablish}, 1749 {ST_L2_4, EV_L2_DL_RELEASE_REQ, l2_release}, 1750 {ST_L2_5, EV_L2_DL_RELEASE_REQ, l2_pend_rel}, 1751 {ST_L2_7, EV_L2_DL_RELEASE_REQ, l2_disconnect}, 1752 {ST_L2_8, EV_L2_DL_RELEASE_REQ, l2_disconnect}, 1753 {ST_L2_5, EV_L2_DL_DATA, l2_feed_i_if_reest}, 1754 {ST_L2_7, EV_L2_DL_DATA, l2_feed_i_pull}, 1755 {ST_L2_8, EV_L2_DL_DATA, l2_feed_iqueue}, 1756 {ST_L2_1, EV_L2_DL_UNITDATA, l2_queue_ui_assign}, 1757 {ST_L2_2, EV_L2_DL_UNITDATA, l2_queue_ui}, 1758 {ST_L2_3, EV_L2_DL_UNITDATA, l2_queue_ui}, 1759 {ST_L2_4, EV_L2_DL_UNITDATA, l2_send_ui}, 1760 {ST_L2_5, EV_L2_DL_UNITDATA, l2_send_ui}, 1761 {ST_L2_6, EV_L2_DL_UNITDATA, l2_send_ui}, 1762 {ST_L2_7, EV_L2_DL_UNITDATA, l2_send_ui}, 1763 {ST_L2_8, EV_L2_DL_UNITDATA, l2_send_ui}, 1764 {ST_L2_1, EV_L2_MDL_ASSIGN, l2_got_tei}, 1765 {ST_L2_2, EV_L2_MDL_ASSIGN, l2_got_tei}, 1766 {ST_L2_3, EV_L2_MDL_ASSIGN, l2_got_tei}, 1767 {ST_L2_2, EV_L2_MDL_ERROR, l2_st24_tei_remove}, 1768 {ST_L2_3, EV_L2_MDL_ERROR, l2_st3_tei_remove}, 1769 {ST_L2_4, EV_L2_MDL_REMOVE, l2_st24_tei_remove}, 1770 {ST_L2_5, EV_L2_MDL_REMOVE, l2_st5_tei_remove}, 1771 {ST_L2_6, EV_L2_MDL_REMOVE, l2_st6_tei_remove}, 1772 {ST_L2_7, EV_L2_MDL_REMOVE, l2_tei_remove}, 1773 {ST_L2_8, EV_L2_MDL_REMOVE, l2_tei_remove}, 1774 {ST_L2_4, EV_L2_SABME, l2_start_multi}, 1775 {ST_L2_5, EV_L2_SABME, l2_send_UA}, 1776 {ST_L2_6, EV_L2_SABME, l2_send_DM}, 1777 {ST_L2_7, EV_L2_SABME, l2_restart_multi}, 1778 {ST_L2_8, EV_L2_SABME, l2_restart_multi}, 1779 {ST_L2_4, EV_L2_DISC, l2_send_DM}, 1780 {ST_L2_5, EV_L2_DISC, l2_send_DM}, 1781 {ST_L2_6, EV_L2_DISC, l2_send_UA}, 1782 {ST_L2_7, EV_L2_DISC, l2_stop_multi}, 1783 {ST_L2_8, EV_L2_DISC, l2_stop_multi}, 1784 {ST_L2_4, EV_L2_UA, l2_mdl_error_ua}, 1785 {ST_L2_5, EV_L2_UA, l2_connected}, 1786 {ST_L2_6, EV_L2_UA, l2_released}, 1787 {ST_L2_7, EV_L2_UA, l2_mdl_error_ua}, 1788 {ST_L2_8, EV_L2_UA, l2_mdl_error_ua}, 1789 {ST_L2_4, EV_L2_DM, l2_reestablish}, 1790 {ST_L2_5, EV_L2_DM, l2_st5_dm_release}, 1791 {ST_L2_6, EV_L2_DM, l2_st6_dm_release}, 1792 {ST_L2_7, EV_L2_DM, l2_mdl_error_dm}, 1793 {ST_L2_8, EV_L2_DM, l2_st8_mdl_error_dm}, 1794 {ST_L2_1, EV_L2_UI, l2_got_ui}, 1795 {ST_L2_2, EV_L2_UI, l2_got_ui}, 1796 {ST_L2_3, EV_L2_UI, l2_got_ui}, 1797 {ST_L2_4, EV_L2_UI, l2_got_ui}, 1798 {ST_L2_5, EV_L2_UI, l2_got_ui}, 1799 {ST_L2_6, EV_L2_UI, l2_got_ui}, 1800 {ST_L2_7, EV_L2_UI, l2_got_ui}, 1801 {ST_L2_8, EV_L2_UI, l2_got_ui}, 1802 {ST_L2_7, EV_L2_FRMR, l2_got_FRMR}, 1803 {ST_L2_8, EV_L2_FRMR, l2_got_FRMR}, 1804 {ST_L2_7, EV_L2_SUPER, l2_st7_got_super}, 1805 {ST_L2_8, EV_L2_SUPER, l2_st8_got_super}, 1806 {ST_L2_7, EV_L2_I, l2_got_iframe}, 1807 {ST_L2_8, EV_L2_I, l2_got_iframe}, 1808 {ST_L2_5, EV_L2_T200, l2_st5_tout_200}, 1809 {ST_L2_6, EV_L2_T200, l2_st6_tout_200}, 1810 {ST_L2_7, EV_L2_T200, l2_st7_tout_200}, 1811 {ST_L2_8, EV_L2_T200, l2_st8_tout_200}, 1812 {ST_L2_7, EV_L2_T203, l2_st7_tout_203}, 1813 {ST_L2_7, EV_L2_ACK_PULL, l2_pull_iqueue}, 1814 {ST_L2_7, EV_L2_SET_OWN_BUSY, l2_set_own_busy}, 1815 {ST_L2_8, EV_L2_SET_OWN_BUSY, l2_set_own_busy}, 1816 {ST_L2_7, EV_L2_CLEAR_OWN_BUSY, l2_clear_own_busy}, 1817 {ST_L2_8, EV_L2_CLEAR_OWN_BUSY, l2_clear_own_busy}, 1818 {ST_L2_4, EV_L2_FRAME_ERROR, l2_frame_error}, 1819 {ST_L2_5, EV_L2_FRAME_ERROR, l2_frame_error}, 1820 {ST_L2_6, EV_L2_FRAME_ERROR, l2_frame_error}, 1821 {ST_L2_7, EV_L2_FRAME_ERROR, l2_frame_error_reest}, 1822 {ST_L2_8, EV_L2_FRAME_ERROR, l2_frame_error_reest}, 1823 {ST_L2_1, EV_L1_DEACTIVATE, l2_st14_persistant_da}, 1824 {ST_L2_2, EV_L1_DEACTIVATE, l2_st24_tei_remove}, 1825 {ST_L2_3, EV_L1_DEACTIVATE, l2_st3_tei_remove}, 1826 {ST_L2_4, EV_L1_DEACTIVATE, l2_st14_persistant_da}, 1827 {ST_L2_5, EV_L1_DEACTIVATE, l2_st5_persistant_da}, 1828 {ST_L2_6, EV_L1_DEACTIVATE, l2_st6_persistant_da}, 1829 {ST_L2_7, EV_L1_DEACTIVATE, l2_persistant_da}, 1830 {ST_L2_8, EV_L1_DEACTIVATE, l2_persistant_da}, 1831 }; 1832 1833 #define L2_FN_COUNT (sizeof(L2FnList)/sizeof(struct FsmNode)) 1834 1835 static int 1836 ph_data_indication(struct layer2 *l2, struct mISDNhead *hh, struct sk_buff *skb) 1837 { 1838 u_char *datap = skb->data; 1839 int ret = -EINVAL; 1840 int psapi, ptei; 1841 u_int l; 1842 int c = 0; 1843 1844 l = l2addrsize(l2); 1845 if (skb->len <= l) { 1846 mISDN_FsmEvent(&l2->l2m, EV_L2_FRAME_ERROR, (void *) 'N'); 1847 return ret; 1848 } 1849 if (test_bit(FLG_LAPD, &l2->flag)) { /* Maybe not needed */ 1850 psapi = *datap++; 1851 ptei = *datap++; 1852 if ((psapi & 1) || !(ptei & 1)) { 1853 printk(KERN_WARNING 1854 "l2 D-channel frame wrong EA0/EA1\n"); 1855 return ret; 1856 } 1857 psapi >>= 2; 1858 ptei >>= 1; 1859 if (psapi != l2->sapi) { 1860 /* not our bussiness 1861 * printk(KERN_DEBUG "%s: sapi %d/%d sapi mismatch\n", 1862 * __func__, 1863 * psapi, l2->sapi); 1864 */ 1865 dev_kfree_skb(skb); 1866 return 0; 1867 } 1868 if ((ptei != l2->tei) && (ptei != GROUP_TEI)) { 1869 /* not our bussiness 1870 * printk(KERN_DEBUG "%s: tei %d/%d sapi %d mismatch\n", 1871 * __func__, 1872 * ptei, l2->tei, psapi); 1873 */ 1874 dev_kfree_skb(skb); 1875 return 0; 1876 } 1877 } else 1878 datap += l; 1879 if (!(*datap & 1)) { /* I-Frame */ 1880 c = iframe_error(l2, skb); 1881 if (!c) 1882 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_I, skb); 1883 } else if (IsSFrame(datap, l2)) { /* S-Frame */ 1884 c = super_error(l2, skb); 1885 if (!c) 1886 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_SUPER, skb); 1887 } else if (IsUI(datap)) { 1888 c = UI_error(l2, skb); 1889 if (!c) 1890 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_UI, skb); 1891 } else if (IsSABME(datap, l2)) { 1892 c = unnum_error(l2, skb, CMD); 1893 if (!c) 1894 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_SABME, skb); 1895 } else if (IsUA(datap)) { 1896 c = unnum_error(l2, skb, RSP); 1897 if (!c) 1898 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_UA, skb); 1899 } else if (IsDISC(datap)) { 1900 c = unnum_error(l2, skb, CMD); 1901 if (!c) 1902 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_DISC, skb); 1903 } else if (IsDM(datap)) { 1904 c = unnum_error(l2, skb, RSP); 1905 if (!c) 1906 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_DM, skb); 1907 } else if (IsFRMR(datap)) { 1908 c = FRMR_error(l2, skb); 1909 if (!c) 1910 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_FRMR, skb); 1911 } else 1912 c = 'L'; 1913 if (c) { 1914 printk(KERN_WARNING "l2 D-channel frame error %c\n", c); 1915 mISDN_FsmEvent(&l2->l2m, EV_L2_FRAME_ERROR, (void *)(long)c); 1916 } 1917 return ret; 1918 } 1919 1920 static int 1921 l2_send(struct mISDNchannel *ch, struct sk_buff *skb) 1922 { 1923 struct layer2 *l2 = container_of(ch, struct layer2, ch); 1924 struct mISDNhead *hh = mISDN_HEAD_P(skb); 1925 int ret = -EINVAL; 1926 1927 if (*debug & DEBUG_L2_RECV) 1928 printk(KERN_DEBUG "%s: prim(%x) id(%x) tei(%d)\n", 1929 __func__, hh->prim, hh->id, l2->tei); 1930 switch (hh->prim) { 1931 case PH_DATA_IND: 1932 ret = ph_data_indication(l2, hh, skb); 1933 break; 1934 case PH_DATA_CNF: 1935 ret = ph_data_confirm(l2, hh, skb); 1936 break; 1937 case PH_ACTIVATE_IND: 1938 test_and_set_bit(FLG_L1_ACTIV, &l2->flag); 1939 l2up_create(l2, MPH_ACTIVATE_IND, 0, NULL); 1940 if (test_and_clear_bit(FLG_ESTAB_PEND, &l2->flag)) 1941 ret = mISDN_FsmEvent(&l2->l2m, 1942 EV_L2_DL_ESTABLISH_REQ, skb); 1943 break; 1944 case PH_DEACTIVATE_IND: 1945 test_and_clear_bit(FLG_L1_ACTIV, &l2->flag); 1946 l2up_create(l2, MPH_DEACTIVATE_IND, 0, NULL); 1947 ret = mISDN_FsmEvent(&l2->l2m, EV_L1_DEACTIVATE, skb); 1948 break; 1949 case MPH_INFORMATION_IND: 1950 if (!l2->up) 1951 break; 1952 ret = l2->up->send(l2->up, skb); 1953 break; 1954 case DL_DATA_REQ: 1955 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_DL_DATA, skb); 1956 break; 1957 case DL_UNITDATA_REQ: 1958 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_DL_UNITDATA, skb); 1959 break; 1960 case DL_ESTABLISH_REQ: 1961 if (test_bit(FLG_LAPB, &l2->flag)) 1962 test_and_set_bit(FLG_ORIG, &l2->flag); 1963 if (test_bit(FLG_L1_ACTIV, &l2->flag)) { 1964 if (test_bit(FLG_LAPD, &l2->flag) || 1965 test_bit(FLG_ORIG, &l2->flag)) 1966 ret = mISDN_FsmEvent(&l2->l2m, 1967 EV_L2_DL_ESTABLISH_REQ, skb); 1968 } else { 1969 if (test_bit(FLG_LAPD, &l2->flag) || 1970 test_bit(FLG_ORIG, &l2->flag)) { 1971 test_and_set_bit(FLG_ESTAB_PEND, 1972 &l2->flag); 1973 } 1974 ret = l2down(l2, PH_ACTIVATE_REQ, l2_newid(l2), 1975 skb); 1976 } 1977 break; 1978 case DL_RELEASE_REQ: 1979 if (test_bit(FLG_LAPB, &l2->flag)) 1980 l2down_create(l2, PH_DEACTIVATE_REQ, 1981 l2_newid(l2), 0, NULL); 1982 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_DL_RELEASE_REQ, 1983 skb); 1984 break; 1985 default: 1986 if (*debug & DEBUG_L2) 1987 l2m_debug(&l2->l2m, "l2 unknown pr %04x", 1988 hh->prim); 1989 } 1990 if (ret) { 1991 dev_kfree_skb(skb); 1992 ret = 0; 1993 } 1994 return ret; 1995 } 1996 1997 int 1998 tei_l2(struct layer2 *l2, u_int cmd, u_long arg) 1999 { 2000 int ret = -EINVAL; 2001 2002 if (*debug & DEBUG_L2_TEI) 2003 printk(KERN_DEBUG "%s: cmd(%x)\n", __func__, cmd); 2004 switch (cmd) { 2005 case (MDL_ASSIGN_REQ): 2006 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_MDL_ASSIGN, (void *)arg); 2007 break; 2008 case (MDL_REMOVE_REQ): 2009 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_MDL_REMOVE, NULL); 2010 break; 2011 case (MDL_ERROR_IND): 2012 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_MDL_ERROR, NULL); 2013 break; 2014 case (MDL_ERROR_RSP): 2015 /* ETS 300-125 5.3.2.1 Test: TC13010 */ 2016 printk(KERN_NOTICE "MDL_ERROR|REQ (tei_l2)\n"); 2017 ret = mISDN_FsmEvent(&l2->l2m, EV_L2_MDL_ERROR, NULL); 2018 break; 2019 } 2020 return ret; 2021 } 2022 2023 static void 2024 release_l2(struct layer2 *l2) 2025 { 2026 mISDN_FsmDelTimer(&l2->t200, 21); 2027 mISDN_FsmDelTimer(&l2->t203, 16); 2028 skb_queue_purge(&l2->i_queue); 2029 skb_queue_purge(&l2->ui_queue); 2030 skb_queue_purge(&l2->down_queue); 2031 ReleaseWin(l2); 2032 if (test_bit(FLG_LAPD, &l2->flag)) { 2033 TEIrelease(l2); 2034 if (l2->ch.st) 2035 l2->ch.st->dev->D.ctrl(&l2->ch.st->dev->D, 2036 CLOSE_CHANNEL, NULL); 2037 } 2038 kfree(l2); 2039 } 2040 2041 static int 2042 l2_ctrl(struct mISDNchannel *ch, u_int cmd, void *arg) 2043 { 2044 struct layer2 *l2 = container_of(ch, struct layer2, ch); 2045 u_int info; 2046 2047 if (*debug & DEBUG_L2_CTRL) 2048 printk(KERN_DEBUG "%s:(%x)\n", __func__, cmd); 2049 2050 switch (cmd) { 2051 case OPEN_CHANNEL: 2052 if (test_bit(FLG_LAPD, &l2->flag)) { 2053 set_channel_address(&l2->ch, l2->sapi, l2->tei); 2054 info = DL_INFO_L2_CONNECT; 2055 l2up_create(l2, DL_INFORMATION_IND, 2056 sizeof(info), &info); 2057 } 2058 break; 2059 case CLOSE_CHANNEL: 2060 if (l2->ch.peer) 2061 l2->ch.peer->ctrl(l2->ch.peer, CLOSE_CHANNEL, NULL); 2062 release_l2(l2); 2063 break; 2064 } 2065 return 0; 2066 } 2067 2068 struct layer2 * 2069 create_l2(struct mISDNchannel *ch, u_int protocol, u_long options, u_long arg) 2070 { 2071 struct layer2 *l2; 2072 struct channel_req rq; 2073 2074 l2 = kzalloc(sizeof(struct layer2), GFP_KERNEL); 2075 if (!l2) { 2076 printk(KERN_ERR "kzalloc layer2 failed\n"); 2077 return NULL; 2078 } 2079 l2->next_id = 1; 2080 l2->down_id = MISDN_ID_NONE; 2081 l2->up = ch; 2082 l2->ch.st = ch->st; 2083 l2->ch.send = l2_send; 2084 l2->ch.ctrl = l2_ctrl; 2085 switch (protocol) { 2086 case ISDN_P_LAPD_NT: 2087 test_and_set_bit(FLG_LAPD, &l2->flag); 2088 test_and_set_bit(FLG_LAPD_NET, &l2->flag); 2089 test_and_set_bit(FLG_MOD128, &l2->flag); 2090 l2->sapi = 0; 2091 l2->maxlen = MAX_DFRAME_LEN; 2092 if (test_bit(OPTION_L2_PMX, &options)) 2093 l2->window = 7; 2094 else 2095 l2->window = 1; 2096 if (test_bit(OPTION_L2_PTP, &options)) 2097 test_and_set_bit(FLG_PTP, &l2->flag); 2098 if (test_bit(OPTION_L2_FIXEDTEI, &options)) 2099 test_and_set_bit(FLG_FIXED_TEI, &l2->flag); 2100 l2->tei = (u_int)arg; 2101 l2->T200 = 1000; 2102 l2->N200 = 3; 2103 l2->T203 = 10000; 2104 if (test_bit(OPTION_L2_PMX, &options)) 2105 rq.protocol = ISDN_P_NT_E1; 2106 else 2107 rq.protocol = ISDN_P_NT_S0; 2108 rq.adr.channel = 0; 2109 l2->ch.st->dev->D.ctrl(&l2->ch.st->dev->D, OPEN_CHANNEL, &rq); 2110 break; 2111 case ISDN_P_LAPD_TE: 2112 test_and_set_bit(FLG_LAPD, &l2->flag); 2113 test_and_set_bit(FLG_MOD128, &l2->flag); 2114 test_and_set_bit(FLG_ORIG, &l2->flag); 2115 l2->sapi = 0; 2116 l2->maxlen = MAX_DFRAME_LEN; 2117 if (test_bit(OPTION_L2_PMX, &options)) 2118 l2->window = 7; 2119 else 2120 l2->window = 1; 2121 if (test_bit(OPTION_L2_PTP, &options)) 2122 test_and_set_bit(FLG_PTP, &l2->flag); 2123 if (test_bit(OPTION_L2_FIXEDTEI, &options)) 2124 test_and_set_bit(FLG_FIXED_TEI, &l2->flag); 2125 l2->tei = (u_int)arg; 2126 l2->T200 = 1000; 2127 l2->N200 = 3; 2128 l2->T203 = 10000; 2129 if (test_bit(OPTION_L2_PMX, &options)) 2130 rq.protocol = ISDN_P_TE_E1; 2131 else 2132 rq.protocol = ISDN_P_TE_S0; 2133 rq.adr.channel = 0; 2134 l2->ch.st->dev->D.ctrl(&l2->ch.st->dev->D, OPEN_CHANNEL, &rq); 2135 break; 2136 case ISDN_P_B_X75SLP: 2137 test_and_set_bit(FLG_LAPB, &l2->flag); 2138 l2->window = 7; 2139 l2->maxlen = MAX_DATA_SIZE; 2140 l2->T200 = 1000; 2141 l2->N200 = 4; 2142 l2->T203 = 5000; 2143 l2->addr.A = 3; 2144 l2->addr.B = 1; 2145 break; 2146 default: 2147 printk(KERN_ERR "layer2 create failed prt %x\n", 2148 protocol); 2149 kfree(l2); 2150 return NULL; 2151 } 2152 skb_queue_head_init(&l2->i_queue); 2153 skb_queue_head_init(&l2->ui_queue); 2154 skb_queue_head_init(&l2->down_queue); 2155 skb_queue_head_init(&l2->tmp_queue); 2156 InitWin(l2); 2157 l2->l2m.fsm = &l2fsm; 2158 if (test_bit(FLG_LAPB, &l2->flag) || 2159 test_bit(FLG_PTP, &l2->flag) || 2160 test_bit(FLG_LAPD_NET, &l2->flag)) 2161 l2->l2m.state = ST_L2_4; 2162 else 2163 l2->l2m.state = ST_L2_1; 2164 l2->l2m.debug = *debug; 2165 l2->l2m.userdata = l2; 2166 l2->l2m.userint = 0; 2167 l2->l2m.printdebug = l2m_debug; 2168 2169 mISDN_FsmInitTimer(&l2->l2m, &l2->t200); 2170 mISDN_FsmInitTimer(&l2->l2m, &l2->t203); 2171 return l2; 2172 } 2173 2174 static int 2175 x75create(struct channel_req *crq) 2176 { 2177 struct layer2 *l2; 2178 2179 if (crq->protocol != ISDN_P_B_X75SLP) 2180 return -EPROTONOSUPPORT; 2181 l2 = create_l2(crq->ch, crq->protocol, 0, 0); 2182 if (!l2) 2183 return -ENOMEM; 2184 crq->ch = &l2->ch; 2185 crq->protocol = ISDN_P_B_HDLC; 2186 return 0; 2187 } 2188 2189 static struct Bprotocol X75SLP = { 2190 .Bprotocols = (1 << (ISDN_P_B_X75SLP & ISDN_P_B_MASK)), 2191 .name = "X75SLP", 2192 .create = x75create 2193 }; 2194 2195 int 2196 Isdnl2_Init(u_int *deb) 2197 { 2198 debug = deb; 2199 mISDN_register_Bprotocol(&X75SLP); 2200 l2fsm.state_count = L2_STATE_COUNT; 2201 l2fsm.event_count = L2_EVENT_COUNT; 2202 l2fsm.strEvent = strL2Event; 2203 l2fsm.strState = strL2State; 2204 mISDN_FsmNew(&l2fsm, L2FnList, ARRAY_SIZE(L2FnList)); 2205 TEIInit(deb); 2206 return 0; 2207 } 2208 2209 void 2210 Isdnl2_cleanup(void) 2211 { 2212 mISDN_unregister_Bprotocol(&X75SLP); 2213 TEIFree(); 2214 mISDN_FsmFree(&l2fsm); 2215 } 2216 2217