1 /* 2 BlueZ - Bluetooth protocol stack for Linux 3 Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved. 4 5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com> 6 7 This program is free software; you can redistribute it and/or modify 8 it under the terms of the GNU General Public License version 2 as 9 published by the Free Software Foundation; 10 11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS 12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS. 14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY 15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES 16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 19 20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS, 21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS 22 SOFTWARE IS DISCLAIMED. 23 */ 24 25 #ifndef __HCI_CORE_H 26 #define __HCI_CORE_H 27 28 #include <linux/interrupt.h> 29 #include <net/bluetooth/hci.h> 30 31 /* HCI upper protocols */ 32 #define HCI_PROTO_L2CAP 0 33 #define HCI_PROTO_SCO 1 34 35 /* HCI priority */ 36 #define HCI_PRIO_MAX 7 37 38 /* HCI Core structures */ 39 struct inquiry_data { 40 bdaddr_t bdaddr; 41 __u8 pscan_rep_mode; 42 __u8 pscan_period_mode; 43 __u8 pscan_mode; 44 __u8 dev_class[3]; 45 __le16 clock_offset; 46 __s8 rssi; 47 __u8 ssp_mode; 48 }; 49 50 struct inquiry_entry { 51 struct inquiry_entry *next; 52 __u32 timestamp; 53 struct inquiry_data data; 54 }; 55 56 struct inquiry_cache { 57 spinlock_t lock; 58 __u32 timestamp; 59 struct inquiry_entry *list; 60 }; 61 62 struct hci_conn_hash { 63 struct list_head list; 64 spinlock_t lock; 65 unsigned int acl_num; 66 unsigned int sco_num; 67 unsigned int le_num; 68 }; 69 70 struct hci_chan_hash { 71 struct list_head list; 72 spinlock_t lock; 73 unsigned int num; 74 }; 75 76 struct bdaddr_list { 77 struct list_head list; 78 bdaddr_t bdaddr; 79 }; 80 81 struct bt_uuid { 82 struct list_head list; 83 u8 uuid[16]; 84 u8 svc_hint; 85 }; 86 87 struct key_master_id { 88 __le16 ediv; 89 u8 rand[8]; 90 } __packed; 91 92 struct link_key_data { 93 bdaddr_t bdaddr; 94 u8 type; 95 u8 val[16]; 96 u8 pin_len; 97 u8 dlen; 98 u8 data[0]; 99 } __packed; 100 101 struct link_key { 102 struct list_head list; 103 bdaddr_t bdaddr; 104 u8 type; 105 u8 val[16]; 106 u8 pin_len; 107 u8 dlen; 108 u8 data[0]; 109 }; 110 111 struct oob_data { 112 struct list_head list; 113 bdaddr_t bdaddr; 114 u8 hash[16]; 115 u8 randomizer[16]; 116 }; 117 118 struct adv_entry { 119 struct list_head list; 120 bdaddr_t bdaddr; 121 u8 bdaddr_type; 122 }; 123 124 #define NUM_REASSEMBLY 4 125 struct hci_dev { 126 struct list_head list; 127 spinlock_t lock; 128 atomic_t refcnt; 129 130 char name[8]; 131 unsigned long flags; 132 __u16 id; 133 __u8 bus; 134 __u8 dev_type; 135 bdaddr_t bdaddr; 136 __u8 dev_name[HCI_MAX_NAME_LENGTH]; 137 __u8 eir[HCI_MAX_EIR_LENGTH]; 138 __u8 dev_class[3]; 139 __u8 major_class; 140 __u8 minor_class; 141 __u8 features[8]; 142 __u8 extfeatures[8]; 143 __u8 commands[64]; 144 __u8 ssp_mode; 145 __u8 hci_ver; 146 __u16 hci_rev; 147 __u8 lmp_ver; 148 __u16 manufacturer; 149 __le16 lmp_subver; 150 __u16 voice_setting; 151 __u8 io_capability; 152 153 __u16 pkt_type; 154 __u16 esco_type; 155 __u16 link_policy; 156 __u16 link_mode; 157 158 __u32 idle_timeout; 159 __u16 sniff_min_interval; 160 __u16 sniff_max_interval; 161 162 __u8 amp_status; 163 __u32 amp_total_bw; 164 __u32 amp_max_bw; 165 __u32 amp_min_latency; 166 __u32 amp_max_pdu; 167 __u8 amp_type; 168 __u16 amp_pal_cap; 169 __u16 amp_assoc_size; 170 __u32 amp_max_flush_to; 171 __u32 amp_be_flush_to; 172 173 __u8 flow_ctl_mode; 174 175 unsigned int auto_accept_delay; 176 177 unsigned long quirks; 178 179 atomic_t cmd_cnt; 180 unsigned int acl_cnt; 181 unsigned int sco_cnt; 182 unsigned int le_cnt; 183 184 unsigned int acl_mtu; 185 unsigned int sco_mtu; 186 unsigned int le_mtu; 187 unsigned int acl_pkts; 188 unsigned int sco_pkts; 189 unsigned int le_pkts; 190 191 unsigned long acl_last_tx; 192 unsigned long sco_last_tx; 193 unsigned long le_last_tx; 194 195 struct workqueue_struct *workqueue; 196 197 struct work_struct power_on; 198 struct delayed_work power_off; 199 200 __u16 discov_timeout; 201 struct delayed_work discov_off; 202 203 struct timer_list cmd_timer; 204 struct tasklet_struct cmd_task; 205 struct tasklet_struct rx_task; 206 struct tasklet_struct tx_task; 207 208 struct sk_buff_head rx_q; 209 struct sk_buff_head raw_q; 210 struct sk_buff_head cmd_q; 211 212 struct sk_buff *sent_cmd; 213 struct sk_buff *reassembly[NUM_REASSEMBLY]; 214 215 struct mutex req_lock; 216 wait_queue_head_t req_wait_q; 217 __u32 req_status; 218 __u32 req_result; 219 220 __u16 init_last_cmd; 221 222 struct list_head mgmt_pending; 223 224 struct inquiry_cache inq_cache; 225 struct hci_conn_hash conn_hash; 226 struct list_head blacklist; 227 228 struct list_head uuids; 229 230 struct list_head link_keys; 231 232 struct list_head remote_oob_data; 233 234 struct list_head adv_entries; 235 struct timer_list adv_timer; 236 237 struct hci_dev_stats stat; 238 239 struct sk_buff_head driver_init; 240 241 void *driver_data; 242 void *core_data; 243 244 atomic_t promisc; 245 246 struct dentry *debugfs; 247 248 struct device *parent; 249 struct device dev; 250 251 struct rfkill *rfkill; 252 253 struct module *owner; 254 255 unsigned long dev_flags; 256 257 int (*open)(struct hci_dev *hdev); 258 int (*close)(struct hci_dev *hdev); 259 int (*flush)(struct hci_dev *hdev); 260 int (*send)(struct sk_buff *skb); 261 void (*destruct)(struct hci_dev *hdev); 262 void (*notify)(struct hci_dev *hdev, unsigned int evt); 263 int (*ioctl)(struct hci_dev *hdev, unsigned int cmd, unsigned long arg); 264 }; 265 266 struct hci_conn { 267 struct list_head list; 268 269 atomic_t refcnt; 270 271 bdaddr_t dst; 272 __u8 dst_type; 273 __u16 handle; 274 __u16 state; 275 __u8 mode; 276 __u8 type; 277 __u8 out; 278 __u8 attempt; 279 __u8 dev_class[3]; 280 __u8 features[8]; 281 __u8 ssp_mode; 282 __u16 interval; 283 __u16 pkt_type; 284 __u16 link_policy; 285 __u32 link_mode; 286 __u8 key_type; 287 __u8 auth_type; 288 __u8 sec_level; 289 __u8 pending_sec_level; 290 __u8 pin_length; 291 __u8 enc_key_size; 292 __u8 io_capability; 293 __u8 power_save; 294 __u16 disc_timeout; 295 unsigned long pend; 296 297 __u8 remote_cap; 298 __u8 remote_oob; 299 __u8 remote_auth; 300 301 unsigned int sent; 302 303 struct sk_buff_head data_q; 304 struct hci_chan_hash chan_hash; 305 306 struct timer_list disc_timer; 307 struct timer_list idle_timer; 308 struct timer_list auto_accept_timer; 309 310 struct work_struct work_add; 311 struct work_struct work_del; 312 313 struct device dev; 314 atomic_t devref; 315 316 struct hci_dev *hdev; 317 void *l2cap_data; 318 void *sco_data; 319 320 struct hci_conn *link; 321 322 void (*connect_cfm_cb) (struct hci_conn *conn, u8 status); 323 void (*security_cfm_cb) (struct hci_conn *conn, u8 status); 324 void (*disconn_cfm_cb) (struct hci_conn *conn, u8 reason); 325 }; 326 327 struct hci_chan { 328 struct list_head list; 329 330 struct hci_conn *conn; 331 struct sk_buff_head data_q; 332 unsigned int sent; 333 }; 334 335 extern struct hci_proto *hci_proto[]; 336 extern struct list_head hci_dev_list; 337 extern struct list_head hci_cb_list; 338 extern rwlock_t hci_dev_list_lock; 339 extern rwlock_t hci_cb_list_lock; 340 341 /* ----- Inquiry cache ----- */ 342 #define INQUIRY_CACHE_AGE_MAX (HZ*30) /* 30 seconds */ 343 #define INQUIRY_ENTRY_AGE_MAX (HZ*60) /* 60 seconds */ 344 345 #define inquiry_cache_lock(c) spin_lock(&c->lock) 346 #define inquiry_cache_unlock(c) spin_unlock(&c->lock) 347 #define inquiry_cache_lock_bh(c) spin_lock_bh(&c->lock) 348 #define inquiry_cache_unlock_bh(c) spin_unlock_bh(&c->lock) 349 350 static inline void inquiry_cache_init(struct hci_dev *hdev) 351 { 352 struct inquiry_cache *c = &hdev->inq_cache; 353 spin_lock_init(&c->lock); 354 c->list = NULL; 355 } 356 357 static inline int inquiry_cache_empty(struct hci_dev *hdev) 358 { 359 struct inquiry_cache *c = &hdev->inq_cache; 360 return c->list == NULL; 361 } 362 363 static inline long inquiry_cache_age(struct hci_dev *hdev) 364 { 365 struct inquiry_cache *c = &hdev->inq_cache; 366 return jiffies - c->timestamp; 367 } 368 369 static inline long inquiry_entry_age(struct inquiry_entry *e) 370 { 371 return jiffies - e->timestamp; 372 } 373 374 struct inquiry_entry *hci_inquiry_cache_lookup(struct hci_dev *hdev, 375 bdaddr_t *bdaddr); 376 void hci_inquiry_cache_update(struct hci_dev *hdev, struct inquiry_data *data); 377 378 /* ----- HCI Connections ----- */ 379 enum { 380 HCI_CONN_AUTH_PEND, 381 HCI_CONN_REAUTH_PEND, 382 HCI_CONN_ENCRYPT_PEND, 383 HCI_CONN_RSWITCH_PEND, 384 HCI_CONN_MODE_CHANGE_PEND, 385 HCI_CONN_SCO_SETUP_PEND, 386 HCI_CONN_LE_SMP_PEND, 387 }; 388 389 static inline void hci_conn_hash_init(struct hci_dev *hdev) 390 { 391 struct hci_conn_hash *h = &hdev->conn_hash; 392 INIT_LIST_HEAD(&h->list); 393 spin_lock_init(&h->lock); 394 h->acl_num = 0; 395 h->sco_num = 0; 396 } 397 398 static inline void hci_conn_hash_add(struct hci_dev *hdev, struct hci_conn *c) 399 { 400 struct hci_conn_hash *h = &hdev->conn_hash; 401 list_add(&c->list, &h->list); 402 switch (c->type) { 403 case ACL_LINK: 404 h->acl_num++; 405 break; 406 case LE_LINK: 407 h->le_num++; 408 break; 409 case SCO_LINK: 410 case ESCO_LINK: 411 h->sco_num++; 412 break; 413 } 414 } 415 416 static inline void hci_conn_hash_del(struct hci_dev *hdev, struct hci_conn *c) 417 { 418 struct hci_conn_hash *h = &hdev->conn_hash; 419 list_del(&c->list); 420 switch (c->type) { 421 case ACL_LINK: 422 h->acl_num--; 423 break; 424 case LE_LINK: 425 h->le_num--; 426 break; 427 case SCO_LINK: 428 case ESCO_LINK: 429 h->sco_num--; 430 break; 431 } 432 } 433 434 static inline unsigned int hci_conn_num(struct hci_dev *hdev, __u8 type) 435 { 436 struct hci_conn_hash *h = &hdev->conn_hash; 437 switch (type) { 438 case ACL_LINK: 439 return h->acl_num; 440 case LE_LINK: 441 return h->le_num; 442 case SCO_LINK: 443 case ESCO_LINK: 444 return h->sco_num; 445 default: 446 return 0; 447 } 448 } 449 450 static inline struct hci_conn *hci_conn_hash_lookup_handle(struct hci_dev *hdev, 451 __u16 handle) 452 { 453 struct hci_conn_hash *h = &hdev->conn_hash; 454 struct list_head *p; 455 struct hci_conn *c; 456 457 list_for_each(p, &h->list) { 458 c = list_entry(p, struct hci_conn, list); 459 if (c->handle == handle) 460 return c; 461 } 462 return NULL; 463 } 464 465 static inline struct hci_conn *hci_conn_hash_lookup_ba(struct hci_dev *hdev, 466 __u8 type, bdaddr_t *ba) 467 { 468 struct hci_conn_hash *h = &hdev->conn_hash; 469 struct list_head *p; 470 struct hci_conn *c; 471 472 list_for_each(p, &h->list) { 473 c = list_entry(p, struct hci_conn, list); 474 if (c->type == type && !bacmp(&c->dst, ba)) 475 return c; 476 } 477 return NULL; 478 } 479 480 static inline struct hci_conn *hci_conn_hash_lookup_state(struct hci_dev *hdev, 481 __u8 type, __u16 state) 482 { 483 struct hci_conn_hash *h = &hdev->conn_hash; 484 struct list_head *p; 485 struct hci_conn *c; 486 487 list_for_each(p, &h->list) { 488 c = list_entry(p, struct hci_conn, list); 489 if (c->type == type && c->state == state) 490 return c; 491 } 492 return NULL; 493 } 494 495 static inline void hci_chan_hash_init(struct hci_conn *c) 496 { 497 struct hci_chan_hash *h = &c->chan_hash; 498 INIT_LIST_HEAD(&h->list); 499 spin_lock_init(&h->lock); 500 h->num = 0; 501 } 502 503 static inline void hci_chan_hash_add(struct hci_conn *c, struct hci_chan *chan) 504 { 505 struct hci_chan_hash *h = &c->chan_hash; 506 list_add(&chan->list, &h->list); 507 h->num++; 508 } 509 510 static inline void hci_chan_hash_del(struct hci_conn *c, struct hci_chan *chan) 511 { 512 struct hci_chan_hash *h = &c->chan_hash; 513 list_del(&chan->list); 514 h->num--; 515 } 516 517 void hci_acl_connect(struct hci_conn *conn); 518 void hci_acl_disconn(struct hci_conn *conn, __u8 reason); 519 void hci_add_sco(struct hci_conn *conn, __u16 handle); 520 void hci_setup_sync(struct hci_conn *conn, __u16 handle); 521 void hci_sco_setup(struct hci_conn *conn, __u8 status); 522 523 struct hci_conn *hci_conn_add(struct hci_dev *hdev, int type, bdaddr_t *dst); 524 int hci_conn_del(struct hci_conn *conn); 525 void hci_conn_hash_flush(struct hci_dev *hdev); 526 void hci_conn_check_pending(struct hci_dev *hdev); 527 528 struct hci_chan *hci_chan_create(struct hci_conn *conn); 529 int hci_chan_del(struct hci_chan *chan); 530 void hci_chan_hash_flush(struct hci_conn *conn); 531 532 struct hci_conn *hci_connect(struct hci_dev *hdev, int type, bdaddr_t *dst, 533 __u8 sec_level, __u8 auth_type); 534 int hci_conn_check_link_mode(struct hci_conn *conn); 535 int hci_conn_check_secure(struct hci_conn *conn, __u8 sec_level); 536 int hci_conn_security(struct hci_conn *conn, __u8 sec_level, __u8 auth_type); 537 int hci_conn_change_link_key(struct hci_conn *conn); 538 int hci_conn_switch_role(struct hci_conn *conn, __u8 role); 539 540 void hci_conn_enter_active_mode(struct hci_conn *conn, __u8 force_active); 541 void hci_conn_enter_sniff_mode(struct hci_conn *conn); 542 543 void hci_conn_hold_device(struct hci_conn *conn); 544 void hci_conn_put_device(struct hci_conn *conn); 545 546 static inline void hci_conn_hold(struct hci_conn *conn) 547 { 548 atomic_inc(&conn->refcnt); 549 del_timer(&conn->disc_timer); 550 } 551 552 static inline void hci_conn_put(struct hci_conn *conn) 553 { 554 if (atomic_dec_and_test(&conn->refcnt)) { 555 unsigned long timeo; 556 if (conn->type == ACL_LINK || conn->type == LE_LINK) { 557 del_timer(&conn->idle_timer); 558 if (conn->state == BT_CONNECTED) { 559 timeo = msecs_to_jiffies(conn->disc_timeout); 560 if (!conn->out) 561 timeo *= 2; 562 } else { 563 timeo = msecs_to_jiffies(10); 564 } 565 } else { 566 timeo = msecs_to_jiffies(10); 567 } 568 mod_timer(&conn->disc_timer, jiffies + timeo); 569 } 570 } 571 572 /* ----- HCI Devices ----- */ 573 static inline void __hci_dev_put(struct hci_dev *d) 574 { 575 if (atomic_dec_and_test(&d->refcnt)) 576 d->destruct(d); 577 } 578 579 /* 580 * hci_dev_put and hci_dev_hold are macros to avoid dragging all the 581 * overhead of all the modular infrastructure into this header. 582 */ 583 #define hci_dev_put(d) \ 584 do { \ 585 __hci_dev_put(d); \ 586 module_put(d->owner); \ 587 } while (0) 588 589 static inline struct hci_dev *__hci_dev_hold(struct hci_dev *d) 590 { 591 atomic_inc(&d->refcnt); 592 return d; 593 } 594 595 #define hci_dev_hold(d) \ 596 ({ \ 597 try_module_get(d->owner) ? __hci_dev_hold(d) : NULL; \ 598 }) 599 600 #define hci_dev_lock(d) spin_lock(&d->lock) 601 #define hci_dev_unlock(d) spin_unlock(&d->lock) 602 #define hci_dev_lock_bh(d) spin_lock_bh(&d->lock) 603 #define hci_dev_unlock_bh(d) spin_unlock_bh(&d->lock) 604 605 struct hci_dev *hci_dev_get(int index); 606 struct hci_dev *hci_get_route(bdaddr_t *src, bdaddr_t *dst); 607 608 struct hci_dev *hci_alloc_dev(void); 609 void hci_free_dev(struct hci_dev *hdev); 610 int hci_register_dev(struct hci_dev *hdev); 611 void hci_unregister_dev(struct hci_dev *hdev); 612 int hci_suspend_dev(struct hci_dev *hdev); 613 int hci_resume_dev(struct hci_dev *hdev); 614 int hci_dev_open(__u16 dev); 615 int hci_dev_close(__u16 dev); 616 int hci_dev_reset(__u16 dev); 617 int hci_dev_reset_stat(__u16 dev); 618 int hci_dev_cmd(unsigned int cmd, void __user *arg); 619 int hci_get_dev_list(void __user *arg); 620 int hci_get_dev_info(void __user *arg); 621 int hci_get_conn_list(void __user *arg); 622 int hci_get_conn_info(struct hci_dev *hdev, void __user *arg); 623 int hci_get_auth_info(struct hci_dev *hdev, void __user *arg); 624 int hci_inquiry(void __user *arg); 625 626 struct bdaddr_list *hci_blacklist_lookup(struct hci_dev *hdev, bdaddr_t *bdaddr); 627 int hci_blacklist_clear(struct hci_dev *hdev); 628 int hci_blacklist_add(struct hci_dev *hdev, bdaddr_t *bdaddr); 629 int hci_blacklist_del(struct hci_dev *hdev, bdaddr_t *bdaddr); 630 631 int hci_uuids_clear(struct hci_dev *hdev); 632 633 int hci_link_keys_clear(struct hci_dev *hdev); 634 struct link_key *hci_find_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr); 635 int hci_add_link_key(struct hci_dev *hdev, struct hci_conn *conn, int new_key, 636 bdaddr_t *bdaddr, u8 *val, u8 type, u8 pin_len); 637 struct link_key *hci_find_ltk(struct hci_dev *hdev, __le16 ediv, u8 rand[8]); 638 struct link_key *hci_find_link_key_type(struct hci_dev *hdev, 639 bdaddr_t *bdaddr, u8 type); 640 int hci_add_ltk(struct hci_dev *hdev, int new_key, bdaddr_t *bdaddr, 641 u8 key_size, __le16 ediv, u8 rand[8], u8 ltk[16]); 642 int hci_remove_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr); 643 644 int hci_remote_oob_data_clear(struct hci_dev *hdev); 645 struct oob_data *hci_find_remote_oob_data(struct hci_dev *hdev, 646 bdaddr_t *bdaddr); 647 int hci_add_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 *hash, 648 u8 *randomizer); 649 int hci_remove_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr); 650 651 #define ADV_CLEAR_TIMEOUT (3*60*HZ) /* Three minutes */ 652 int hci_adv_entries_clear(struct hci_dev *hdev); 653 struct adv_entry *hci_find_adv_entry(struct hci_dev *hdev, bdaddr_t *bdaddr); 654 int hci_add_adv_entry(struct hci_dev *hdev, 655 struct hci_ev_le_advertising_info *ev); 656 657 void hci_del_off_timer(struct hci_dev *hdev); 658 659 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb); 660 661 int hci_recv_frame(struct sk_buff *skb); 662 int hci_recv_fragment(struct hci_dev *hdev, int type, void *data, int count); 663 int hci_recv_stream_fragment(struct hci_dev *hdev, void *data, int count); 664 665 void hci_init_sysfs(struct hci_dev *hdev); 666 int hci_add_sysfs(struct hci_dev *hdev); 667 void hci_del_sysfs(struct hci_dev *hdev); 668 void hci_conn_init_sysfs(struct hci_conn *conn); 669 void hci_conn_add_sysfs(struct hci_conn *conn); 670 void hci_conn_del_sysfs(struct hci_conn *conn); 671 672 #define SET_HCIDEV_DEV(hdev, pdev) ((hdev)->parent = (pdev)) 673 674 /* ----- LMP capabilities ----- */ 675 #define lmp_rswitch_capable(dev) ((dev)->features[0] & LMP_RSWITCH) 676 #define lmp_encrypt_capable(dev) ((dev)->features[0] & LMP_ENCRYPT) 677 #define lmp_sniff_capable(dev) ((dev)->features[0] & LMP_SNIFF) 678 #define lmp_sniffsubr_capable(dev) ((dev)->features[5] & LMP_SNIFF_SUBR) 679 #define lmp_esco_capable(dev) ((dev)->features[3] & LMP_ESCO) 680 #define lmp_ssp_capable(dev) ((dev)->features[6] & LMP_SIMPLE_PAIR) 681 #define lmp_no_flush_capable(dev) ((dev)->features[6] & LMP_NO_FLUSH) 682 #define lmp_le_capable(dev) ((dev)->features[4] & LMP_LE) 683 684 /* ----- Extended LMP capabilities ----- */ 685 #define lmp_host_le_capable(dev) ((dev)->extfeatures[0] & LMP_HOST_LE) 686 687 /* ----- HCI protocols ----- */ 688 struct hci_proto { 689 char *name; 690 unsigned int id; 691 unsigned long flags; 692 693 void *priv; 694 695 int (*connect_ind) (struct hci_dev *hdev, bdaddr_t *bdaddr, 696 __u8 type); 697 int (*connect_cfm) (struct hci_conn *conn, __u8 status); 698 int (*disconn_ind) (struct hci_conn *conn); 699 int (*disconn_cfm) (struct hci_conn *conn, __u8 reason); 700 int (*recv_acldata) (struct hci_conn *conn, struct sk_buff *skb, 701 __u16 flags); 702 int (*recv_scodata) (struct hci_conn *conn, struct sk_buff *skb); 703 int (*security_cfm) (struct hci_conn *conn, __u8 status, 704 __u8 encrypt); 705 }; 706 707 static inline int hci_proto_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr, 708 __u8 type) 709 { 710 register struct hci_proto *hp; 711 int mask = 0; 712 713 hp = hci_proto[HCI_PROTO_L2CAP]; 714 if (hp && hp->connect_ind) 715 mask |= hp->connect_ind(hdev, bdaddr, type); 716 717 hp = hci_proto[HCI_PROTO_SCO]; 718 if (hp && hp->connect_ind) 719 mask |= hp->connect_ind(hdev, bdaddr, type); 720 721 return mask; 722 } 723 724 static inline void hci_proto_connect_cfm(struct hci_conn *conn, __u8 status) 725 { 726 register struct hci_proto *hp; 727 728 hp = hci_proto[HCI_PROTO_L2CAP]; 729 if (hp && hp->connect_cfm) 730 hp->connect_cfm(conn, status); 731 732 hp = hci_proto[HCI_PROTO_SCO]; 733 if (hp && hp->connect_cfm) 734 hp->connect_cfm(conn, status); 735 736 if (conn->connect_cfm_cb) 737 conn->connect_cfm_cb(conn, status); 738 } 739 740 static inline int hci_proto_disconn_ind(struct hci_conn *conn) 741 { 742 register struct hci_proto *hp; 743 int reason = HCI_ERROR_REMOTE_USER_TERM; 744 745 hp = hci_proto[HCI_PROTO_L2CAP]; 746 if (hp && hp->disconn_ind) 747 reason = hp->disconn_ind(conn); 748 749 hp = hci_proto[HCI_PROTO_SCO]; 750 if (hp && hp->disconn_ind) 751 reason = hp->disconn_ind(conn); 752 753 return reason; 754 } 755 756 static inline void hci_proto_disconn_cfm(struct hci_conn *conn, __u8 reason) 757 { 758 register struct hci_proto *hp; 759 760 hp = hci_proto[HCI_PROTO_L2CAP]; 761 if (hp && hp->disconn_cfm) 762 hp->disconn_cfm(conn, reason); 763 764 hp = hci_proto[HCI_PROTO_SCO]; 765 if (hp && hp->disconn_cfm) 766 hp->disconn_cfm(conn, reason); 767 768 if (conn->disconn_cfm_cb) 769 conn->disconn_cfm_cb(conn, reason); 770 } 771 772 static inline void hci_proto_auth_cfm(struct hci_conn *conn, __u8 status) 773 { 774 register struct hci_proto *hp; 775 __u8 encrypt; 776 777 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend)) 778 return; 779 780 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00; 781 782 hp = hci_proto[HCI_PROTO_L2CAP]; 783 if (hp && hp->security_cfm) 784 hp->security_cfm(conn, status, encrypt); 785 786 hp = hci_proto[HCI_PROTO_SCO]; 787 if (hp && hp->security_cfm) 788 hp->security_cfm(conn, status, encrypt); 789 790 if (conn->security_cfm_cb) 791 conn->security_cfm_cb(conn, status); 792 } 793 794 static inline void hci_proto_encrypt_cfm(struct hci_conn *conn, __u8 status, 795 __u8 encrypt) 796 { 797 register struct hci_proto *hp; 798 799 hp = hci_proto[HCI_PROTO_L2CAP]; 800 if (hp && hp->security_cfm) 801 hp->security_cfm(conn, status, encrypt); 802 803 hp = hci_proto[HCI_PROTO_SCO]; 804 if (hp && hp->security_cfm) 805 hp->security_cfm(conn, status, encrypt); 806 807 if (conn->security_cfm_cb) 808 conn->security_cfm_cb(conn, status); 809 } 810 811 int hci_register_proto(struct hci_proto *hproto); 812 int hci_unregister_proto(struct hci_proto *hproto); 813 814 /* ----- HCI callbacks ----- */ 815 struct hci_cb { 816 struct list_head list; 817 818 char *name; 819 820 void (*security_cfm) (struct hci_conn *conn, __u8 status, 821 __u8 encrypt); 822 void (*key_change_cfm) (struct hci_conn *conn, __u8 status); 823 void (*role_switch_cfm) (struct hci_conn *conn, __u8 status, __u8 role); 824 }; 825 826 static inline void hci_auth_cfm(struct hci_conn *conn, __u8 status) 827 { 828 struct list_head *p; 829 __u8 encrypt; 830 831 hci_proto_auth_cfm(conn, status); 832 833 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->pend)) 834 return; 835 836 encrypt = (conn->link_mode & HCI_LM_ENCRYPT) ? 0x01 : 0x00; 837 838 read_lock_bh(&hci_cb_list_lock); 839 list_for_each(p, &hci_cb_list) { 840 struct hci_cb *cb = list_entry(p, struct hci_cb, list); 841 if (cb->security_cfm) 842 cb->security_cfm(conn, status, encrypt); 843 } 844 read_unlock_bh(&hci_cb_list_lock); 845 } 846 847 static inline void hci_encrypt_cfm(struct hci_conn *conn, __u8 status, 848 __u8 encrypt) 849 { 850 struct list_head *p; 851 852 if (conn->sec_level == BT_SECURITY_SDP) 853 conn->sec_level = BT_SECURITY_LOW; 854 855 if (conn->pending_sec_level > conn->sec_level) 856 conn->sec_level = conn->pending_sec_level; 857 858 hci_proto_encrypt_cfm(conn, status, encrypt); 859 860 read_lock_bh(&hci_cb_list_lock); 861 list_for_each(p, &hci_cb_list) { 862 struct hci_cb *cb = list_entry(p, struct hci_cb, list); 863 if (cb->security_cfm) 864 cb->security_cfm(conn, status, encrypt); 865 } 866 read_unlock_bh(&hci_cb_list_lock); 867 } 868 869 static inline void hci_key_change_cfm(struct hci_conn *conn, __u8 status) 870 { 871 struct list_head *p; 872 873 read_lock_bh(&hci_cb_list_lock); 874 list_for_each(p, &hci_cb_list) { 875 struct hci_cb *cb = list_entry(p, struct hci_cb, list); 876 if (cb->key_change_cfm) 877 cb->key_change_cfm(conn, status); 878 } 879 read_unlock_bh(&hci_cb_list_lock); 880 } 881 882 static inline void hci_role_switch_cfm(struct hci_conn *conn, __u8 status, 883 __u8 role) 884 { 885 struct list_head *p; 886 887 read_lock_bh(&hci_cb_list_lock); 888 list_for_each(p, &hci_cb_list) { 889 struct hci_cb *cb = list_entry(p, struct hci_cb, list); 890 if (cb->role_switch_cfm) 891 cb->role_switch_cfm(conn, status, role); 892 } 893 read_unlock_bh(&hci_cb_list_lock); 894 } 895 896 int hci_register_cb(struct hci_cb *hcb); 897 int hci_unregister_cb(struct hci_cb *hcb); 898 899 int hci_register_notifier(struct notifier_block *nb); 900 int hci_unregister_notifier(struct notifier_block *nb); 901 902 int hci_send_cmd(struct hci_dev *hdev, __u16 opcode, __u32 plen, void *param); 903 void hci_send_acl(struct hci_chan *chan, struct sk_buff *skb, __u16 flags); 904 void hci_send_sco(struct hci_conn *conn, struct sk_buff *skb); 905 906 void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 opcode); 907 908 void hci_si_event(struct hci_dev *hdev, int type, int dlen, void *data); 909 910 /* ----- HCI Sockets ----- */ 911 void hci_send_to_sock(struct hci_dev *hdev, struct sk_buff *skb, 912 struct sock *skip_sk); 913 914 /* Management interface */ 915 int mgmt_control(struct sock *sk, struct msghdr *msg, size_t len); 916 int mgmt_index_added(struct hci_dev *hdev); 917 int mgmt_index_removed(struct hci_dev *hdev); 918 int mgmt_powered(struct hci_dev *hdev, u8 powered); 919 int mgmt_discoverable(struct hci_dev *hdev, u8 discoverable); 920 int mgmt_connectable(struct hci_dev *hdev, u8 connectable); 921 int mgmt_write_scan_failed(struct hci_dev *hdev, u8 scan, u8 status); 922 int mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key, 923 u8 persistent); 924 int mgmt_connected(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 925 u8 addr_type); 926 int mgmt_disconnected(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 927 u8 addr_type); 928 int mgmt_disconnect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 status); 929 int mgmt_connect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 930 u8 addr_type, u8 status); 931 int mgmt_pin_code_request(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 secure); 932 int mgmt_pin_code_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 933 u8 status); 934 int mgmt_pin_code_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 935 u8 status); 936 int mgmt_user_confirm_request(struct hci_dev *hdev, bdaddr_t *bdaddr, 937 __le32 value, u8 confirm_hint); 938 int mgmt_user_confirm_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 939 u8 status); 940 int mgmt_user_confirm_neg_reply_complete(struct hci_dev *hdev, 941 bdaddr_t *bdaddr, u8 status); 942 int mgmt_user_passkey_request(struct hci_dev *hdev, bdaddr_t *bdaddr); 943 int mgmt_user_passkey_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 944 u8 status); 945 int mgmt_user_passkey_neg_reply_complete(struct hci_dev *hdev, 946 bdaddr_t *bdaddr, u8 status); 947 int mgmt_auth_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 status); 948 int mgmt_set_local_name_complete(struct hci_dev *hdev, u8 *name, u8 status); 949 int mgmt_read_local_oob_data_reply_complete(struct hci_dev *hdev, u8 *hash, 950 u8 *randomizer, u8 status); 951 int mgmt_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 952 u8 addr_type, u8 *dev_class, s8 rssi, u8 *eir); 953 int mgmt_remote_name(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 *name); 954 int mgmt_start_discovery_failed(struct hci_dev *hdev, u8 status); 955 int mgmt_stop_discovery_failed(struct hci_dev *hdev, u8 status); 956 int mgmt_discovering(struct hci_dev *hdev, u8 discovering); 957 int mgmt_device_blocked(struct hci_dev *hdev, bdaddr_t *bdaddr); 958 int mgmt_device_unblocked(struct hci_dev *hdev, bdaddr_t *bdaddr); 959 960 /* HCI info for socket */ 961 #define hci_pi(sk) ((struct hci_pinfo *) sk) 962 963 struct hci_pinfo { 964 struct bt_sock bt; 965 struct hci_dev *hdev; 966 struct hci_filter filter; 967 __u32 cmsg_mask; 968 unsigned short channel; 969 }; 970 971 /* HCI security filter */ 972 #define HCI_SFLT_MAX_OGF 5 973 974 struct hci_sec_filter { 975 __u32 type_mask; 976 __u32 event_mask[2]; 977 __u32 ocf_mask[HCI_SFLT_MAX_OGF + 1][4]; 978 }; 979 980 /* ----- HCI requests ----- */ 981 #define HCI_REQ_DONE 0 982 #define HCI_REQ_PEND 1 983 #define HCI_REQ_CANCELED 2 984 985 #define hci_req_lock(d) mutex_lock(&d->req_lock) 986 #define hci_req_unlock(d) mutex_unlock(&d->req_lock) 987 988 void hci_req_complete(struct hci_dev *hdev, __u16 cmd, int result); 989 990 void hci_le_conn_update(struct hci_conn *conn, u16 min, u16 max, 991 u16 latency, u16 to_multiplier); 992 void hci_le_start_enc(struct hci_conn *conn, __le16 ediv, __u8 rand[8], 993 __u8 ltk[16]); 994 void hci_le_ltk_reply(struct hci_conn *conn, u8 ltk[16]); 995 void hci_le_ltk_neg_reply(struct hci_conn *conn); 996 997 int hci_do_inquiry(struct hci_dev *hdev, u8 length); 998 int hci_cancel_inquiry(struct hci_dev *hdev); 999 1000 #endif /* __HCI_CORE_H */ 1001