xref: /linux/include/net/bluetooth/hci_core.h (revision b2ddeb11738464ce8f75c15384a3b8132cb80357)
1 /*
2    BlueZ - Bluetooth protocol stack for Linux
3    Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved.
4 
5    Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6 
7    This program is free software; you can redistribute it and/or modify
8    it under the terms of the GNU General Public License version 2 as
9    published by the Free Software Foundation;
10 
11    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12    OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14    IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15    CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16    WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17    ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18    OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19 
20    ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21    COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22    SOFTWARE IS DISCLAIMED.
23 */
24 
25 #ifndef __HCI_CORE_H
26 #define __HCI_CORE_H
27 
28 #include <net/bluetooth/hci.h>
29 #include <net/bluetooth/hci_sock.h>
30 
31 /* HCI priority */
32 #define HCI_PRIO_MAX	7
33 
34 /* HCI Core structures */
35 struct inquiry_data {
36 	bdaddr_t	bdaddr;
37 	__u8		pscan_rep_mode;
38 	__u8		pscan_period_mode;
39 	__u8		pscan_mode;
40 	__u8		dev_class[3];
41 	__le16		clock_offset;
42 	__s8		rssi;
43 	__u8		ssp_mode;
44 };
45 
46 struct inquiry_entry {
47 	struct list_head	all;		/* inq_cache.all */
48 	struct list_head	list;		/* unknown or resolve */
49 	enum {
50 		NAME_NOT_KNOWN,
51 		NAME_NEEDED,
52 		NAME_PENDING,
53 		NAME_KNOWN,
54 	} name_state;
55 	__u32			timestamp;
56 	struct inquiry_data	data;
57 };
58 
59 struct discovery_state {
60 	int			type;
61 	enum {
62 		DISCOVERY_STOPPED,
63 		DISCOVERY_STARTING,
64 		DISCOVERY_FINDING,
65 		DISCOVERY_RESOLVING,
66 		DISCOVERY_STOPPING,
67 	} state;
68 	struct list_head	all;	/* All devices found during inquiry */
69 	struct list_head	unknown;	/* Name state not known */
70 	struct list_head	resolve;	/* Name needs to be resolved */
71 	__u32			timestamp;
72 	bdaddr_t		last_adv_addr;
73 	u8			last_adv_addr_type;
74 	s8			last_adv_rssi;
75 	u32			last_adv_flags;
76 	u8			last_adv_data[HCI_MAX_AD_LENGTH];
77 	u8			last_adv_data_len;
78 	bool			report_invalid_rssi;
79 	bool			result_filtering;
80 	s8			rssi;
81 	u16			uuid_count;
82 	u8			(*uuids)[16];
83 	unsigned long		scan_start;
84 	unsigned long		scan_duration;
85 };
86 
87 struct hci_conn_hash {
88 	struct list_head list;
89 	unsigned int     acl_num;
90 	unsigned int     amp_num;
91 	unsigned int     sco_num;
92 	unsigned int     le_num;
93 	unsigned int     le_num_slave;
94 };
95 
96 struct bdaddr_list {
97 	struct list_head list;
98 	bdaddr_t bdaddr;
99 	u8 bdaddr_type;
100 };
101 
102 struct bt_uuid {
103 	struct list_head list;
104 	u8 uuid[16];
105 	u8 size;
106 	u8 svc_hint;
107 };
108 
109 struct smp_csrk {
110 	bdaddr_t bdaddr;
111 	u8 bdaddr_type;
112 	u8 type;
113 	u8 val[16];
114 };
115 
116 struct smp_ltk {
117 	struct list_head list;
118 	struct rcu_head rcu;
119 	bdaddr_t bdaddr;
120 	u8 bdaddr_type;
121 	u8 authenticated;
122 	u8 type;
123 	u8 enc_size;
124 	__le16 ediv;
125 	__le64 rand;
126 	u8 val[16];
127 };
128 
129 struct smp_irk {
130 	struct list_head list;
131 	struct rcu_head rcu;
132 	bdaddr_t rpa;
133 	bdaddr_t bdaddr;
134 	u8 addr_type;
135 	u8 val[16];
136 };
137 
138 struct link_key {
139 	struct list_head list;
140 	struct rcu_head rcu;
141 	bdaddr_t bdaddr;
142 	u8 type;
143 	u8 val[HCI_LINK_KEY_SIZE];
144 	u8 pin_len;
145 };
146 
147 struct oob_data {
148 	struct list_head list;
149 	bdaddr_t bdaddr;
150 	u8 bdaddr_type;
151 	u8 present;
152 	u8 hash192[16];
153 	u8 rand192[16];
154 	u8 hash256[16];
155 	u8 rand256[16];
156 };
157 
158 struct adv_info {
159 	struct delayed_work timeout_exp;
160 	__u8	instance;
161 	__u32	flags;
162 	__u16	timeout;
163 	__u16	adv_data_len;
164 	__u8	adv_data[HCI_MAX_AD_LENGTH];
165 	__u16	scan_rsp_len;
166 	__u8	scan_rsp_data[HCI_MAX_AD_LENGTH];
167 };
168 
169 #define HCI_MAX_SHORT_NAME_LENGTH	10
170 
171 /* Default LE RPA expiry time, 15 minutes */
172 #define HCI_DEFAULT_RPA_TIMEOUT		(15 * 60)
173 
174 /* Default min/max age of connection information (1s/3s) */
175 #define DEFAULT_CONN_INFO_MIN_AGE	1000
176 #define DEFAULT_CONN_INFO_MAX_AGE	3000
177 
178 struct amp_assoc {
179 	__u16	len;
180 	__u16	offset;
181 	__u16	rem_len;
182 	__u16	len_so_far;
183 	__u8	data[HCI_MAX_AMP_ASSOC_SIZE];
184 };
185 
186 #define HCI_MAX_PAGES	3
187 
188 #define NUM_REASSEMBLY 4
189 struct hci_dev {
190 	struct list_head list;
191 	struct mutex	lock;
192 
193 	char		name[8];
194 	unsigned long	flags;
195 	__u16		id;
196 	__u8		bus;
197 	__u8		dev_type;
198 	bdaddr_t	bdaddr;
199 	bdaddr_t	setup_addr;
200 	bdaddr_t	public_addr;
201 	bdaddr_t	random_addr;
202 	bdaddr_t	static_addr;
203 	__u8		adv_addr_type;
204 	__u8		dev_name[HCI_MAX_NAME_LENGTH];
205 	__u8		short_name[HCI_MAX_SHORT_NAME_LENGTH];
206 	__u8		eir[HCI_MAX_EIR_LENGTH];
207 	__u8		dev_class[3];
208 	__u8		major_class;
209 	__u8		minor_class;
210 	__u8		max_page;
211 	__u8		features[HCI_MAX_PAGES][8];
212 	__u8		le_features[8];
213 	__u8		le_white_list_size;
214 	__u8		le_states[8];
215 	__u8		commands[64];
216 	__u8		hci_ver;
217 	__u16		hci_rev;
218 	__u8		lmp_ver;
219 	__u16		manufacturer;
220 	__u16		lmp_subver;
221 	__u16		voice_setting;
222 	__u8		num_iac;
223 	__u8		stored_max_keys;
224 	__u8		stored_num_keys;
225 	__u8		io_capability;
226 	__s8		inq_tx_power;
227 	__u16		page_scan_interval;
228 	__u16		page_scan_window;
229 	__u8		page_scan_type;
230 	__u8		le_adv_channel_map;
231 	__u16		le_adv_min_interval;
232 	__u16		le_adv_max_interval;
233 	__u8		le_scan_type;
234 	__u16		le_scan_interval;
235 	__u16		le_scan_window;
236 	__u16		le_conn_min_interval;
237 	__u16		le_conn_max_interval;
238 	__u16		le_conn_latency;
239 	__u16		le_supv_timeout;
240 	__u16		le_def_tx_len;
241 	__u16		le_def_tx_time;
242 	__u16		le_max_tx_len;
243 	__u16		le_max_tx_time;
244 	__u16		le_max_rx_len;
245 	__u16		le_max_rx_time;
246 	__u16		discov_interleaved_timeout;
247 	__u16		conn_info_min_age;
248 	__u16		conn_info_max_age;
249 	__u8		ssp_debug_mode;
250 	__u8		hw_error_code;
251 	__u32		clock;
252 
253 	__u16		devid_source;
254 	__u16		devid_vendor;
255 	__u16		devid_product;
256 	__u16		devid_version;
257 
258 	__u16		pkt_type;
259 	__u16		esco_type;
260 	__u16		link_policy;
261 	__u16		link_mode;
262 
263 	__u32		idle_timeout;
264 	__u16		sniff_min_interval;
265 	__u16		sniff_max_interval;
266 
267 	__u8		amp_status;
268 	__u32		amp_total_bw;
269 	__u32		amp_max_bw;
270 	__u32		amp_min_latency;
271 	__u32		amp_max_pdu;
272 	__u8		amp_type;
273 	__u16		amp_pal_cap;
274 	__u16		amp_assoc_size;
275 	__u32		amp_max_flush_to;
276 	__u32		amp_be_flush_to;
277 
278 	struct amp_assoc	loc_assoc;
279 
280 	__u8		flow_ctl_mode;
281 
282 	unsigned int	auto_accept_delay;
283 
284 	unsigned long	quirks;
285 
286 	atomic_t	cmd_cnt;
287 	unsigned int	acl_cnt;
288 	unsigned int	sco_cnt;
289 	unsigned int	le_cnt;
290 
291 	unsigned int	acl_mtu;
292 	unsigned int	sco_mtu;
293 	unsigned int	le_mtu;
294 	unsigned int	acl_pkts;
295 	unsigned int	sco_pkts;
296 	unsigned int	le_pkts;
297 
298 	__u16		block_len;
299 	__u16		block_mtu;
300 	__u16		num_blocks;
301 	__u16		block_cnt;
302 
303 	unsigned long	acl_last_tx;
304 	unsigned long	sco_last_tx;
305 	unsigned long	le_last_tx;
306 
307 	struct workqueue_struct	*workqueue;
308 	struct workqueue_struct	*req_workqueue;
309 
310 	struct work_struct	power_on;
311 	struct delayed_work	power_off;
312 	struct work_struct	error_reset;
313 
314 	__u16			discov_timeout;
315 	struct delayed_work	discov_off;
316 
317 	struct delayed_work	service_cache;
318 
319 	struct delayed_work	cmd_timer;
320 
321 	struct work_struct	rx_work;
322 	struct work_struct	cmd_work;
323 	struct work_struct	tx_work;
324 
325 	struct sk_buff_head	rx_q;
326 	struct sk_buff_head	raw_q;
327 	struct sk_buff_head	cmd_q;
328 
329 	struct sk_buff		*sent_cmd;
330 	struct sk_buff		*reassembly[NUM_REASSEMBLY];
331 
332 	struct mutex		req_lock;
333 	wait_queue_head_t	req_wait_q;
334 	__u32			req_status;
335 	__u32			req_result;
336 	struct sk_buff		*req_skb;
337 
338 	void			*smp_data;
339 	void			*smp_bredr_data;
340 
341 	struct discovery_state	discovery;
342 	struct hci_conn_hash	conn_hash;
343 
344 	struct list_head	mgmt_pending;
345 	struct list_head	blacklist;
346 	struct list_head	whitelist;
347 	struct list_head	uuids;
348 	struct list_head	link_keys;
349 	struct list_head	long_term_keys;
350 	struct list_head	identity_resolving_keys;
351 	struct list_head	remote_oob_data;
352 	struct list_head	le_white_list;
353 	struct list_head	le_conn_params;
354 	struct list_head	pend_le_conns;
355 	struct list_head	pend_le_reports;
356 
357 	struct hci_dev_stats	stat;
358 
359 	atomic_t		promisc;
360 
361 	struct dentry		*debugfs;
362 
363 	struct device		dev;
364 
365 	struct rfkill		*rfkill;
366 
367 	DECLARE_BITMAP(dev_flags, __HCI_NUM_FLAGS);
368 
369 	struct delayed_work	le_scan_disable;
370 	struct delayed_work	le_scan_restart;
371 
372 	__s8			adv_tx_power;
373 	__u8			adv_data[HCI_MAX_AD_LENGTH];
374 	__u8			adv_data_len;
375 	__u8			scan_rsp_data[HCI_MAX_AD_LENGTH];
376 	__u8			scan_rsp_data_len;
377 
378 	struct adv_info		adv_instance;
379 
380 	__u8			irk[16];
381 	__u32			rpa_timeout;
382 	struct delayed_work	rpa_expired;
383 	bdaddr_t		rpa;
384 
385 	int (*open)(struct hci_dev *hdev);
386 	int (*close)(struct hci_dev *hdev);
387 	int (*flush)(struct hci_dev *hdev);
388 	int (*setup)(struct hci_dev *hdev);
389 	int (*shutdown)(struct hci_dev *hdev);
390 	int (*send)(struct hci_dev *hdev, struct sk_buff *skb);
391 	void (*notify)(struct hci_dev *hdev, unsigned int evt);
392 	void (*hw_error)(struct hci_dev *hdev, u8 code);
393 	int (*set_bdaddr)(struct hci_dev *hdev, const bdaddr_t *bdaddr);
394 };
395 
396 #define HCI_PHY_HANDLE(handle)	(handle & 0xff)
397 
398 struct hci_conn {
399 	struct list_head list;
400 
401 	atomic_t	refcnt;
402 
403 	bdaddr_t	dst;
404 	__u8		dst_type;
405 	bdaddr_t	src;
406 	__u8		src_type;
407 	bdaddr_t	init_addr;
408 	__u8		init_addr_type;
409 	bdaddr_t	resp_addr;
410 	__u8		resp_addr_type;
411 	__u16		handle;
412 	__u16		state;
413 	__u8		mode;
414 	__u8		type;
415 	__u8		role;
416 	bool		out;
417 	__u8		attempt;
418 	__u8		dev_class[3];
419 	__u8		features[HCI_MAX_PAGES][8];
420 	__u16		pkt_type;
421 	__u16		link_policy;
422 	__u8		key_type;
423 	__u8		auth_type;
424 	__u8		sec_level;
425 	__u8		pending_sec_level;
426 	__u8		pin_length;
427 	__u8		enc_key_size;
428 	__u8		io_capability;
429 	__u32		passkey_notify;
430 	__u8		passkey_entered;
431 	__u16		disc_timeout;
432 	__u16		conn_timeout;
433 	__u16		setting;
434 	__u16		le_conn_min_interval;
435 	__u16		le_conn_max_interval;
436 	__u16		le_conn_interval;
437 	__u16		le_conn_latency;
438 	__u16		le_supv_timeout;
439 	__u8		le_adv_data[HCI_MAX_AD_LENGTH];
440 	__u8		le_adv_data_len;
441 	__s8		rssi;
442 	__s8		tx_power;
443 	__s8		max_tx_power;
444 	unsigned long	flags;
445 
446 	__u32		clock;
447 	__u16		clock_accuracy;
448 
449 	unsigned long	conn_info_timestamp;
450 
451 	__u8		remote_cap;
452 	__u8		remote_auth;
453 	__u8		remote_id;
454 
455 	unsigned int	sent;
456 
457 	struct sk_buff_head data_q;
458 	struct list_head chan_list;
459 
460 	struct delayed_work disc_work;
461 	struct delayed_work auto_accept_work;
462 	struct delayed_work idle_work;
463 	struct delayed_work le_conn_timeout;
464 
465 	struct device	dev;
466 	struct dentry	*debugfs;
467 
468 	struct hci_dev	*hdev;
469 	void		*l2cap_data;
470 	void		*sco_data;
471 	struct amp_mgr	*amp_mgr;
472 
473 	struct hci_conn	*link;
474 
475 	void (*connect_cfm_cb)	(struct hci_conn *conn, u8 status);
476 	void (*security_cfm_cb)	(struct hci_conn *conn, u8 status);
477 	void (*disconn_cfm_cb)	(struct hci_conn *conn, u8 reason);
478 };
479 
480 struct hci_chan {
481 	struct list_head list;
482 	__u16 handle;
483 	struct hci_conn *conn;
484 	struct sk_buff_head data_q;
485 	unsigned int	sent;
486 	__u8		state;
487 };
488 
489 struct hci_conn_params {
490 	struct list_head list;
491 	struct list_head action;
492 
493 	bdaddr_t addr;
494 	u8 addr_type;
495 
496 	u16 conn_min_interval;
497 	u16 conn_max_interval;
498 	u16 conn_latency;
499 	u16 supervision_timeout;
500 
501 	enum {
502 		HCI_AUTO_CONN_DISABLED,
503 		HCI_AUTO_CONN_REPORT,
504 		HCI_AUTO_CONN_DIRECT,
505 		HCI_AUTO_CONN_ALWAYS,
506 		HCI_AUTO_CONN_LINK_LOSS,
507 	} auto_connect;
508 
509 	struct hci_conn *conn;
510 };
511 
512 extern struct list_head hci_dev_list;
513 extern struct list_head hci_cb_list;
514 extern rwlock_t hci_dev_list_lock;
515 extern struct mutex hci_cb_list_lock;
516 
517 #define hci_dev_set_flag(hdev, nr)             set_bit((nr), (hdev)->dev_flags)
518 #define hci_dev_clear_flag(hdev, nr)           clear_bit((nr), (hdev)->dev_flags)
519 #define hci_dev_change_flag(hdev, nr)          change_bit((nr), (hdev)->dev_flags)
520 #define hci_dev_test_flag(hdev, nr)            test_bit((nr), (hdev)->dev_flags)
521 #define hci_dev_test_and_set_flag(hdev, nr)    test_and_set_bit((nr), (hdev)->dev_flags)
522 #define hci_dev_test_and_clear_flag(hdev, nr)  test_and_clear_bit((nr), (hdev)->dev_flags)
523 #define hci_dev_test_and_change_flag(hdev, nr) test_and_change_bit((nr), (hdev)->dev_flags)
524 
525 #define hci_dev_clear_volatile_flags(hdev)			\
526 	do {							\
527 		hci_dev_clear_flag(hdev, HCI_LE_SCAN);		\
528 		hci_dev_clear_flag(hdev, HCI_LE_ADV);		\
529 		hci_dev_clear_flag(hdev, HCI_PERIODIC_INQ);	\
530 	} while (0)
531 
532 /* ----- HCI interface to upper protocols ----- */
533 int l2cap_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr);
534 int l2cap_disconn_ind(struct hci_conn *hcon);
535 int l2cap_recv_acldata(struct hci_conn *hcon, struct sk_buff *skb, u16 flags);
536 
537 int sco_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr, __u8 *flags);
538 int sco_recv_scodata(struct hci_conn *hcon, struct sk_buff *skb);
539 
540 /* ----- Inquiry cache ----- */
541 #define INQUIRY_CACHE_AGE_MAX   (HZ*30)   /* 30 seconds */
542 #define INQUIRY_ENTRY_AGE_MAX   (HZ*60)   /* 60 seconds */
543 
544 static inline void discovery_init(struct hci_dev *hdev)
545 {
546 	hdev->discovery.state = DISCOVERY_STOPPED;
547 	INIT_LIST_HEAD(&hdev->discovery.all);
548 	INIT_LIST_HEAD(&hdev->discovery.unknown);
549 	INIT_LIST_HEAD(&hdev->discovery.resolve);
550 	hdev->discovery.report_invalid_rssi = true;
551 	hdev->discovery.rssi = HCI_RSSI_INVALID;
552 }
553 
554 static inline void hci_discovery_filter_clear(struct hci_dev *hdev)
555 {
556 	hdev->discovery.result_filtering = false;
557 	hdev->discovery.report_invalid_rssi = true;
558 	hdev->discovery.rssi = HCI_RSSI_INVALID;
559 	hdev->discovery.uuid_count = 0;
560 	kfree(hdev->discovery.uuids);
561 	hdev->discovery.uuids = NULL;
562 	hdev->discovery.scan_start = 0;
563 	hdev->discovery.scan_duration = 0;
564 }
565 
566 static inline void adv_info_init(struct hci_dev *hdev)
567 {
568 	memset(&hdev->adv_instance, 0, sizeof(struct adv_info));
569 }
570 
571 bool hci_discovery_active(struct hci_dev *hdev);
572 
573 void hci_discovery_set_state(struct hci_dev *hdev, int state);
574 
575 static inline int inquiry_cache_empty(struct hci_dev *hdev)
576 {
577 	return list_empty(&hdev->discovery.all);
578 }
579 
580 static inline long inquiry_cache_age(struct hci_dev *hdev)
581 {
582 	struct discovery_state *c = &hdev->discovery;
583 	return jiffies - c->timestamp;
584 }
585 
586 static inline long inquiry_entry_age(struct inquiry_entry *e)
587 {
588 	return jiffies - e->timestamp;
589 }
590 
591 struct inquiry_entry *hci_inquiry_cache_lookup(struct hci_dev *hdev,
592 					       bdaddr_t *bdaddr);
593 struct inquiry_entry *hci_inquiry_cache_lookup_unknown(struct hci_dev *hdev,
594 						       bdaddr_t *bdaddr);
595 struct inquiry_entry *hci_inquiry_cache_lookup_resolve(struct hci_dev *hdev,
596 						       bdaddr_t *bdaddr,
597 						       int state);
598 void hci_inquiry_cache_update_resolve(struct hci_dev *hdev,
599 				      struct inquiry_entry *ie);
600 u32 hci_inquiry_cache_update(struct hci_dev *hdev, struct inquiry_data *data,
601 			     bool name_known);
602 void hci_inquiry_cache_flush(struct hci_dev *hdev);
603 
604 /* ----- HCI Connections ----- */
605 enum {
606 	HCI_CONN_AUTH_PEND,
607 	HCI_CONN_REAUTH_PEND,
608 	HCI_CONN_ENCRYPT_PEND,
609 	HCI_CONN_RSWITCH_PEND,
610 	HCI_CONN_MODE_CHANGE_PEND,
611 	HCI_CONN_SCO_SETUP_PEND,
612 	HCI_CONN_MGMT_CONNECTED,
613 	HCI_CONN_SSP_ENABLED,
614 	HCI_CONN_SC_ENABLED,
615 	HCI_CONN_AES_CCM,
616 	HCI_CONN_POWER_SAVE,
617 	HCI_CONN_FLUSH_KEY,
618 	HCI_CONN_ENCRYPT,
619 	HCI_CONN_AUTH,
620 	HCI_CONN_SECURE,
621 	HCI_CONN_FIPS,
622 	HCI_CONN_STK_ENCRYPT,
623 	HCI_CONN_AUTH_INITIATOR,
624 	HCI_CONN_DROP,
625 	HCI_CONN_PARAM_REMOVAL_PEND,
626 	HCI_CONN_NEW_LINK_KEY,
627 };
628 
629 static inline bool hci_conn_ssp_enabled(struct hci_conn *conn)
630 {
631 	struct hci_dev *hdev = conn->hdev;
632 	return hci_dev_test_flag(hdev, HCI_SSP_ENABLED) &&
633 	       test_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
634 }
635 
636 static inline bool hci_conn_sc_enabled(struct hci_conn *conn)
637 {
638 	struct hci_dev *hdev = conn->hdev;
639 	return hci_dev_test_flag(hdev, HCI_SC_ENABLED) &&
640 	       test_bit(HCI_CONN_SC_ENABLED, &conn->flags);
641 }
642 
643 static inline void hci_conn_hash_add(struct hci_dev *hdev, struct hci_conn *c)
644 {
645 	struct hci_conn_hash *h = &hdev->conn_hash;
646 	list_add_rcu(&c->list, &h->list);
647 	switch (c->type) {
648 	case ACL_LINK:
649 		h->acl_num++;
650 		break;
651 	case AMP_LINK:
652 		h->amp_num++;
653 		break;
654 	case LE_LINK:
655 		h->le_num++;
656 		if (c->role == HCI_ROLE_SLAVE)
657 			h->le_num_slave++;
658 		break;
659 	case SCO_LINK:
660 	case ESCO_LINK:
661 		h->sco_num++;
662 		break;
663 	}
664 }
665 
666 static inline void hci_conn_hash_del(struct hci_dev *hdev, struct hci_conn *c)
667 {
668 	struct hci_conn_hash *h = &hdev->conn_hash;
669 
670 	list_del_rcu(&c->list);
671 	synchronize_rcu();
672 
673 	switch (c->type) {
674 	case ACL_LINK:
675 		h->acl_num--;
676 		break;
677 	case AMP_LINK:
678 		h->amp_num--;
679 		break;
680 	case LE_LINK:
681 		h->le_num--;
682 		if (c->role == HCI_ROLE_SLAVE)
683 			h->le_num_slave--;
684 		break;
685 	case SCO_LINK:
686 	case ESCO_LINK:
687 		h->sco_num--;
688 		break;
689 	}
690 }
691 
692 static inline unsigned int hci_conn_num(struct hci_dev *hdev, __u8 type)
693 {
694 	struct hci_conn_hash *h = &hdev->conn_hash;
695 	switch (type) {
696 	case ACL_LINK:
697 		return h->acl_num;
698 	case AMP_LINK:
699 		return h->amp_num;
700 	case LE_LINK:
701 		return h->le_num;
702 	case SCO_LINK:
703 	case ESCO_LINK:
704 		return h->sco_num;
705 	default:
706 		return 0;
707 	}
708 }
709 
710 static inline unsigned int hci_conn_count(struct hci_dev *hdev)
711 {
712 	struct hci_conn_hash *c = &hdev->conn_hash;
713 
714 	return c->acl_num + c->amp_num + c->sco_num + c->le_num;
715 }
716 
717 static inline __u8 hci_conn_lookup_type(struct hci_dev *hdev, __u16 handle)
718 {
719 	struct hci_conn_hash *h = &hdev->conn_hash;
720 	struct hci_conn *c;
721 	__u8 type = INVALID_LINK;
722 
723 	rcu_read_lock();
724 
725 	list_for_each_entry_rcu(c, &h->list, list) {
726 		if (c->handle == handle) {
727 			type = c->type;
728 			break;
729 		}
730 	}
731 
732 	rcu_read_unlock();
733 
734 	return type;
735 }
736 
737 static inline struct hci_conn *hci_conn_hash_lookup_handle(struct hci_dev *hdev,
738 								__u16 handle)
739 {
740 	struct hci_conn_hash *h = &hdev->conn_hash;
741 	struct hci_conn  *c;
742 
743 	rcu_read_lock();
744 
745 	list_for_each_entry_rcu(c, &h->list, list) {
746 		if (c->handle == handle) {
747 			rcu_read_unlock();
748 			return c;
749 		}
750 	}
751 	rcu_read_unlock();
752 
753 	return NULL;
754 }
755 
756 static inline struct hci_conn *hci_conn_hash_lookup_ba(struct hci_dev *hdev,
757 							__u8 type, bdaddr_t *ba)
758 {
759 	struct hci_conn_hash *h = &hdev->conn_hash;
760 	struct hci_conn  *c;
761 
762 	rcu_read_lock();
763 
764 	list_for_each_entry_rcu(c, &h->list, list) {
765 		if (c->type == type && !bacmp(&c->dst, ba)) {
766 			rcu_read_unlock();
767 			return c;
768 		}
769 	}
770 
771 	rcu_read_unlock();
772 
773 	return NULL;
774 }
775 
776 static inline struct hci_conn *hci_conn_hash_lookup_state(struct hci_dev *hdev,
777 							__u8 type, __u16 state)
778 {
779 	struct hci_conn_hash *h = &hdev->conn_hash;
780 	struct hci_conn  *c;
781 
782 	rcu_read_lock();
783 
784 	list_for_each_entry_rcu(c, &h->list, list) {
785 		if (c->type == type && c->state == state) {
786 			rcu_read_unlock();
787 			return c;
788 		}
789 	}
790 
791 	rcu_read_unlock();
792 
793 	return NULL;
794 }
795 
796 int hci_disconnect(struct hci_conn *conn, __u8 reason);
797 bool hci_setup_sync(struct hci_conn *conn, __u16 handle);
798 void hci_sco_setup(struct hci_conn *conn, __u8 status);
799 
800 struct hci_conn *hci_conn_add(struct hci_dev *hdev, int type, bdaddr_t *dst,
801 			      u8 role);
802 int hci_conn_del(struct hci_conn *conn);
803 void hci_conn_hash_flush(struct hci_dev *hdev);
804 void hci_conn_check_pending(struct hci_dev *hdev);
805 
806 struct hci_chan *hci_chan_create(struct hci_conn *conn);
807 void hci_chan_del(struct hci_chan *chan);
808 void hci_chan_list_flush(struct hci_conn *conn);
809 struct hci_chan *hci_chan_lookup_handle(struct hci_dev *hdev, __u16 handle);
810 
811 struct hci_conn *hci_connect_le(struct hci_dev *hdev, bdaddr_t *dst,
812 				u8 dst_type, u8 sec_level, u16 conn_timeout,
813 				u8 role);
814 struct hci_conn *hci_connect_acl(struct hci_dev *hdev, bdaddr_t *dst,
815 				 u8 sec_level, u8 auth_type);
816 struct hci_conn *hci_connect_sco(struct hci_dev *hdev, int type, bdaddr_t *dst,
817 				 __u16 setting);
818 int hci_conn_check_link_mode(struct hci_conn *conn);
819 int hci_conn_check_secure(struct hci_conn *conn, __u8 sec_level);
820 int hci_conn_security(struct hci_conn *conn, __u8 sec_level, __u8 auth_type,
821 		      bool initiator);
822 int hci_conn_switch_role(struct hci_conn *conn, __u8 role);
823 
824 void hci_conn_enter_active_mode(struct hci_conn *conn, __u8 force_active);
825 
826 void hci_le_conn_failed(struct hci_conn *conn, u8 status);
827 
828 /*
829  * hci_conn_get() and hci_conn_put() are used to control the life-time of an
830  * "hci_conn" object. They do not guarantee that the hci_conn object is running,
831  * working or anything else. They just guarantee that the object is available
832  * and can be dereferenced. So you can use its locks, local variables and any
833  * other constant data.
834  * Before accessing runtime data, you _must_ lock the object and then check that
835  * it is still running. As soon as you release the locks, the connection might
836  * get dropped, though.
837  *
838  * On the other hand, hci_conn_hold() and hci_conn_drop() are used to control
839  * how long the underlying connection is held. So every channel that runs on the
840  * hci_conn object calls this to prevent the connection from disappearing. As
841  * long as you hold a device, you must also guarantee that you have a valid
842  * reference to the device via hci_conn_get() (or the initial reference from
843  * hci_conn_add()).
844  * The hold()/drop() ref-count is known to drop below 0 sometimes, which doesn't
845  * break because nobody cares for that. But this means, we cannot use
846  * _get()/_drop() in it, but require the caller to have a valid ref (FIXME).
847  */
848 
849 static inline struct hci_conn *hci_conn_get(struct hci_conn *conn)
850 {
851 	get_device(&conn->dev);
852 	return conn;
853 }
854 
855 static inline void hci_conn_put(struct hci_conn *conn)
856 {
857 	put_device(&conn->dev);
858 }
859 
860 static inline void hci_conn_hold(struct hci_conn *conn)
861 {
862 	BT_DBG("hcon %p orig refcnt %d", conn, atomic_read(&conn->refcnt));
863 
864 	atomic_inc(&conn->refcnt);
865 	cancel_delayed_work(&conn->disc_work);
866 }
867 
868 static inline void hci_conn_drop(struct hci_conn *conn)
869 {
870 	BT_DBG("hcon %p orig refcnt %d", conn, atomic_read(&conn->refcnt));
871 
872 	if (atomic_dec_and_test(&conn->refcnt)) {
873 		unsigned long timeo;
874 
875 		switch (conn->type) {
876 		case ACL_LINK:
877 		case LE_LINK:
878 			cancel_delayed_work(&conn->idle_work);
879 			if (conn->state == BT_CONNECTED) {
880 				timeo = conn->disc_timeout;
881 				if (!conn->out)
882 					timeo *= 2;
883 			} else {
884 				timeo = 0;
885 			}
886 			break;
887 
888 		case AMP_LINK:
889 			timeo = conn->disc_timeout;
890 			break;
891 
892 		default:
893 			timeo = 0;
894 			break;
895 		}
896 
897 		cancel_delayed_work(&conn->disc_work);
898 		queue_delayed_work(conn->hdev->workqueue,
899 				   &conn->disc_work, timeo);
900 	}
901 }
902 
903 /* ----- HCI Devices ----- */
904 static inline void hci_dev_put(struct hci_dev *d)
905 {
906 	BT_DBG("%s orig refcnt %d", d->name,
907 	       atomic_read(&d->dev.kobj.kref.refcount));
908 
909 	put_device(&d->dev);
910 }
911 
912 static inline struct hci_dev *hci_dev_hold(struct hci_dev *d)
913 {
914 	BT_DBG("%s orig refcnt %d", d->name,
915 	       atomic_read(&d->dev.kobj.kref.refcount));
916 
917 	get_device(&d->dev);
918 	return d;
919 }
920 
921 #define hci_dev_lock(d)		mutex_lock(&d->lock)
922 #define hci_dev_unlock(d)	mutex_unlock(&d->lock)
923 
924 #define to_hci_dev(d) container_of(d, struct hci_dev, dev)
925 #define to_hci_conn(c) container_of(c, struct hci_conn, dev)
926 
927 static inline void *hci_get_drvdata(struct hci_dev *hdev)
928 {
929 	return dev_get_drvdata(&hdev->dev);
930 }
931 
932 static inline void hci_set_drvdata(struct hci_dev *hdev, void *data)
933 {
934 	dev_set_drvdata(&hdev->dev, data);
935 }
936 
937 struct hci_dev *hci_dev_get(int index);
938 struct hci_dev *hci_get_route(bdaddr_t *dst, bdaddr_t *src);
939 
940 struct hci_dev *hci_alloc_dev(void);
941 void hci_free_dev(struct hci_dev *hdev);
942 int hci_register_dev(struct hci_dev *hdev);
943 void hci_unregister_dev(struct hci_dev *hdev);
944 int hci_suspend_dev(struct hci_dev *hdev);
945 int hci_resume_dev(struct hci_dev *hdev);
946 int hci_reset_dev(struct hci_dev *hdev);
947 int hci_dev_open(__u16 dev);
948 int hci_dev_close(__u16 dev);
949 int hci_dev_reset(__u16 dev);
950 int hci_dev_reset_stat(__u16 dev);
951 int hci_dev_cmd(unsigned int cmd, void __user *arg);
952 int hci_get_dev_list(void __user *arg);
953 int hci_get_dev_info(void __user *arg);
954 int hci_get_conn_list(void __user *arg);
955 int hci_get_conn_info(struct hci_dev *hdev, void __user *arg);
956 int hci_get_auth_info(struct hci_dev *hdev, void __user *arg);
957 int hci_inquiry(void __user *arg);
958 
959 struct bdaddr_list *hci_bdaddr_list_lookup(struct list_head *list,
960 					   bdaddr_t *bdaddr, u8 type);
961 int hci_bdaddr_list_add(struct list_head *list, bdaddr_t *bdaddr, u8 type);
962 int hci_bdaddr_list_del(struct list_head *list, bdaddr_t *bdaddr, u8 type);
963 void hci_bdaddr_list_clear(struct list_head *list);
964 
965 struct hci_conn_params *hci_conn_params_lookup(struct hci_dev *hdev,
966 					       bdaddr_t *addr, u8 addr_type);
967 struct hci_conn_params *hci_conn_params_add(struct hci_dev *hdev,
968 					    bdaddr_t *addr, u8 addr_type);
969 void hci_conn_params_del(struct hci_dev *hdev, bdaddr_t *addr, u8 addr_type);
970 void hci_conn_params_clear_all(struct hci_dev *hdev);
971 void hci_conn_params_clear_disabled(struct hci_dev *hdev);
972 
973 struct hci_conn_params *hci_pend_le_action_lookup(struct list_head *list,
974 						  bdaddr_t *addr,
975 						  u8 addr_type);
976 
977 void hci_uuids_clear(struct hci_dev *hdev);
978 
979 void hci_link_keys_clear(struct hci_dev *hdev);
980 struct link_key *hci_find_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr);
981 struct link_key *hci_add_link_key(struct hci_dev *hdev, struct hci_conn *conn,
982 				  bdaddr_t *bdaddr, u8 *val, u8 type,
983 				  u8 pin_len, bool *persistent);
984 struct smp_ltk *hci_add_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr,
985 			    u8 addr_type, u8 type, u8 authenticated,
986 			    u8 tk[16], u8 enc_size, __le16 ediv, __le64 rand);
987 struct smp_ltk *hci_find_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr,
988 			     u8 addr_type, u8 role);
989 int hci_remove_ltk(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 bdaddr_type);
990 void hci_smp_ltks_clear(struct hci_dev *hdev);
991 int hci_remove_link_key(struct hci_dev *hdev, bdaddr_t *bdaddr);
992 
993 struct smp_irk *hci_find_irk_by_rpa(struct hci_dev *hdev, bdaddr_t *rpa);
994 struct smp_irk *hci_find_irk_by_addr(struct hci_dev *hdev, bdaddr_t *bdaddr,
995 				     u8 addr_type);
996 struct smp_irk *hci_add_irk(struct hci_dev *hdev, bdaddr_t *bdaddr,
997 			    u8 addr_type, u8 val[16], bdaddr_t *rpa);
998 void hci_remove_irk(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 addr_type);
999 void hci_smp_irks_clear(struct hci_dev *hdev);
1000 
1001 bool hci_bdaddr_is_paired(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 type);
1002 
1003 void hci_remote_oob_data_clear(struct hci_dev *hdev);
1004 struct oob_data *hci_find_remote_oob_data(struct hci_dev *hdev,
1005 					  bdaddr_t *bdaddr, u8 bdaddr_type);
1006 int hci_add_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr,
1007 			    u8 bdaddr_type, u8 *hash192, u8 *rand192,
1008 			    u8 *hash256, u8 *rand256);
1009 int hci_remove_remote_oob_data(struct hci_dev *hdev, bdaddr_t *bdaddr,
1010 			       u8 bdaddr_type);
1011 
1012 void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb);
1013 
1014 int hci_recv_frame(struct hci_dev *hdev, struct sk_buff *skb);
1015 int hci_recv_stream_fragment(struct hci_dev *hdev, void *data, int count);
1016 
1017 void hci_init_sysfs(struct hci_dev *hdev);
1018 void hci_conn_init_sysfs(struct hci_conn *conn);
1019 void hci_conn_add_sysfs(struct hci_conn *conn);
1020 void hci_conn_del_sysfs(struct hci_conn *conn);
1021 
1022 #define SET_HCIDEV_DEV(hdev, pdev) ((hdev)->dev.parent = (pdev))
1023 
1024 /* ----- LMP capabilities ----- */
1025 #define lmp_encrypt_capable(dev)   ((dev)->features[0][0] & LMP_ENCRYPT)
1026 #define lmp_rswitch_capable(dev)   ((dev)->features[0][0] & LMP_RSWITCH)
1027 #define lmp_hold_capable(dev)      ((dev)->features[0][0] & LMP_HOLD)
1028 #define lmp_sniff_capable(dev)     ((dev)->features[0][0] & LMP_SNIFF)
1029 #define lmp_park_capable(dev)      ((dev)->features[0][1] & LMP_PARK)
1030 #define lmp_inq_rssi_capable(dev)  ((dev)->features[0][3] & LMP_RSSI_INQ)
1031 #define lmp_esco_capable(dev)      ((dev)->features[0][3] & LMP_ESCO)
1032 #define lmp_bredr_capable(dev)     (!((dev)->features[0][4] & LMP_NO_BREDR))
1033 #define lmp_le_capable(dev)        ((dev)->features[0][4] & LMP_LE)
1034 #define lmp_sniffsubr_capable(dev) ((dev)->features[0][5] & LMP_SNIFF_SUBR)
1035 #define lmp_pause_enc_capable(dev) ((dev)->features[0][5] & LMP_PAUSE_ENC)
1036 #define lmp_ext_inq_capable(dev)   ((dev)->features[0][6] & LMP_EXT_INQ)
1037 #define lmp_le_br_capable(dev)     (!!((dev)->features[0][6] & LMP_SIMUL_LE_BR))
1038 #define lmp_ssp_capable(dev)       ((dev)->features[0][6] & LMP_SIMPLE_PAIR)
1039 #define lmp_no_flush_capable(dev)  ((dev)->features[0][6] & LMP_NO_FLUSH)
1040 #define lmp_lsto_capable(dev)      ((dev)->features[0][7] & LMP_LSTO)
1041 #define lmp_inq_tx_pwr_capable(dev) ((dev)->features[0][7] & LMP_INQ_TX_PWR)
1042 #define lmp_ext_feat_capable(dev)  ((dev)->features[0][7] & LMP_EXTFEATURES)
1043 #define lmp_transp_capable(dev)    ((dev)->features[0][2] & LMP_TRANSPARENT)
1044 
1045 /* ----- Extended LMP capabilities ----- */
1046 #define lmp_csb_master_capable(dev) ((dev)->features[2][0] & LMP_CSB_MASTER)
1047 #define lmp_csb_slave_capable(dev)  ((dev)->features[2][0] & LMP_CSB_SLAVE)
1048 #define lmp_sync_train_capable(dev) ((dev)->features[2][0] & LMP_SYNC_TRAIN)
1049 #define lmp_sync_scan_capable(dev)  ((dev)->features[2][0] & LMP_SYNC_SCAN)
1050 #define lmp_sc_capable(dev)         ((dev)->features[2][1] & LMP_SC)
1051 #define lmp_ping_capable(dev)       ((dev)->features[2][1] & LMP_PING)
1052 
1053 /* ----- Host capabilities ----- */
1054 #define lmp_host_ssp_capable(dev)  ((dev)->features[1][0] & LMP_HOST_SSP)
1055 #define lmp_host_sc_capable(dev)   ((dev)->features[1][0] & LMP_HOST_SC)
1056 #define lmp_host_le_capable(dev)   (!!((dev)->features[1][0] & LMP_HOST_LE))
1057 #define lmp_host_le_br_capable(dev) (!!((dev)->features[1][0] & LMP_HOST_LE_BREDR))
1058 
1059 #define hdev_is_powered(dev)   (test_bit(HCI_UP, &(dev)->flags) && \
1060 				!hci_dev_test_flag(dev, HCI_AUTO_OFF))
1061 #define bredr_sc_enabled(dev)  (lmp_sc_capable(dev) && \
1062 				hci_dev_test_flag(dev, HCI_SC_ENABLED))
1063 
1064 /* ----- HCI protocols ----- */
1065 #define HCI_PROTO_DEFER             0x01
1066 
1067 static inline int hci_proto_connect_ind(struct hci_dev *hdev, bdaddr_t *bdaddr,
1068 					__u8 type, __u8 *flags)
1069 {
1070 	switch (type) {
1071 	case ACL_LINK:
1072 		return l2cap_connect_ind(hdev, bdaddr);
1073 
1074 	case SCO_LINK:
1075 	case ESCO_LINK:
1076 		return sco_connect_ind(hdev, bdaddr, flags);
1077 
1078 	default:
1079 		BT_ERR("unknown link type %d", type);
1080 		return -EINVAL;
1081 	}
1082 }
1083 
1084 static inline int hci_proto_disconn_ind(struct hci_conn *conn)
1085 {
1086 	if (conn->type != ACL_LINK && conn->type != LE_LINK)
1087 		return HCI_ERROR_REMOTE_USER_TERM;
1088 
1089 	return l2cap_disconn_ind(conn);
1090 }
1091 
1092 /* ----- HCI callbacks ----- */
1093 struct hci_cb {
1094 	struct list_head list;
1095 
1096 	char *name;
1097 
1098 	void (*connect_cfm)	(struct hci_conn *conn, __u8 status);
1099 	void (*disconn_cfm)	(struct hci_conn *conn, __u8 status);
1100 	void (*security_cfm)	(struct hci_conn *conn, __u8 status,
1101 								__u8 encrypt);
1102 	void (*key_change_cfm)	(struct hci_conn *conn, __u8 status);
1103 	void (*role_switch_cfm)	(struct hci_conn *conn, __u8 status, __u8 role);
1104 };
1105 
1106 static inline void hci_connect_cfm(struct hci_conn *conn, __u8 status)
1107 {
1108 	struct hci_cb *cb;
1109 
1110 	mutex_lock(&hci_cb_list_lock);
1111 	list_for_each_entry(cb, &hci_cb_list, list) {
1112 		if (cb->connect_cfm)
1113 			cb->connect_cfm(conn, status);
1114 	}
1115 	mutex_unlock(&hci_cb_list_lock);
1116 
1117 	if (conn->connect_cfm_cb)
1118 		conn->connect_cfm_cb(conn, status);
1119 }
1120 
1121 static inline void hci_disconn_cfm(struct hci_conn *conn, __u8 reason)
1122 {
1123 	struct hci_cb *cb;
1124 
1125 	mutex_lock(&hci_cb_list_lock);
1126 	list_for_each_entry(cb, &hci_cb_list, list) {
1127 		if (cb->disconn_cfm)
1128 			cb->disconn_cfm(conn, reason);
1129 	}
1130 	mutex_unlock(&hci_cb_list_lock);
1131 
1132 	if (conn->disconn_cfm_cb)
1133 		conn->disconn_cfm_cb(conn, reason);
1134 }
1135 
1136 static inline void hci_auth_cfm(struct hci_conn *conn, __u8 status)
1137 {
1138 	struct hci_cb *cb;
1139 	__u8 encrypt;
1140 
1141 	if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags))
1142 		return;
1143 
1144 	encrypt = test_bit(HCI_CONN_ENCRYPT, &conn->flags) ? 0x01 : 0x00;
1145 
1146 	mutex_lock(&hci_cb_list_lock);
1147 	list_for_each_entry(cb, &hci_cb_list, list) {
1148 		if (cb->security_cfm)
1149 			cb->security_cfm(conn, status, encrypt);
1150 	}
1151 	mutex_unlock(&hci_cb_list_lock);
1152 
1153 	if (conn->security_cfm_cb)
1154 		conn->security_cfm_cb(conn, status);
1155 }
1156 
1157 static inline void hci_encrypt_cfm(struct hci_conn *conn, __u8 status,
1158 								__u8 encrypt)
1159 {
1160 	struct hci_cb *cb;
1161 
1162 	if (conn->sec_level == BT_SECURITY_SDP)
1163 		conn->sec_level = BT_SECURITY_LOW;
1164 
1165 	if (conn->pending_sec_level > conn->sec_level)
1166 		conn->sec_level = conn->pending_sec_level;
1167 
1168 	mutex_lock(&hci_cb_list_lock);
1169 	list_for_each_entry(cb, &hci_cb_list, list) {
1170 		if (cb->security_cfm)
1171 			cb->security_cfm(conn, status, encrypt);
1172 	}
1173 	mutex_unlock(&hci_cb_list_lock);
1174 
1175 	if (conn->security_cfm_cb)
1176 		conn->security_cfm_cb(conn, status);
1177 }
1178 
1179 static inline void hci_key_change_cfm(struct hci_conn *conn, __u8 status)
1180 {
1181 	struct hci_cb *cb;
1182 
1183 	mutex_lock(&hci_cb_list_lock);
1184 	list_for_each_entry(cb, &hci_cb_list, list) {
1185 		if (cb->key_change_cfm)
1186 			cb->key_change_cfm(conn, status);
1187 	}
1188 	mutex_unlock(&hci_cb_list_lock);
1189 }
1190 
1191 static inline void hci_role_switch_cfm(struct hci_conn *conn, __u8 status,
1192 								__u8 role)
1193 {
1194 	struct hci_cb *cb;
1195 
1196 	mutex_lock(&hci_cb_list_lock);
1197 	list_for_each_entry(cb, &hci_cb_list, list) {
1198 		if (cb->role_switch_cfm)
1199 			cb->role_switch_cfm(conn, status, role);
1200 	}
1201 	mutex_unlock(&hci_cb_list_lock);
1202 }
1203 
1204 static inline bool eir_has_data_type(u8 *data, size_t data_len, u8 type)
1205 {
1206 	size_t parsed = 0;
1207 
1208 	if (data_len < 2)
1209 		return false;
1210 
1211 	while (parsed < data_len - 1) {
1212 		u8 field_len = data[0];
1213 
1214 		if (field_len == 0)
1215 			break;
1216 
1217 		parsed += field_len + 1;
1218 
1219 		if (parsed > data_len)
1220 			break;
1221 
1222 		if (data[1] == type)
1223 			return true;
1224 
1225 		data += field_len + 1;
1226 	}
1227 
1228 	return false;
1229 }
1230 
1231 static inline bool hci_bdaddr_is_rpa(bdaddr_t *bdaddr, u8 addr_type)
1232 {
1233 	if (addr_type != ADDR_LE_DEV_RANDOM)
1234 		return false;
1235 
1236 	if ((bdaddr->b[5] & 0xc0) == 0x40)
1237 	       return true;
1238 
1239 	return false;
1240 }
1241 
1242 static inline bool hci_is_identity_address(bdaddr_t *addr, u8 addr_type)
1243 {
1244 	if (addr_type == ADDR_LE_DEV_PUBLIC)
1245 		return true;
1246 
1247 	/* Check for Random Static address type */
1248 	if ((addr->b[5] & 0xc0) == 0xc0)
1249 		return true;
1250 
1251 	return false;
1252 }
1253 
1254 static inline struct smp_irk *hci_get_irk(struct hci_dev *hdev,
1255 					  bdaddr_t *bdaddr, u8 addr_type)
1256 {
1257 	if (!hci_bdaddr_is_rpa(bdaddr, addr_type))
1258 		return NULL;
1259 
1260 	return hci_find_irk_by_rpa(hdev, bdaddr);
1261 }
1262 
1263 static inline int hci_check_conn_params(u16 min, u16 max, u16 latency,
1264 					u16 to_multiplier)
1265 {
1266 	u16 max_latency;
1267 
1268 	if (min > max || min < 6 || max > 3200)
1269 		return -EINVAL;
1270 
1271 	if (to_multiplier < 10 || to_multiplier > 3200)
1272 		return -EINVAL;
1273 
1274 	if (max >= to_multiplier * 8)
1275 		return -EINVAL;
1276 
1277 	max_latency = (to_multiplier * 8 / max) - 1;
1278 	if (latency > 499 || latency > max_latency)
1279 		return -EINVAL;
1280 
1281 	return 0;
1282 }
1283 
1284 int hci_register_cb(struct hci_cb *hcb);
1285 int hci_unregister_cb(struct hci_cb *hcb);
1286 
1287 struct sk_buff *__hci_cmd_sync(struct hci_dev *hdev, u16 opcode, u32 plen,
1288 			       const void *param, u32 timeout);
1289 struct sk_buff *__hci_cmd_sync_ev(struct hci_dev *hdev, u16 opcode, u32 plen,
1290 				  const void *param, u8 event, u32 timeout);
1291 
1292 int hci_send_cmd(struct hci_dev *hdev, __u16 opcode, __u32 plen,
1293 		 const void *param);
1294 void hci_send_acl(struct hci_chan *chan, struct sk_buff *skb, __u16 flags);
1295 void hci_send_sco(struct hci_conn *conn, struct sk_buff *skb);
1296 
1297 void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 opcode);
1298 
1299 /* ----- HCI Sockets ----- */
1300 void hci_send_to_sock(struct hci_dev *hdev, struct sk_buff *skb);
1301 void hci_send_to_channel(unsigned short channel, struct sk_buff *skb,
1302 			 int flag, struct sock *skip_sk);
1303 void hci_send_to_monitor(struct hci_dev *hdev, struct sk_buff *skb);
1304 
1305 void hci_sock_dev_event(struct hci_dev *hdev, int event);
1306 
1307 #define HCI_MGMT_VAR_LEN	BIT(0)
1308 #define HCI_MGMT_NO_HDEV	BIT(1)
1309 #define HCI_MGMT_UNTRUSTED	BIT(2)
1310 #define HCI_MGMT_UNCONFIGURED	BIT(3)
1311 
1312 struct hci_mgmt_handler {
1313 	int (*func) (struct sock *sk, struct hci_dev *hdev, void *data,
1314 		     u16 data_len);
1315 	size_t data_len;
1316 	unsigned long flags;
1317 };
1318 
1319 struct hci_mgmt_chan {
1320 	struct list_head list;
1321 	unsigned short channel;
1322 	size_t handler_count;
1323 	const struct hci_mgmt_handler *handlers;
1324 	void (*hdev_init) (struct sock *sk, struct hci_dev *hdev);
1325 };
1326 
1327 int hci_mgmt_chan_register(struct hci_mgmt_chan *c);
1328 void hci_mgmt_chan_unregister(struct hci_mgmt_chan *c);
1329 
1330 /* Management interface */
1331 #define DISCOV_TYPE_BREDR		(BIT(BDADDR_BREDR))
1332 #define DISCOV_TYPE_LE			(BIT(BDADDR_LE_PUBLIC) | \
1333 					 BIT(BDADDR_LE_RANDOM))
1334 #define DISCOV_TYPE_INTERLEAVED		(BIT(BDADDR_BREDR) | \
1335 					 BIT(BDADDR_LE_PUBLIC) | \
1336 					 BIT(BDADDR_LE_RANDOM))
1337 
1338 /* These LE scan and inquiry parameters were chosen according to LE General
1339  * Discovery Procedure specification.
1340  */
1341 #define DISCOV_LE_SCAN_WIN		0x12
1342 #define DISCOV_LE_SCAN_INT		0x12
1343 #define DISCOV_LE_TIMEOUT		10240	/* msec */
1344 #define DISCOV_INTERLEAVED_TIMEOUT	5120	/* msec */
1345 #define DISCOV_INTERLEAVED_INQUIRY_LEN	0x04
1346 #define DISCOV_BREDR_INQUIRY_LEN	0x08
1347 #define DISCOV_LE_RESTART_DELAY		msecs_to_jiffies(200)	/* msec */
1348 
1349 int mgmt_new_settings(struct hci_dev *hdev);
1350 void mgmt_index_added(struct hci_dev *hdev);
1351 void mgmt_index_removed(struct hci_dev *hdev);
1352 void mgmt_set_powered_failed(struct hci_dev *hdev, int err);
1353 int mgmt_powered(struct hci_dev *hdev, u8 powered);
1354 int mgmt_update_adv_data(struct hci_dev *hdev);
1355 void mgmt_discoverable_timeout(struct hci_dev *hdev);
1356 void mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key,
1357 		       bool persistent);
1358 void mgmt_device_connected(struct hci_dev *hdev, struct hci_conn *conn,
1359 			   u32 flags, u8 *name, u8 name_len);
1360 void mgmt_device_disconnected(struct hci_dev *hdev, bdaddr_t *bdaddr,
1361 			      u8 link_type, u8 addr_type, u8 reason,
1362 			      bool mgmt_connected);
1363 void mgmt_disconnect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr,
1364 			    u8 link_type, u8 addr_type, u8 status);
1365 void mgmt_connect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type,
1366 			 u8 addr_type, u8 status);
1367 void mgmt_pin_code_request(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 secure);
1368 void mgmt_pin_code_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
1369 				  u8 status);
1370 void mgmt_pin_code_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
1371 				      u8 status);
1372 int mgmt_user_confirm_request(struct hci_dev *hdev, bdaddr_t *bdaddr,
1373 			      u8 link_type, u8 addr_type, u32 value,
1374 			      u8 confirm_hint);
1375 int mgmt_user_confirm_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
1376 				     u8 link_type, u8 addr_type, u8 status);
1377 int mgmt_user_confirm_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
1378 					 u8 link_type, u8 addr_type, u8 status);
1379 int mgmt_user_passkey_request(struct hci_dev *hdev, bdaddr_t *bdaddr,
1380 			      u8 link_type, u8 addr_type);
1381 int mgmt_user_passkey_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
1382 				     u8 link_type, u8 addr_type, u8 status);
1383 int mgmt_user_passkey_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
1384 					 u8 link_type, u8 addr_type, u8 status);
1385 int mgmt_user_passkey_notify(struct hci_dev *hdev, bdaddr_t *bdaddr,
1386 			     u8 link_type, u8 addr_type, u32 passkey,
1387 			     u8 entered);
1388 void mgmt_auth_failed(struct hci_conn *conn, u8 status);
1389 void mgmt_auth_enable_complete(struct hci_dev *hdev, u8 status);
1390 void mgmt_ssp_enable_complete(struct hci_dev *hdev, u8 enable, u8 status);
1391 void mgmt_set_class_of_dev_complete(struct hci_dev *hdev, u8 *dev_class,
1392 				    u8 status);
1393 void mgmt_set_local_name_complete(struct hci_dev *hdev, u8 *name, u8 status);
1394 void mgmt_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type,
1395 		       u8 addr_type, u8 *dev_class, s8 rssi, u32 flags,
1396 		       u8 *eir, u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len);
1397 void mgmt_remote_name(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type,
1398 		      u8 addr_type, s8 rssi, u8 *name, u8 name_len);
1399 void mgmt_discovering(struct hci_dev *hdev, u8 discovering);
1400 bool mgmt_powering_down(struct hci_dev *hdev);
1401 void mgmt_new_ltk(struct hci_dev *hdev, struct smp_ltk *key, bool persistent);
1402 void mgmt_new_irk(struct hci_dev *hdev, struct smp_irk *irk);
1403 void mgmt_new_csrk(struct hci_dev *hdev, struct smp_csrk *csrk,
1404 		   bool persistent);
1405 void mgmt_new_conn_param(struct hci_dev *hdev, bdaddr_t *bdaddr,
1406 			 u8 bdaddr_type, u8 store_hint, u16 min_interval,
1407 			 u16 max_interval, u16 latency, u16 timeout);
1408 void mgmt_reenable_advertising(struct hci_dev *hdev);
1409 void mgmt_smp_complete(struct hci_conn *conn, bool complete);
1410 
1411 u8 hci_le_conn_update(struct hci_conn *conn, u16 min, u16 max, u16 latency,
1412 		      u16 to_multiplier);
1413 void hci_le_start_enc(struct hci_conn *conn, __le16 ediv, __le64 rand,
1414 							__u8 ltk[16]);
1415 
1416 void hci_copy_identity_address(struct hci_dev *hdev, bdaddr_t *bdaddr,
1417 			       u8 *bdaddr_type);
1418 
1419 #define SCO_AIRMODE_MASK       0x0003
1420 #define SCO_AIRMODE_CVSD       0x0000
1421 #define SCO_AIRMODE_TRANSP     0x0003
1422 
1423 #endif /* __HCI_CORE_H */
1424