1 /* 2 BlueZ - Bluetooth protocol stack for Linux 3 4 Copyright (C) 2010 Nokia Corporation 5 Copyright (C) 2011-2012 Intel Corporation 6 7 This program is free software; you can redistribute it and/or modify 8 it under the terms of the GNU General Public License version 2 as 9 published by the Free Software Foundation; 10 11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS 12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS. 14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY 15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES 16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 19 20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS, 21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS 22 SOFTWARE IS DISCLAIMED. 23 */ 24 25 /* Bluetooth HCI Management interface */ 26 27 #include <linux/module.h> 28 #include <asm/unaligned.h> 29 30 #include <net/bluetooth/bluetooth.h> 31 #include <net/bluetooth/hci_core.h> 32 #include <net/bluetooth/hci_sock.h> 33 #include <net/bluetooth/l2cap.h> 34 #include <net/bluetooth/mgmt.h> 35 36 #include "hci_request.h" 37 #include "smp.h" 38 #include "mgmt_util.h" 39 #include "mgmt_config.h" 40 #include "msft.h" 41 #include "eir.h" 42 #include "aosp.h" 43 44 #define MGMT_VERSION 1 45 #define MGMT_REVISION 22 46 47 static const u16 mgmt_commands[] = { 48 MGMT_OP_READ_INDEX_LIST, 49 MGMT_OP_READ_INFO, 50 MGMT_OP_SET_POWERED, 51 MGMT_OP_SET_DISCOVERABLE, 52 MGMT_OP_SET_CONNECTABLE, 53 MGMT_OP_SET_FAST_CONNECTABLE, 54 MGMT_OP_SET_BONDABLE, 55 MGMT_OP_SET_LINK_SECURITY, 56 MGMT_OP_SET_SSP, 57 MGMT_OP_SET_HS, 58 MGMT_OP_SET_LE, 59 MGMT_OP_SET_DEV_CLASS, 60 MGMT_OP_SET_LOCAL_NAME, 61 MGMT_OP_ADD_UUID, 62 MGMT_OP_REMOVE_UUID, 63 MGMT_OP_LOAD_LINK_KEYS, 64 MGMT_OP_LOAD_LONG_TERM_KEYS, 65 MGMT_OP_DISCONNECT, 66 MGMT_OP_GET_CONNECTIONS, 67 MGMT_OP_PIN_CODE_REPLY, 68 MGMT_OP_PIN_CODE_NEG_REPLY, 69 MGMT_OP_SET_IO_CAPABILITY, 70 MGMT_OP_PAIR_DEVICE, 71 MGMT_OP_CANCEL_PAIR_DEVICE, 72 MGMT_OP_UNPAIR_DEVICE, 73 MGMT_OP_USER_CONFIRM_REPLY, 74 MGMT_OP_USER_CONFIRM_NEG_REPLY, 75 MGMT_OP_USER_PASSKEY_REPLY, 76 MGMT_OP_USER_PASSKEY_NEG_REPLY, 77 MGMT_OP_READ_LOCAL_OOB_DATA, 78 MGMT_OP_ADD_REMOTE_OOB_DATA, 79 MGMT_OP_REMOVE_REMOTE_OOB_DATA, 80 MGMT_OP_START_DISCOVERY, 81 MGMT_OP_STOP_DISCOVERY, 82 MGMT_OP_CONFIRM_NAME, 83 MGMT_OP_BLOCK_DEVICE, 84 MGMT_OP_UNBLOCK_DEVICE, 85 MGMT_OP_SET_DEVICE_ID, 86 MGMT_OP_SET_ADVERTISING, 87 MGMT_OP_SET_BREDR, 88 MGMT_OP_SET_STATIC_ADDRESS, 89 MGMT_OP_SET_SCAN_PARAMS, 90 MGMT_OP_SET_SECURE_CONN, 91 MGMT_OP_SET_DEBUG_KEYS, 92 MGMT_OP_SET_PRIVACY, 93 MGMT_OP_LOAD_IRKS, 94 MGMT_OP_GET_CONN_INFO, 95 MGMT_OP_GET_CLOCK_INFO, 96 MGMT_OP_ADD_DEVICE, 97 MGMT_OP_REMOVE_DEVICE, 98 MGMT_OP_LOAD_CONN_PARAM, 99 MGMT_OP_READ_UNCONF_INDEX_LIST, 100 MGMT_OP_READ_CONFIG_INFO, 101 MGMT_OP_SET_EXTERNAL_CONFIG, 102 MGMT_OP_SET_PUBLIC_ADDRESS, 103 MGMT_OP_START_SERVICE_DISCOVERY, 104 MGMT_OP_READ_LOCAL_OOB_EXT_DATA, 105 MGMT_OP_READ_EXT_INDEX_LIST, 106 MGMT_OP_READ_ADV_FEATURES, 107 MGMT_OP_ADD_ADVERTISING, 108 MGMT_OP_REMOVE_ADVERTISING, 109 MGMT_OP_GET_ADV_SIZE_INFO, 110 MGMT_OP_START_LIMITED_DISCOVERY, 111 MGMT_OP_READ_EXT_INFO, 112 MGMT_OP_SET_APPEARANCE, 113 MGMT_OP_GET_PHY_CONFIGURATION, 114 MGMT_OP_SET_PHY_CONFIGURATION, 115 MGMT_OP_SET_BLOCKED_KEYS, 116 MGMT_OP_SET_WIDEBAND_SPEECH, 117 MGMT_OP_READ_CONTROLLER_CAP, 118 MGMT_OP_READ_EXP_FEATURES_INFO, 119 MGMT_OP_SET_EXP_FEATURE, 120 MGMT_OP_READ_DEF_SYSTEM_CONFIG, 121 MGMT_OP_SET_DEF_SYSTEM_CONFIG, 122 MGMT_OP_READ_DEF_RUNTIME_CONFIG, 123 MGMT_OP_SET_DEF_RUNTIME_CONFIG, 124 MGMT_OP_GET_DEVICE_FLAGS, 125 MGMT_OP_SET_DEVICE_FLAGS, 126 MGMT_OP_READ_ADV_MONITOR_FEATURES, 127 MGMT_OP_ADD_ADV_PATTERNS_MONITOR, 128 MGMT_OP_REMOVE_ADV_MONITOR, 129 MGMT_OP_ADD_EXT_ADV_PARAMS, 130 MGMT_OP_ADD_EXT_ADV_DATA, 131 MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, 132 MGMT_OP_SET_MESH_RECEIVER, 133 MGMT_OP_MESH_READ_FEATURES, 134 MGMT_OP_MESH_SEND, 135 MGMT_OP_MESH_SEND_CANCEL, 136 }; 137 138 static const u16 mgmt_events[] = { 139 MGMT_EV_CONTROLLER_ERROR, 140 MGMT_EV_INDEX_ADDED, 141 MGMT_EV_INDEX_REMOVED, 142 MGMT_EV_NEW_SETTINGS, 143 MGMT_EV_CLASS_OF_DEV_CHANGED, 144 MGMT_EV_LOCAL_NAME_CHANGED, 145 MGMT_EV_NEW_LINK_KEY, 146 MGMT_EV_NEW_LONG_TERM_KEY, 147 MGMT_EV_DEVICE_CONNECTED, 148 MGMT_EV_DEVICE_DISCONNECTED, 149 MGMT_EV_CONNECT_FAILED, 150 MGMT_EV_PIN_CODE_REQUEST, 151 MGMT_EV_USER_CONFIRM_REQUEST, 152 MGMT_EV_USER_PASSKEY_REQUEST, 153 MGMT_EV_AUTH_FAILED, 154 MGMT_EV_DEVICE_FOUND, 155 MGMT_EV_DISCOVERING, 156 MGMT_EV_DEVICE_BLOCKED, 157 MGMT_EV_DEVICE_UNBLOCKED, 158 MGMT_EV_DEVICE_UNPAIRED, 159 MGMT_EV_PASSKEY_NOTIFY, 160 MGMT_EV_NEW_IRK, 161 MGMT_EV_NEW_CSRK, 162 MGMT_EV_DEVICE_ADDED, 163 MGMT_EV_DEVICE_REMOVED, 164 MGMT_EV_NEW_CONN_PARAM, 165 MGMT_EV_UNCONF_INDEX_ADDED, 166 MGMT_EV_UNCONF_INDEX_REMOVED, 167 MGMT_EV_NEW_CONFIG_OPTIONS, 168 MGMT_EV_EXT_INDEX_ADDED, 169 MGMT_EV_EXT_INDEX_REMOVED, 170 MGMT_EV_LOCAL_OOB_DATA_UPDATED, 171 MGMT_EV_ADVERTISING_ADDED, 172 MGMT_EV_ADVERTISING_REMOVED, 173 MGMT_EV_EXT_INFO_CHANGED, 174 MGMT_EV_PHY_CONFIGURATION_CHANGED, 175 MGMT_EV_EXP_FEATURE_CHANGED, 176 MGMT_EV_DEVICE_FLAGS_CHANGED, 177 MGMT_EV_ADV_MONITOR_ADDED, 178 MGMT_EV_ADV_MONITOR_REMOVED, 179 MGMT_EV_CONTROLLER_SUSPEND, 180 MGMT_EV_CONTROLLER_RESUME, 181 MGMT_EV_ADV_MONITOR_DEVICE_FOUND, 182 MGMT_EV_ADV_MONITOR_DEVICE_LOST, 183 }; 184 185 static const u16 mgmt_untrusted_commands[] = { 186 MGMT_OP_READ_INDEX_LIST, 187 MGMT_OP_READ_INFO, 188 MGMT_OP_READ_UNCONF_INDEX_LIST, 189 MGMT_OP_READ_CONFIG_INFO, 190 MGMT_OP_READ_EXT_INDEX_LIST, 191 MGMT_OP_READ_EXT_INFO, 192 MGMT_OP_READ_CONTROLLER_CAP, 193 MGMT_OP_READ_EXP_FEATURES_INFO, 194 MGMT_OP_READ_DEF_SYSTEM_CONFIG, 195 MGMT_OP_READ_DEF_RUNTIME_CONFIG, 196 }; 197 198 static const u16 mgmt_untrusted_events[] = { 199 MGMT_EV_INDEX_ADDED, 200 MGMT_EV_INDEX_REMOVED, 201 MGMT_EV_NEW_SETTINGS, 202 MGMT_EV_CLASS_OF_DEV_CHANGED, 203 MGMT_EV_LOCAL_NAME_CHANGED, 204 MGMT_EV_UNCONF_INDEX_ADDED, 205 MGMT_EV_UNCONF_INDEX_REMOVED, 206 MGMT_EV_NEW_CONFIG_OPTIONS, 207 MGMT_EV_EXT_INDEX_ADDED, 208 MGMT_EV_EXT_INDEX_REMOVED, 209 MGMT_EV_EXT_INFO_CHANGED, 210 MGMT_EV_EXP_FEATURE_CHANGED, 211 }; 212 213 #define CACHE_TIMEOUT msecs_to_jiffies(2 * 1000) 214 215 #define ZERO_KEY "\x00\x00\x00\x00\x00\x00\x00\x00" \ 216 "\x00\x00\x00\x00\x00\x00\x00\x00" 217 218 /* HCI to MGMT error code conversion table */ 219 static const u8 mgmt_status_table[] = { 220 MGMT_STATUS_SUCCESS, 221 MGMT_STATUS_UNKNOWN_COMMAND, /* Unknown Command */ 222 MGMT_STATUS_NOT_CONNECTED, /* No Connection */ 223 MGMT_STATUS_FAILED, /* Hardware Failure */ 224 MGMT_STATUS_CONNECT_FAILED, /* Page Timeout */ 225 MGMT_STATUS_AUTH_FAILED, /* Authentication Failed */ 226 MGMT_STATUS_AUTH_FAILED, /* PIN or Key Missing */ 227 MGMT_STATUS_NO_RESOURCES, /* Memory Full */ 228 MGMT_STATUS_TIMEOUT, /* Connection Timeout */ 229 MGMT_STATUS_NO_RESOURCES, /* Max Number of Connections */ 230 MGMT_STATUS_NO_RESOURCES, /* Max Number of SCO Connections */ 231 MGMT_STATUS_ALREADY_CONNECTED, /* ACL Connection Exists */ 232 MGMT_STATUS_BUSY, /* Command Disallowed */ 233 MGMT_STATUS_NO_RESOURCES, /* Rejected Limited Resources */ 234 MGMT_STATUS_REJECTED, /* Rejected Security */ 235 MGMT_STATUS_REJECTED, /* Rejected Personal */ 236 MGMT_STATUS_TIMEOUT, /* Host Timeout */ 237 MGMT_STATUS_NOT_SUPPORTED, /* Unsupported Feature */ 238 MGMT_STATUS_INVALID_PARAMS, /* Invalid Parameters */ 239 MGMT_STATUS_DISCONNECTED, /* OE User Ended Connection */ 240 MGMT_STATUS_NO_RESOURCES, /* OE Low Resources */ 241 MGMT_STATUS_DISCONNECTED, /* OE Power Off */ 242 MGMT_STATUS_DISCONNECTED, /* Connection Terminated */ 243 MGMT_STATUS_BUSY, /* Repeated Attempts */ 244 MGMT_STATUS_REJECTED, /* Pairing Not Allowed */ 245 MGMT_STATUS_FAILED, /* Unknown LMP PDU */ 246 MGMT_STATUS_NOT_SUPPORTED, /* Unsupported Remote Feature */ 247 MGMT_STATUS_REJECTED, /* SCO Offset Rejected */ 248 MGMT_STATUS_REJECTED, /* SCO Interval Rejected */ 249 MGMT_STATUS_REJECTED, /* Air Mode Rejected */ 250 MGMT_STATUS_INVALID_PARAMS, /* Invalid LMP Parameters */ 251 MGMT_STATUS_FAILED, /* Unspecified Error */ 252 MGMT_STATUS_NOT_SUPPORTED, /* Unsupported LMP Parameter Value */ 253 MGMT_STATUS_FAILED, /* Role Change Not Allowed */ 254 MGMT_STATUS_TIMEOUT, /* LMP Response Timeout */ 255 MGMT_STATUS_FAILED, /* LMP Error Transaction Collision */ 256 MGMT_STATUS_FAILED, /* LMP PDU Not Allowed */ 257 MGMT_STATUS_REJECTED, /* Encryption Mode Not Accepted */ 258 MGMT_STATUS_FAILED, /* Unit Link Key Used */ 259 MGMT_STATUS_NOT_SUPPORTED, /* QoS Not Supported */ 260 MGMT_STATUS_TIMEOUT, /* Instant Passed */ 261 MGMT_STATUS_NOT_SUPPORTED, /* Pairing Not Supported */ 262 MGMT_STATUS_FAILED, /* Transaction Collision */ 263 MGMT_STATUS_FAILED, /* Reserved for future use */ 264 MGMT_STATUS_INVALID_PARAMS, /* Unacceptable Parameter */ 265 MGMT_STATUS_REJECTED, /* QoS Rejected */ 266 MGMT_STATUS_NOT_SUPPORTED, /* Classification Not Supported */ 267 MGMT_STATUS_REJECTED, /* Insufficient Security */ 268 MGMT_STATUS_INVALID_PARAMS, /* Parameter Out Of Range */ 269 MGMT_STATUS_FAILED, /* Reserved for future use */ 270 MGMT_STATUS_BUSY, /* Role Switch Pending */ 271 MGMT_STATUS_FAILED, /* Reserved for future use */ 272 MGMT_STATUS_FAILED, /* Slot Violation */ 273 MGMT_STATUS_FAILED, /* Role Switch Failed */ 274 MGMT_STATUS_INVALID_PARAMS, /* EIR Too Large */ 275 MGMT_STATUS_NOT_SUPPORTED, /* Simple Pairing Not Supported */ 276 MGMT_STATUS_BUSY, /* Host Busy Pairing */ 277 MGMT_STATUS_REJECTED, /* Rejected, No Suitable Channel */ 278 MGMT_STATUS_BUSY, /* Controller Busy */ 279 MGMT_STATUS_INVALID_PARAMS, /* Unsuitable Connection Interval */ 280 MGMT_STATUS_TIMEOUT, /* Directed Advertising Timeout */ 281 MGMT_STATUS_AUTH_FAILED, /* Terminated Due to MIC Failure */ 282 MGMT_STATUS_CONNECT_FAILED, /* Connection Establishment Failed */ 283 MGMT_STATUS_CONNECT_FAILED, /* MAC Connection Failed */ 284 }; 285 286 static u8 mgmt_errno_status(int err) 287 { 288 switch (err) { 289 case 0: 290 return MGMT_STATUS_SUCCESS; 291 case -EPERM: 292 return MGMT_STATUS_REJECTED; 293 case -EINVAL: 294 return MGMT_STATUS_INVALID_PARAMS; 295 case -EOPNOTSUPP: 296 return MGMT_STATUS_NOT_SUPPORTED; 297 case -EBUSY: 298 return MGMT_STATUS_BUSY; 299 case -ETIMEDOUT: 300 return MGMT_STATUS_AUTH_FAILED; 301 case -ENOMEM: 302 return MGMT_STATUS_NO_RESOURCES; 303 case -EISCONN: 304 return MGMT_STATUS_ALREADY_CONNECTED; 305 case -ENOTCONN: 306 return MGMT_STATUS_DISCONNECTED; 307 } 308 309 return MGMT_STATUS_FAILED; 310 } 311 312 static u8 mgmt_status(int err) 313 { 314 if (err < 0) 315 return mgmt_errno_status(err); 316 317 if (err < ARRAY_SIZE(mgmt_status_table)) 318 return mgmt_status_table[err]; 319 320 return MGMT_STATUS_FAILED; 321 } 322 323 static int mgmt_index_event(u16 event, struct hci_dev *hdev, void *data, 324 u16 len, int flag) 325 { 326 return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len, 327 flag, NULL); 328 } 329 330 static int mgmt_limited_event(u16 event, struct hci_dev *hdev, void *data, 331 u16 len, int flag, struct sock *skip_sk) 332 { 333 return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len, 334 flag, skip_sk); 335 } 336 337 static int mgmt_event(u16 event, struct hci_dev *hdev, void *data, u16 len, 338 struct sock *skip_sk) 339 { 340 return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len, 341 HCI_SOCK_TRUSTED, skip_sk); 342 } 343 344 static int mgmt_event_skb(struct sk_buff *skb, struct sock *skip_sk) 345 { 346 return mgmt_send_event_skb(HCI_CHANNEL_CONTROL, skb, HCI_SOCK_TRUSTED, 347 skip_sk); 348 } 349 350 static u8 le_addr_type(u8 mgmt_addr_type) 351 { 352 if (mgmt_addr_type == BDADDR_LE_PUBLIC) 353 return ADDR_LE_DEV_PUBLIC; 354 else 355 return ADDR_LE_DEV_RANDOM; 356 } 357 358 void mgmt_fill_version_info(void *ver) 359 { 360 struct mgmt_rp_read_version *rp = ver; 361 362 rp->version = MGMT_VERSION; 363 rp->revision = cpu_to_le16(MGMT_REVISION); 364 } 365 366 static int read_version(struct sock *sk, struct hci_dev *hdev, void *data, 367 u16 data_len) 368 { 369 struct mgmt_rp_read_version rp; 370 371 bt_dev_dbg(hdev, "sock %p", sk); 372 373 mgmt_fill_version_info(&rp); 374 375 return mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_VERSION, 0, 376 &rp, sizeof(rp)); 377 } 378 379 static int read_commands(struct sock *sk, struct hci_dev *hdev, void *data, 380 u16 data_len) 381 { 382 struct mgmt_rp_read_commands *rp; 383 u16 num_commands, num_events; 384 size_t rp_size; 385 int i, err; 386 387 bt_dev_dbg(hdev, "sock %p", sk); 388 389 if (hci_sock_test_flag(sk, HCI_SOCK_TRUSTED)) { 390 num_commands = ARRAY_SIZE(mgmt_commands); 391 num_events = ARRAY_SIZE(mgmt_events); 392 } else { 393 num_commands = ARRAY_SIZE(mgmt_untrusted_commands); 394 num_events = ARRAY_SIZE(mgmt_untrusted_events); 395 } 396 397 rp_size = sizeof(*rp) + ((num_commands + num_events) * sizeof(u16)); 398 399 rp = kmalloc(rp_size, GFP_KERNEL); 400 if (!rp) 401 return -ENOMEM; 402 403 rp->num_commands = cpu_to_le16(num_commands); 404 rp->num_events = cpu_to_le16(num_events); 405 406 if (hci_sock_test_flag(sk, HCI_SOCK_TRUSTED)) { 407 __le16 *opcode = rp->opcodes; 408 409 for (i = 0; i < num_commands; i++, opcode++) 410 put_unaligned_le16(mgmt_commands[i], opcode); 411 412 for (i = 0; i < num_events; i++, opcode++) 413 put_unaligned_le16(mgmt_events[i], opcode); 414 } else { 415 __le16 *opcode = rp->opcodes; 416 417 for (i = 0; i < num_commands; i++, opcode++) 418 put_unaligned_le16(mgmt_untrusted_commands[i], opcode); 419 420 for (i = 0; i < num_events; i++, opcode++) 421 put_unaligned_le16(mgmt_untrusted_events[i], opcode); 422 } 423 424 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_COMMANDS, 0, 425 rp, rp_size); 426 kfree(rp); 427 428 return err; 429 } 430 431 static int read_index_list(struct sock *sk, struct hci_dev *hdev, void *data, 432 u16 data_len) 433 { 434 struct mgmt_rp_read_index_list *rp; 435 struct hci_dev *d; 436 size_t rp_len; 437 u16 count; 438 int err; 439 440 bt_dev_dbg(hdev, "sock %p", sk); 441 442 read_lock(&hci_dev_list_lock); 443 444 count = 0; 445 list_for_each_entry(d, &hci_dev_list, list) { 446 if (d->dev_type == HCI_PRIMARY && 447 !hci_dev_test_flag(d, HCI_UNCONFIGURED)) 448 count++; 449 } 450 451 rp_len = sizeof(*rp) + (2 * count); 452 rp = kmalloc(rp_len, GFP_ATOMIC); 453 if (!rp) { 454 read_unlock(&hci_dev_list_lock); 455 return -ENOMEM; 456 } 457 458 count = 0; 459 list_for_each_entry(d, &hci_dev_list, list) { 460 if (hci_dev_test_flag(d, HCI_SETUP) || 461 hci_dev_test_flag(d, HCI_CONFIG) || 462 hci_dev_test_flag(d, HCI_USER_CHANNEL)) 463 continue; 464 465 /* Devices marked as raw-only are neither configured 466 * nor unconfigured controllers. 467 */ 468 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks)) 469 continue; 470 471 if (d->dev_type == HCI_PRIMARY && 472 !hci_dev_test_flag(d, HCI_UNCONFIGURED)) { 473 rp->index[count++] = cpu_to_le16(d->id); 474 bt_dev_dbg(hdev, "Added hci%u", d->id); 475 } 476 } 477 478 rp->num_controllers = cpu_to_le16(count); 479 rp_len = sizeof(*rp) + (2 * count); 480 481 read_unlock(&hci_dev_list_lock); 482 483 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_INDEX_LIST, 484 0, rp, rp_len); 485 486 kfree(rp); 487 488 return err; 489 } 490 491 static int read_unconf_index_list(struct sock *sk, struct hci_dev *hdev, 492 void *data, u16 data_len) 493 { 494 struct mgmt_rp_read_unconf_index_list *rp; 495 struct hci_dev *d; 496 size_t rp_len; 497 u16 count; 498 int err; 499 500 bt_dev_dbg(hdev, "sock %p", sk); 501 502 read_lock(&hci_dev_list_lock); 503 504 count = 0; 505 list_for_each_entry(d, &hci_dev_list, list) { 506 if (d->dev_type == HCI_PRIMARY && 507 hci_dev_test_flag(d, HCI_UNCONFIGURED)) 508 count++; 509 } 510 511 rp_len = sizeof(*rp) + (2 * count); 512 rp = kmalloc(rp_len, GFP_ATOMIC); 513 if (!rp) { 514 read_unlock(&hci_dev_list_lock); 515 return -ENOMEM; 516 } 517 518 count = 0; 519 list_for_each_entry(d, &hci_dev_list, list) { 520 if (hci_dev_test_flag(d, HCI_SETUP) || 521 hci_dev_test_flag(d, HCI_CONFIG) || 522 hci_dev_test_flag(d, HCI_USER_CHANNEL)) 523 continue; 524 525 /* Devices marked as raw-only are neither configured 526 * nor unconfigured controllers. 527 */ 528 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks)) 529 continue; 530 531 if (d->dev_type == HCI_PRIMARY && 532 hci_dev_test_flag(d, HCI_UNCONFIGURED)) { 533 rp->index[count++] = cpu_to_le16(d->id); 534 bt_dev_dbg(hdev, "Added hci%u", d->id); 535 } 536 } 537 538 rp->num_controllers = cpu_to_le16(count); 539 rp_len = sizeof(*rp) + (2 * count); 540 541 read_unlock(&hci_dev_list_lock); 542 543 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 544 MGMT_OP_READ_UNCONF_INDEX_LIST, 0, rp, rp_len); 545 546 kfree(rp); 547 548 return err; 549 } 550 551 static int read_ext_index_list(struct sock *sk, struct hci_dev *hdev, 552 void *data, u16 data_len) 553 { 554 struct mgmt_rp_read_ext_index_list *rp; 555 struct hci_dev *d; 556 u16 count; 557 int err; 558 559 bt_dev_dbg(hdev, "sock %p", sk); 560 561 read_lock(&hci_dev_list_lock); 562 563 count = 0; 564 list_for_each_entry(d, &hci_dev_list, list) { 565 if (d->dev_type == HCI_PRIMARY || d->dev_type == HCI_AMP) 566 count++; 567 } 568 569 rp = kmalloc(struct_size(rp, entry, count), GFP_ATOMIC); 570 if (!rp) { 571 read_unlock(&hci_dev_list_lock); 572 return -ENOMEM; 573 } 574 575 count = 0; 576 list_for_each_entry(d, &hci_dev_list, list) { 577 if (hci_dev_test_flag(d, HCI_SETUP) || 578 hci_dev_test_flag(d, HCI_CONFIG) || 579 hci_dev_test_flag(d, HCI_USER_CHANNEL)) 580 continue; 581 582 /* Devices marked as raw-only are neither configured 583 * nor unconfigured controllers. 584 */ 585 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks)) 586 continue; 587 588 if (d->dev_type == HCI_PRIMARY) { 589 if (hci_dev_test_flag(d, HCI_UNCONFIGURED)) 590 rp->entry[count].type = 0x01; 591 else 592 rp->entry[count].type = 0x00; 593 } else if (d->dev_type == HCI_AMP) { 594 rp->entry[count].type = 0x02; 595 } else { 596 continue; 597 } 598 599 rp->entry[count].bus = d->bus; 600 rp->entry[count++].index = cpu_to_le16(d->id); 601 bt_dev_dbg(hdev, "Added hci%u", d->id); 602 } 603 604 rp->num_controllers = cpu_to_le16(count); 605 606 read_unlock(&hci_dev_list_lock); 607 608 /* If this command is called at least once, then all the 609 * default index and unconfigured index events are disabled 610 * and from now on only extended index events are used. 611 */ 612 hci_sock_set_flag(sk, HCI_MGMT_EXT_INDEX_EVENTS); 613 hci_sock_clear_flag(sk, HCI_MGMT_INDEX_EVENTS); 614 hci_sock_clear_flag(sk, HCI_MGMT_UNCONF_INDEX_EVENTS); 615 616 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 617 MGMT_OP_READ_EXT_INDEX_LIST, 0, rp, 618 struct_size(rp, entry, count)); 619 620 kfree(rp); 621 622 return err; 623 } 624 625 static bool is_configured(struct hci_dev *hdev) 626 { 627 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) && 628 !hci_dev_test_flag(hdev, HCI_EXT_CONFIGURED)) 629 return false; 630 631 if ((test_bit(HCI_QUIRK_INVALID_BDADDR, &hdev->quirks) || 632 test_bit(HCI_QUIRK_USE_BDADDR_PROPERTY, &hdev->quirks)) && 633 !bacmp(&hdev->public_addr, BDADDR_ANY)) 634 return false; 635 636 return true; 637 } 638 639 static __le32 get_missing_options(struct hci_dev *hdev) 640 { 641 u32 options = 0; 642 643 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) && 644 !hci_dev_test_flag(hdev, HCI_EXT_CONFIGURED)) 645 options |= MGMT_OPTION_EXTERNAL_CONFIG; 646 647 if ((test_bit(HCI_QUIRK_INVALID_BDADDR, &hdev->quirks) || 648 test_bit(HCI_QUIRK_USE_BDADDR_PROPERTY, &hdev->quirks)) && 649 !bacmp(&hdev->public_addr, BDADDR_ANY)) 650 options |= MGMT_OPTION_PUBLIC_ADDRESS; 651 652 return cpu_to_le32(options); 653 } 654 655 static int new_options(struct hci_dev *hdev, struct sock *skip) 656 { 657 __le32 options = get_missing_options(hdev); 658 659 return mgmt_limited_event(MGMT_EV_NEW_CONFIG_OPTIONS, hdev, &options, 660 sizeof(options), HCI_MGMT_OPTION_EVENTS, skip); 661 } 662 663 static int send_options_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev) 664 { 665 __le32 options = get_missing_options(hdev); 666 667 return mgmt_cmd_complete(sk, hdev->id, opcode, 0, &options, 668 sizeof(options)); 669 } 670 671 static int read_config_info(struct sock *sk, struct hci_dev *hdev, 672 void *data, u16 data_len) 673 { 674 struct mgmt_rp_read_config_info rp; 675 u32 options = 0; 676 677 bt_dev_dbg(hdev, "sock %p", sk); 678 679 hci_dev_lock(hdev); 680 681 memset(&rp, 0, sizeof(rp)); 682 rp.manufacturer = cpu_to_le16(hdev->manufacturer); 683 684 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks)) 685 options |= MGMT_OPTION_EXTERNAL_CONFIG; 686 687 if (hdev->set_bdaddr) 688 options |= MGMT_OPTION_PUBLIC_ADDRESS; 689 690 rp.supported_options = cpu_to_le32(options); 691 rp.missing_options = get_missing_options(hdev); 692 693 hci_dev_unlock(hdev); 694 695 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_CONFIG_INFO, 0, 696 &rp, sizeof(rp)); 697 } 698 699 static u32 get_supported_phys(struct hci_dev *hdev) 700 { 701 u32 supported_phys = 0; 702 703 if (lmp_bredr_capable(hdev)) { 704 supported_phys |= MGMT_PHY_BR_1M_1SLOT; 705 706 if (hdev->features[0][0] & LMP_3SLOT) 707 supported_phys |= MGMT_PHY_BR_1M_3SLOT; 708 709 if (hdev->features[0][0] & LMP_5SLOT) 710 supported_phys |= MGMT_PHY_BR_1M_5SLOT; 711 712 if (lmp_edr_2m_capable(hdev)) { 713 supported_phys |= MGMT_PHY_EDR_2M_1SLOT; 714 715 if (lmp_edr_3slot_capable(hdev)) 716 supported_phys |= MGMT_PHY_EDR_2M_3SLOT; 717 718 if (lmp_edr_5slot_capable(hdev)) 719 supported_phys |= MGMT_PHY_EDR_2M_5SLOT; 720 721 if (lmp_edr_3m_capable(hdev)) { 722 supported_phys |= MGMT_PHY_EDR_3M_1SLOT; 723 724 if (lmp_edr_3slot_capable(hdev)) 725 supported_phys |= MGMT_PHY_EDR_3M_3SLOT; 726 727 if (lmp_edr_5slot_capable(hdev)) 728 supported_phys |= MGMT_PHY_EDR_3M_5SLOT; 729 } 730 } 731 } 732 733 if (lmp_le_capable(hdev)) { 734 supported_phys |= MGMT_PHY_LE_1M_TX; 735 supported_phys |= MGMT_PHY_LE_1M_RX; 736 737 if (hdev->le_features[1] & HCI_LE_PHY_2M) { 738 supported_phys |= MGMT_PHY_LE_2M_TX; 739 supported_phys |= MGMT_PHY_LE_2M_RX; 740 } 741 742 if (hdev->le_features[1] & HCI_LE_PHY_CODED) { 743 supported_phys |= MGMT_PHY_LE_CODED_TX; 744 supported_phys |= MGMT_PHY_LE_CODED_RX; 745 } 746 } 747 748 return supported_phys; 749 } 750 751 static u32 get_selected_phys(struct hci_dev *hdev) 752 { 753 u32 selected_phys = 0; 754 755 if (lmp_bredr_capable(hdev)) { 756 selected_phys |= MGMT_PHY_BR_1M_1SLOT; 757 758 if (hdev->pkt_type & (HCI_DM3 | HCI_DH3)) 759 selected_phys |= MGMT_PHY_BR_1M_3SLOT; 760 761 if (hdev->pkt_type & (HCI_DM5 | HCI_DH5)) 762 selected_phys |= MGMT_PHY_BR_1M_5SLOT; 763 764 if (lmp_edr_2m_capable(hdev)) { 765 if (!(hdev->pkt_type & HCI_2DH1)) 766 selected_phys |= MGMT_PHY_EDR_2M_1SLOT; 767 768 if (lmp_edr_3slot_capable(hdev) && 769 !(hdev->pkt_type & HCI_2DH3)) 770 selected_phys |= MGMT_PHY_EDR_2M_3SLOT; 771 772 if (lmp_edr_5slot_capable(hdev) && 773 !(hdev->pkt_type & HCI_2DH5)) 774 selected_phys |= MGMT_PHY_EDR_2M_5SLOT; 775 776 if (lmp_edr_3m_capable(hdev)) { 777 if (!(hdev->pkt_type & HCI_3DH1)) 778 selected_phys |= MGMT_PHY_EDR_3M_1SLOT; 779 780 if (lmp_edr_3slot_capable(hdev) && 781 !(hdev->pkt_type & HCI_3DH3)) 782 selected_phys |= MGMT_PHY_EDR_3M_3SLOT; 783 784 if (lmp_edr_5slot_capable(hdev) && 785 !(hdev->pkt_type & HCI_3DH5)) 786 selected_phys |= MGMT_PHY_EDR_3M_5SLOT; 787 } 788 } 789 } 790 791 if (lmp_le_capable(hdev)) { 792 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_1M) 793 selected_phys |= MGMT_PHY_LE_1M_TX; 794 795 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_1M) 796 selected_phys |= MGMT_PHY_LE_1M_RX; 797 798 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_2M) 799 selected_phys |= MGMT_PHY_LE_2M_TX; 800 801 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_2M) 802 selected_phys |= MGMT_PHY_LE_2M_RX; 803 804 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_CODED) 805 selected_phys |= MGMT_PHY_LE_CODED_TX; 806 807 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_CODED) 808 selected_phys |= MGMT_PHY_LE_CODED_RX; 809 } 810 811 return selected_phys; 812 } 813 814 static u32 get_configurable_phys(struct hci_dev *hdev) 815 { 816 return (get_supported_phys(hdev) & ~MGMT_PHY_BR_1M_1SLOT & 817 ~MGMT_PHY_LE_1M_TX & ~MGMT_PHY_LE_1M_RX); 818 } 819 820 static u32 get_supported_settings(struct hci_dev *hdev) 821 { 822 u32 settings = 0; 823 824 settings |= MGMT_SETTING_POWERED; 825 settings |= MGMT_SETTING_BONDABLE; 826 settings |= MGMT_SETTING_DEBUG_KEYS; 827 settings |= MGMT_SETTING_CONNECTABLE; 828 settings |= MGMT_SETTING_DISCOVERABLE; 829 830 if (lmp_bredr_capable(hdev)) { 831 if (hdev->hci_ver >= BLUETOOTH_VER_1_2) 832 settings |= MGMT_SETTING_FAST_CONNECTABLE; 833 settings |= MGMT_SETTING_BREDR; 834 settings |= MGMT_SETTING_LINK_SECURITY; 835 836 if (lmp_ssp_capable(hdev)) { 837 settings |= MGMT_SETTING_SSP; 838 if (IS_ENABLED(CONFIG_BT_HS)) 839 settings |= MGMT_SETTING_HS; 840 } 841 842 if (lmp_sc_capable(hdev)) 843 settings |= MGMT_SETTING_SECURE_CONN; 844 845 if (test_bit(HCI_QUIRK_WIDEBAND_SPEECH_SUPPORTED, 846 &hdev->quirks)) 847 settings |= MGMT_SETTING_WIDEBAND_SPEECH; 848 } 849 850 if (lmp_le_capable(hdev)) { 851 settings |= MGMT_SETTING_LE; 852 settings |= MGMT_SETTING_SECURE_CONN; 853 settings |= MGMT_SETTING_PRIVACY; 854 settings |= MGMT_SETTING_STATIC_ADDRESS; 855 settings |= MGMT_SETTING_ADVERTISING; 856 } 857 858 if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) || 859 hdev->set_bdaddr) 860 settings |= MGMT_SETTING_CONFIGURATION; 861 862 if (cis_central_capable(hdev)) 863 settings |= MGMT_SETTING_CIS_CENTRAL; 864 865 if (cis_peripheral_capable(hdev)) 866 settings |= MGMT_SETTING_CIS_PERIPHERAL; 867 868 settings |= MGMT_SETTING_PHY_CONFIGURATION; 869 870 return settings; 871 } 872 873 static u32 get_current_settings(struct hci_dev *hdev) 874 { 875 u32 settings = 0; 876 877 if (hdev_is_powered(hdev)) 878 settings |= MGMT_SETTING_POWERED; 879 880 if (hci_dev_test_flag(hdev, HCI_CONNECTABLE)) 881 settings |= MGMT_SETTING_CONNECTABLE; 882 883 if (hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE)) 884 settings |= MGMT_SETTING_FAST_CONNECTABLE; 885 886 if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE)) 887 settings |= MGMT_SETTING_DISCOVERABLE; 888 889 if (hci_dev_test_flag(hdev, HCI_BONDABLE)) 890 settings |= MGMT_SETTING_BONDABLE; 891 892 if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 893 settings |= MGMT_SETTING_BREDR; 894 895 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 896 settings |= MGMT_SETTING_LE; 897 898 if (hci_dev_test_flag(hdev, HCI_LINK_SECURITY)) 899 settings |= MGMT_SETTING_LINK_SECURITY; 900 901 if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) 902 settings |= MGMT_SETTING_SSP; 903 904 if (hci_dev_test_flag(hdev, HCI_HS_ENABLED)) 905 settings |= MGMT_SETTING_HS; 906 907 if (hci_dev_test_flag(hdev, HCI_ADVERTISING)) 908 settings |= MGMT_SETTING_ADVERTISING; 909 910 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED)) 911 settings |= MGMT_SETTING_SECURE_CONN; 912 913 if (hci_dev_test_flag(hdev, HCI_KEEP_DEBUG_KEYS)) 914 settings |= MGMT_SETTING_DEBUG_KEYS; 915 916 if (hci_dev_test_flag(hdev, HCI_PRIVACY)) 917 settings |= MGMT_SETTING_PRIVACY; 918 919 /* The current setting for static address has two purposes. The 920 * first is to indicate if the static address will be used and 921 * the second is to indicate if it is actually set. 922 * 923 * This means if the static address is not configured, this flag 924 * will never be set. If the address is configured, then if the 925 * address is actually used decides if the flag is set or not. 926 * 927 * For single mode LE only controllers and dual-mode controllers 928 * with BR/EDR disabled, the existence of the static address will 929 * be evaluated. 930 */ 931 if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) || 932 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) || 933 !bacmp(&hdev->bdaddr, BDADDR_ANY)) { 934 if (bacmp(&hdev->static_addr, BDADDR_ANY)) 935 settings |= MGMT_SETTING_STATIC_ADDRESS; 936 } 937 938 if (hci_dev_test_flag(hdev, HCI_WIDEBAND_SPEECH_ENABLED)) 939 settings |= MGMT_SETTING_WIDEBAND_SPEECH; 940 941 if (cis_central_capable(hdev)) 942 settings |= MGMT_SETTING_CIS_CENTRAL; 943 944 if (cis_peripheral_capable(hdev)) 945 settings |= MGMT_SETTING_CIS_PERIPHERAL; 946 947 if (bis_capable(hdev)) 948 settings |= MGMT_SETTING_ISO_BROADCASTER; 949 950 if (sync_recv_capable(hdev)) 951 settings |= MGMT_SETTING_ISO_SYNC_RECEIVER; 952 953 return settings; 954 } 955 956 static struct mgmt_pending_cmd *pending_find(u16 opcode, struct hci_dev *hdev) 957 { 958 return mgmt_pending_find(HCI_CHANNEL_CONTROL, opcode, hdev); 959 } 960 961 u8 mgmt_get_adv_discov_flags(struct hci_dev *hdev) 962 { 963 struct mgmt_pending_cmd *cmd; 964 965 /* If there's a pending mgmt command the flags will not yet have 966 * their final values, so check for this first. 967 */ 968 cmd = pending_find(MGMT_OP_SET_DISCOVERABLE, hdev); 969 if (cmd) { 970 struct mgmt_mode *cp = cmd->param; 971 if (cp->val == 0x01) 972 return LE_AD_GENERAL; 973 else if (cp->val == 0x02) 974 return LE_AD_LIMITED; 975 } else { 976 if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE)) 977 return LE_AD_LIMITED; 978 else if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE)) 979 return LE_AD_GENERAL; 980 } 981 982 return 0; 983 } 984 985 bool mgmt_get_connectable(struct hci_dev *hdev) 986 { 987 struct mgmt_pending_cmd *cmd; 988 989 /* If there's a pending mgmt command the flag will not yet have 990 * it's final value, so check for this first. 991 */ 992 cmd = pending_find(MGMT_OP_SET_CONNECTABLE, hdev); 993 if (cmd) { 994 struct mgmt_mode *cp = cmd->param; 995 996 return cp->val; 997 } 998 999 return hci_dev_test_flag(hdev, HCI_CONNECTABLE); 1000 } 1001 1002 static int service_cache_sync(struct hci_dev *hdev, void *data) 1003 { 1004 hci_update_eir_sync(hdev); 1005 hci_update_class_sync(hdev); 1006 1007 return 0; 1008 } 1009 1010 static void service_cache_off(struct work_struct *work) 1011 { 1012 struct hci_dev *hdev = container_of(work, struct hci_dev, 1013 service_cache.work); 1014 1015 if (!hci_dev_test_and_clear_flag(hdev, HCI_SERVICE_CACHE)) 1016 return; 1017 1018 hci_cmd_sync_queue(hdev, service_cache_sync, NULL, NULL); 1019 } 1020 1021 static int rpa_expired_sync(struct hci_dev *hdev, void *data) 1022 { 1023 /* The generation of a new RPA and programming it into the 1024 * controller happens in the hci_req_enable_advertising() 1025 * function. 1026 */ 1027 if (ext_adv_capable(hdev)) 1028 return hci_start_ext_adv_sync(hdev, hdev->cur_adv_instance); 1029 else 1030 return hci_enable_advertising_sync(hdev); 1031 } 1032 1033 static void rpa_expired(struct work_struct *work) 1034 { 1035 struct hci_dev *hdev = container_of(work, struct hci_dev, 1036 rpa_expired.work); 1037 1038 bt_dev_dbg(hdev, ""); 1039 1040 hci_dev_set_flag(hdev, HCI_RPA_EXPIRED); 1041 1042 if (!hci_dev_test_flag(hdev, HCI_ADVERTISING)) 1043 return; 1044 1045 hci_cmd_sync_queue(hdev, rpa_expired_sync, NULL, NULL); 1046 } 1047 1048 static int set_discoverable_sync(struct hci_dev *hdev, void *data); 1049 1050 static void discov_off(struct work_struct *work) 1051 { 1052 struct hci_dev *hdev = container_of(work, struct hci_dev, 1053 discov_off.work); 1054 1055 bt_dev_dbg(hdev, ""); 1056 1057 hci_dev_lock(hdev); 1058 1059 /* When discoverable timeout triggers, then just make sure 1060 * the limited discoverable flag is cleared. Even in the case 1061 * of a timeout triggered from general discoverable, it is 1062 * safe to unconditionally clear the flag. 1063 */ 1064 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1065 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1066 hdev->discov_timeout = 0; 1067 1068 hci_cmd_sync_queue(hdev, set_discoverable_sync, NULL, NULL); 1069 1070 mgmt_new_settings(hdev); 1071 1072 hci_dev_unlock(hdev); 1073 } 1074 1075 static int send_settings_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev); 1076 1077 static void mesh_send_complete(struct hci_dev *hdev, 1078 struct mgmt_mesh_tx *mesh_tx, bool silent) 1079 { 1080 u8 handle = mesh_tx->handle; 1081 1082 if (!silent) 1083 mgmt_event(MGMT_EV_MESH_PACKET_CMPLT, hdev, &handle, 1084 sizeof(handle), NULL); 1085 1086 mgmt_mesh_remove(mesh_tx); 1087 } 1088 1089 static int mesh_send_done_sync(struct hci_dev *hdev, void *data) 1090 { 1091 struct mgmt_mesh_tx *mesh_tx; 1092 1093 hci_dev_clear_flag(hdev, HCI_MESH_SENDING); 1094 hci_disable_advertising_sync(hdev); 1095 mesh_tx = mgmt_mesh_next(hdev, NULL); 1096 1097 if (mesh_tx) 1098 mesh_send_complete(hdev, mesh_tx, false); 1099 1100 return 0; 1101 } 1102 1103 static int mesh_send_sync(struct hci_dev *hdev, void *data); 1104 static void mesh_send_start_complete(struct hci_dev *hdev, void *data, int err); 1105 static void mesh_next(struct hci_dev *hdev, void *data, int err) 1106 { 1107 struct mgmt_mesh_tx *mesh_tx = mgmt_mesh_next(hdev, NULL); 1108 1109 if (!mesh_tx) 1110 return; 1111 1112 err = hci_cmd_sync_queue(hdev, mesh_send_sync, mesh_tx, 1113 mesh_send_start_complete); 1114 1115 if (err < 0) 1116 mesh_send_complete(hdev, mesh_tx, false); 1117 else 1118 hci_dev_set_flag(hdev, HCI_MESH_SENDING); 1119 } 1120 1121 static void mesh_send_done(struct work_struct *work) 1122 { 1123 struct hci_dev *hdev = container_of(work, struct hci_dev, 1124 mesh_send_done.work); 1125 1126 if (!hci_dev_test_flag(hdev, HCI_MESH_SENDING)) 1127 return; 1128 1129 hci_cmd_sync_queue(hdev, mesh_send_done_sync, NULL, mesh_next); 1130 } 1131 1132 static void mgmt_init_hdev(struct sock *sk, struct hci_dev *hdev) 1133 { 1134 if (hci_dev_test_flag(hdev, HCI_MGMT)) 1135 return; 1136 1137 BT_INFO("MGMT ver %d.%d", MGMT_VERSION, MGMT_REVISION); 1138 1139 INIT_DELAYED_WORK(&hdev->discov_off, discov_off); 1140 INIT_DELAYED_WORK(&hdev->service_cache, service_cache_off); 1141 INIT_DELAYED_WORK(&hdev->rpa_expired, rpa_expired); 1142 INIT_DELAYED_WORK(&hdev->mesh_send_done, mesh_send_done); 1143 1144 /* Non-mgmt controlled devices get this bit set 1145 * implicitly so that pairing works for them, however 1146 * for mgmt we require user-space to explicitly enable 1147 * it 1148 */ 1149 hci_dev_clear_flag(hdev, HCI_BONDABLE); 1150 1151 hci_dev_set_flag(hdev, HCI_MGMT); 1152 } 1153 1154 static int read_controller_info(struct sock *sk, struct hci_dev *hdev, 1155 void *data, u16 data_len) 1156 { 1157 struct mgmt_rp_read_info rp; 1158 1159 bt_dev_dbg(hdev, "sock %p", sk); 1160 1161 hci_dev_lock(hdev); 1162 1163 memset(&rp, 0, sizeof(rp)); 1164 1165 bacpy(&rp.bdaddr, &hdev->bdaddr); 1166 1167 rp.version = hdev->hci_ver; 1168 rp.manufacturer = cpu_to_le16(hdev->manufacturer); 1169 1170 rp.supported_settings = cpu_to_le32(get_supported_settings(hdev)); 1171 rp.current_settings = cpu_to_le32(get_current_settings(hdev)); 1172 1173 memcpy(rp.dev_class, hdev->dev_class, 3); 1174 1175 memcpy(rp.name, hdev->dev_name, sizeof(hdev->dev_name)); 1176 memcpy(rp.short_name, hdev->short_name, sizeof(hdev->short_name)); 1177 1178 hci_dev_unlock(hdev); 1179 1180 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_INFO, 0, &rp, 1181 sizeof(rp)); 1182 } 1183 1184 static u16 append_eir_data_to_buf(struct hci_dev *hdev, u8 *eir) 1185 { 1186 u16 eir_len = 0; 1187 size_t name_len; 1188 1189 if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1190 eir_len = eir_append_data(eir, eir_len, EIR_CLASS_OF_DEV, 1191 hdev->dev_class, 3); 1192 1193 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 1194 eir_len = eir_append_le16(eir, eir_len, EIR_APPEARANCE, 1195 hdev->appearance); 1196 1197 name_len = strnlen(hdev->dev_name, sizeof(hdev->dev_name)); 1198 eir_len = eir_append_data(eir, eir_len, EIR_NAME_COMPLETE, 1199 hdev->dev_name, name_len); 1200 1201 name_len = strnlen(hdev->short_name, sizeof(hdev->short_name)); 1202 eir_len = eir_append_data(eir, eir_len, EIR_NAME_SHORT, 1203 hdev->short_name, name_len); 1204 1205 return eir_len; 1206 } 1207 1208 static int read_ext_controller_info(struct sock *sk, struct hci_dev *hdev, 1209 void *data, u16 data_len) 1210 { 1211 char buf[512]; 1212 struct mgmt_rp_read_ext_info *rp = (void *)buf; 1213 u16 eir_len; 1214 1215 bt_dev_dbg(hdev, "sock %p", sk); 1216 1217 memset(&buf, 0, sizeof(buf)); 1218 1219 hci_dev_lock(hdev); 1220 1221 bacpy(&rp->bdaddr, &hdev->bdaddr); 1222 1223 rp->version = hdev->hci_ver; 1224 rp->manufacturer = cpu_to_le16(hdev->manufacturer); 1225 1226 rp->supported_settings = cpu_to_le32(get_supported_settings(hdev)); 1227 rp->current_settings = cpu_to_le32(get_current_settings(hdev)); 1228 1229 1230 eir_len = append_eir_data_to_buf(hdev, rp->eir); 1231 rp->eir_len = cpu_to_le16(eir_len); 1232 1233 hci_dev_unlock(hdev); 1234 1235 /* If this command is called at least once, then the events 1236 * for class of device and local name changes are disabled 1237 * and only the new extended controller information event 1238 * is used. 1239 */ 1240 hci_sock_set_flag(sk, HCI_MGMT_EXT_INFO_EVENTS); 1241 hci_sock_clear_flag(sk, HCI_MGMT_DEV_CLASS_EVENTS); 1242 hci_sock_clear_flag(sk, HCI_MGMT_LOCAL_NAME_EVENTS); 1243 1244 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_EXT_INFO, 0, rp, 1245 sizeof(*rp) + eir_len); 1246 } 1247 1248 static int ext_info_changed(struct hci_dev *hdev, struct sock *skip) 1249 { 1250 char buf[512]; 1251 struct mgmt_ev_ext_info_changed *ev = (void *)buf; 1252 u16 eir_len; 1253 1254 memset(buf, 0, sizeof(buf)); 1255 1256 eir_len = append_eir_data_to_buf(hdev, ev->eir); 1257 ev->eir_len = cpu_to_le16(eir_len); 1258 1259 return mgmt_limited_event(MGMT_EV_EXT_INFO_CHANGED, hdev, ev, 1260 sizeof(*ev) + eir_len, 1261 HCI_MGMT_EXT_INFO_EVENTS, skip); 1262 } 1263 1264 static int send_settings_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev) 1265 { 1266 __le32 settings = cpu_to_le32(get_current_settings(hdev)); 1267 1268 return mgmt_cmd_complete(sk, hdev->id, opcode, 0, &settings, 1269 sizeof(settings)); 1270 } 1271 1272 void mgmt_advertising_added(struct sock *sk, struct hci_dev *hdev, u8 instance) 1273 { 1274 struct mgmt_ev_advertising_added ev; 1275 1276 ev.instance = instance; 1277 1278 mgmt_event(MGMT_EV_ADVERTISING_ADDED, hdev, &ev, sizeof(ev), sk); 1279 } 1280 1281 void mgmt_advertising_removed(struct sock *sk, struct hci_dev *hdev, 1282 u8 instance) 1283 { 1284 struct mgmt_ev_advertising_removed ev; 1285 1286 ev.instance = instance; 1287 1288 mgmt_event(MGMT_EV_ADVERTISING_REMOVED, hdev, &ev, sizeof(ev), sk); 1289 } 1290 1291 static void cancel_adv_timeout(struct hci_dev *hdev) 1292 { 1293 if (hdev->adv_instance_timeout) { 1294 hdev->adv_instance_timeout = 0; 1295 cancel_delayed_work(&hdev->adv_instance_expire); 1296 } 1297 } 1298 1299 /* This function requires the caller holds hdev->lock */ 1300 static void restart_le_actions(struct hci_dev *hdev) 1301 { 1302 struct hci_conn_params *p; 1303 1304 list_for_each_entry(p, &hdev->le_conn_params, list) { 1305 /* Needed for AUTO_OFF case where might not "really" 1306 * have been powered off. 1307 */ 1308 hci_pend_le_list_del_init(p); 1309 1310 switch (p->auto_connect) { 1311 case HCI_AUTO_CONN_DIRECT: 1312 case HCI_AUTO_CONN_ALWAYS: 1313 hci_pend_le_list_add(p, &hdev->pend_le_conns); 1314 break; 1315 case HCI_AUTO_CONN_REPORT: 1316 hci_pend_le_list_add(p, &hdev->pend_le_reports); 1317 break; 1318 default: 1319 break; 1320 } 1321 } 1322 } 1323 1324 static int new_settings(struct hci_dev *hdev, struct sock *skip) 1325 { 1326 __le32 ev = cpu_to_le32(get_current_settings(hdev)); 1327 1328 return mgmt_limited_event(MGMT_EV_NEW_SETTINGS, hdev, &ev, 1329 sizeof(ev), HCI_MGMT_SETTING_EVENTS, skip); 1330 } 1331 1332 static void mgmt_set_powered_complete(struct hci_dev *hdev, void *data, int err) 1333 { 1334 struct mgmt_pending_cmd *cmd = data; 1335 struct mgmt_mode *cp; 1336 1337 /* Make sure cmd still outstanding. */ 1338 if (cmd != pending_find(MGMT_OP_SET_POWERED, hdev)) 1339 return; 1340 1341 cp = cmd->param; 1342 1343 bt_dev_dbg(hdev, "err %d", err); 1344 1345 if (!err) { 1346 if (cp->val) { 1347 hci_dev_lock(hdev); 1348 restart_le_actions(hdev); 1349 hci_update_passive_scan(hdev); 1350 hci_dev_unlock(hdev); 1351 } 1352 1353 send_settings_rsp(cmd->sk, cmd->opcode, hdev); 1354 1355 /* Only call new_setting for power on as power off is deferred 1356 * to hdev->power_off work which does call hci_dev_do_close. 1357 */ 1358 if (cp->val) 1359 new_settings(hdev, cmd->sk); 1360 } else { 1361 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_POWERED, 1362 mgmt_status(err)); 1363 } 1364 1365 mgmt_pending_remove(cmd); 1366 } 1367 1368 static int set_powered_sync(struct hci_dev *hdev, void *data) 1369 { 1370 struct mgmt_pending_cmd *cmd = data; 1371 struct mgmt_mode *cp = cmd->param; 1372 1373 BT_DBG("%s", hdev->name); 1374 1375 return hci_set_powered_sync(hdev, cp->val); 1376 } 1377 1378 static int set_powered(struct sock *sk, struct hci_dev *hdev, void *data, 1379 u16 len) 1380 { 1381 struct mgmt_mode *cp = data; 1382 struct mgmt_pending_cmd *cmd; 1383 int err; 1384 1385 bt_dev_dbg(hdev, "sock %p", sk); 1386 1387 if (cp->val != 0x00 && cp->val != 0x01) 1388 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_POWERED, 1389 MGMT_STATUS_INVALID_PARAMS); 1390 1391 hci_dev_lock(hdev); 1392 1393 if (pending_find(MGMT_OP_SET_POWERED, hdev)) { 1394 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_POWERED, 1395 MGMT_STATUS_BUSY); 1396 goto failed; 1397 } 1398 1399 if (!!cp->val == hdev_is_powered(hdev)) { 1400 err = send_settings_rsp(sk, MGMT_OP_SET_POWERED, hdev); 1401 goto failed; 1402 } 1403 1404 cmd = mgmt_pending_add(sk, MGMT_OP_SET_POWERED, hdev, data, len); 1405 if (!cmd) { 1406 err = -ENOMEM; 1407 goto failed; 1408 } 1409 1410 /* Cancel potentially blocking sync operation before power off */ 1411 if (cp->val == 0x00) { 1412 __hci_cmd_sync_cancel(hdev, -EHOSTDOWN); 1413 err = hci_cmd_sync_queue(hdev, set_powered_sync, cmd, 1414 mgmt_set_powered_complete); 1415 } else { 1416 /* Use hci_cmd_sync_submit since hdev might not be running */ 1417 err = hci_cmd_sync_submit(hdev, set_powered_sync, cmd, 1418 mgmt_set_powered_complete); 1419 } 1420 1421 if (err < 0) 1422 mgmt_pending_remove(cmd); 1423 1424 failed: 1425 hci_dev_unlock(hdev); 1426 return err; 1427 } 1428 1429 int mgmt_new_settings(struct hci_dev *hdev) 1430 { 1431 return new_settings(hdev, NULL); 1432 } 1433 1434 struct cmd_lookup { 1435 struct sock *sk; 1436 struct hci_dev *hdev; 1437 u8 mgmt_status; 1438 }; 1439 1440 static void settings_rsp(struct mgmt_pending_cmd *cmd, void *data) 1441 { 1442 struct cmd_lookup *match = data; 1443 1444 send_settings_rsp(cmd->sk, cmd->opcode, match->hdev); 1445 1446 list_del(&cmd->list); 1447 1448 if (match->sk == NULL) { 1449 match->sk = cmd->sk; 1450 sock_hold(match->sk); 1451 } 1452 1453 mgmt_pending_free(cmd); 1454 } 1455 1456 static void cmd_status_rsp(struct mgmt_pending_cmd *cmd, void *data) 1457 { 1458 u8 *status = data; 1459 1460 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, *status); 1461 mgmt_pending_remove(cmd); 1462 } 1463 1464 static void cmd_complete_rsp(struct mgmt_pending_cmd *cmd, void *data) 1465 { 1466 if (cmd->cmd_complete) { 1467 u8 *status = data; 1468 1469 cmd->cmd_complete(cmd, *status); 1470 mgmt_pending_remove(cmd); 1471 1472 return; 1473 } 1474 1475 cmd_status_rsp(cmd, data); 1476 } 1477 1478 static int generic_cmd_complete(struct mgmt_pending_cmd *cmd, u8 status) 1479 { 1480 return mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status, 1481 cmd->param, cmd->param_len); 1482 } 1483 1484 static int addr_cmd_complete(struct mgmt_pending_cmd *cmd, u8 status) 1485 { 1486 return mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status, 1487 cmd->param, sizeof(struct mgmt_addr_info)); 1488 } 1489 1490 static u8 mgmt_bredr_support(struct hci_dev *hdev) 1491 { 1492 if (!lmp_bredr_capable(hdev)) 1493 return MGMT_STATUS_NOT_SUPPORTED; 1494 else if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1495 return MGMT_STATUS_REJECTED; 1496 else 1497 return MGMT_STATUS_SUCCESS; 1498 } 1499 1500 static u8 mgmt_le_support(struct hci_dev *hdev) 1501 { 1502 if (!lmp_le_capable(hdev)) 1503 return MGMT_STATUS_NOT_SUPPORTED; 1504 else if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 1505 return MGMT_STATUS_REJECTED; 1506 else 1507 return MGMT_STATUS_SUCCESS; 1508 } 1509 1510 static void mgmt_set_discoverable_complete(struct hci_dev *hdev, void *data, 1511 int err) 1512 { 1513 struct mgmt_pending_cmd *cmd = data; 1514 1515 bt_dev_dbg(hdev, "err %d", err); 1516 1517 /* Make sure cmd still outstanding. */ 1518 if (cmd != pending_find(MGMT_OP_SET_DISCOVERABLE, hdev)) 1519 return; 1520 1521 hci_dev_lock(hdev); 1522 1523 if (err) { 1524 u8 mgmt_err = mgmt_status(err); 1525 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 1526 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1527 goto done; 1528 } 1529 1530 if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE) && 1531 hdev->discov_timeout > 0) { 1532 int to = msecs_to_jiffies(hdev->discov_timeout * 1000); 1533 queue_delayed_work(hdev->req_workqueue, &hdev->discov_off, to); 1534 } 1535 1536 send_settings_rsp(cmd->sk, MGMT_OP_SET_DISCOVERABLE, hdev); 1537 new_settings(hdev, cmd->sk); 1538 1539 done: 1540 mgmt_pending_remove(cmd); 1541 hci_dev_unlock(hdev); 1542 } 1543 1544 static int set_discoverable_sync(struct hci_dev *hdev, void *data) 1545 { 1546 BT_DBG("%s", hdev->name); 1547 1548 return hci_update_discoverable_sync(hdev); 1549 } 1550 1551 static int set_discoverable(struct sock *sk, struct hci_dev *hdev, void *data, 1552 u16 len) 1553 { 1554 struct mgmt_cp_set_discoverable *cp = data; 1555 struct mgmt_pending_cmd *cmd; 1556 u16 timeout; 1557 int err; 1558 1559 bt_dev_dbg(hdev, "sock %p", sk); 1560 1561 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) && 1562 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1563 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1564 MGMT_STATUS_REJECTED); 1565 1566 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 1567 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1568 MGMT_STATUS_INVALID_PARAMS); 1569 1570 timeout = __le16_to_cpu(cp->timeout); 1571 1572 /* Disabling discoverable requires that no timeout is set, 1573 * and enabling limited discoverable requires a timeout. 1574 */ 1575 if ((cp->val == 0x00 && timeout > 0) || 1576 (cp->val == 0x02 && timeout == 0)) 1577 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1578 MGMT_STATUS_INVALID_PARAMS); 1579 1580 hci_dev_lock(hdev); 1581 1582 if (!hdev_is_powered(hdev) && timeout > 0) { 1583 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1584 MGMT_STATUS_NOT_POWERED); 1585 goto failed; 1586 } 1587 1588 if (pending_find(MGMT_OP_SET_DISCOVERABLE, hdev) || 1589 pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) { 1590 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1591 MGMT_STATUS_BUSY); 1592 goto failed; 1593 } 1594 1595 if (!hci_dev_test_flag(hdev, HCI_CONNECTABLE)) { 1596 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1597 MGMT_STATUS_REJECTED); 1598 goto failed; 1599 } 1600 1601 if (hdev->advertising_paused) { 1602 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE, 1603 MGMT_STATUS_BUSY); 1604 goto failed; 1605 } 1606 1607 if (!hdev_is_powered(hdev)) { 1608 bool changed = false; 1609 1610 /* Setting limited discoverable when powered off is 1611 * not a valid operation since it requires a timeout 1612 * and so no need to check HCI_LIMITED_DISCOVERABLE. 1613 */ 1614 if (!!cp->val != hci_dev_test_flag(hdev, HCI_DISCOVERABLE)) { 1615 hci_dev_change_flag(hdev, HCI_DISCOVERABLE); 1616 changed = true; 1617 } 1618 1619 err = send_settings_rsp(sk, MGMT_OP_SET_DISCOVERABLE, hdev); 1620 if (err < 0) 1621 goto failed; 1622 1623 if (changed) 1624 err = new_settings(hdev, sk); 1625 1626 goto failed; 1627 } 1628 1629 /* If the current mode is the same, then just update the timeout 1630 * value with the new value. And if only the timeout gets updated, 1631 * then no need for any HCI transactions. 1632 */ 1633 if (!!cp->val == hci_dev_test_flag(hdev, HCI_DISCOVERABLE) && 1634 (cp->val == 0x02) == hci_dev_test_flag(hdev, 1635 HCI_LIMITED_DISCOVERABLE)) { 1636 cancel_delayed_work(&hdev->discov_off); 1637 hdev->discov_timeout = timeout; 1638 1639 if (cp->val && hdev->discov_timeout > 0) { 1640 int to = msecs_to_jiffies(hdev->discov_timeout * 1000); 1641 queue_delayed_work(hdev->req_workqueue, 1642 &hdev->discov_off, to); 1643 } 1644 1645 err = send_settings_rsp(sk, MGMT_OP_SET_DISCOVERABLE, hdev); 1646 goto failed; 1647 } 1648 1649 cmd = mgmt_pending_add(sk, MGMT_OP_SET_DISCOVERABLE, hdev, data, len); 1650 if (!cmd) { 1651 err = -ENOMEM; 1652 goto failed; 1653 } 1654 1655 /* Cancel any potential discoverable timeout that might be 1656 * still active and store new timeout value. The arming of 1657 * the timeout happens in the complete handler. 1658 */ 1659 cancel_delayed_work(&hdev->discov_off); 1660 hdev->discov_timeout = timeout; 1661 1662 if (cp->val) 1663 hci_dev_set_flag(hdev, HCI_DISCOVERABLE); 1664 else 1665 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1666 1667 /* Limited discoverable mode */ 1668 if (cp->val == 0x02) 1669 hci_dev_set_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1670 else 1671 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1672 1673 err = hci_cmd_sync_queue(hdev, set_discoverable_sync, cmd, 1674 mgmt_set_discoverable_complete); 1675 1676 if (err < 0) 1677 mgmt_pending_remove(cmd); 1678 1679 failed: 1680 hci_dev_unlock(hdev); 1681 return err; 1682 } 1683 1684 static void mgmt_set_connectable_complete(struct hci_dev *hdev, void *data, 1685 int err) 1686 { 1687 struct mgmt_pending_cmd *cmd = data; 1688 1689 bt_dev_dbg(hdev, "err %d", err); 1690 1691 /* Make sure cmd still outstanding. */ 1692 if (cmd != pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) 1693 return; 1694 1695 hci_dev_lock(hdev); 1696 1697 if (err) { 1698 u8 mgmt_err = mgmt_status(err); 1699 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 1700 goto done; 1701 } 1702 1703 send_settings_rsp(cmd->sk, MGMT_OP_SET_CONNECTABLE, hdev); 1704 new_settings(hdev, cmd->sk); 1705 1706 done: 1707 if (cmd) 1708 mgmt_pending_remove(cmd); 1709 1710 hci_dev_unlock(hdev); 1711 } 1712 1713 static int set_connectable_update_settings(struct hci_dev *hdev, 1714 struct sock *sk, u8 val) 1715 { 1716 bool changed = false; 1717 int err; 1718 1719 if (!!val != hci_dev_test_flag(hdev, HCI_CONNECTABLE)) 1720 changed = true; 1721 1722 if (val) { 1723 hci_dev_set_flag(hdev, HCI_CONNECTABLE); 1724 } else { 1725 hci_dev_clear_flag(hdev, HCI_CONNECTABLE); 1726 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1727 } 1728 1729 err = send_settings_rsp(sk, MGMT_OP_SET_CONNECTABLE, hdev); 1730 if (err < 0) 1731 return err; 1732 1733 if (changed) { 1734 hci_update_scan(hdev); 1735 hci_update_passive_scan(hdev); 1736 return new_settings(hdev, sk); 1737 } 1738 1739 return 0; 1740 } 1741 1742 static int set_connectable_sync(struct hci_dev *hdev, void *data) 1743 { 1744 BT_DBG("%s", hdev->name); 1745 1746 return hci_update_connectable_sync(hdev); 1747 } 1748 1749 static int set_connectable(struct sock *sk, struct hci_dev *hdev, void *data, 1750 u16 len) 1751 { 1752 struct mgmt_mode *cp = data; 1753 struct mgmt_pending_cmd *cmd; 1754 int err; 1755 1756 bt_dev_dbg(hdev, "sock %p", sk); 1757 1758 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) && 1759 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 1760 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE, 1761 MGMT_STATUS_REJECTED); 1762 1763 if (cp->val != 0x00 && cp->val != 0x01) 1764 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE, 1765 MGMT_STATUS_INVALID_PARAMS); 1766 1767 hci_dev_lock(hdev); 1768 1769 if (!hdev_is_powered(hdev)) { 1770 err = set_connectable_update_settings(hdev, sk, cp->val); 1771 goto failed; 1772 } 1773 1774 if (pending_find(MGMT_OP_SET_DISCOVERABLE, hdev) || 1775 pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) { 1776 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE, 1777 MGMT_STATUS_BUSY); 1778 goto failed; 1779 } 1780 1781 cmd = mgmt_pending_add(sk, MGMT_OP_SET_CONNECTABLE, hdev, data, len); 1782 if (!cmd) { 1783 err = -ENOMEM; 1784 goto failed; 1785 } 1786 1787 if (cp->val) { 1788 hci_dev_set_flag(hdev, HCI_CONNECTABLE); 1789 } else { 1790 if (hdev->discov_timeout > 0) 1791 cancel_delayed_work(&hdev->discov_off); 1792 1793 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE); 1794 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 1795 hci_dev_clear_flag(hdev, HCI_CONNECTABLE); 1796 } 1797 1798 err = hci_cmd_sync_queue(hdev, set_connectable_sync, cmd, 1799 mgmt_set_connectable_complete); 1800 1801 if (err < 0) 1802 mgmt_pending_remove(cmd); 1803 1804 failed: 1805 hci_dev_unlock(hdev); 1806 return err; 1807 } 1808 1809 static int set_bondable(struct sock *sk, struct hci_dev *hdev, void *data, 1810 u16 len) 1811 { 1812 struct mgmt_mode *cp = data; 1813 bool changed; 1814 int err; 1815 1816 bt_dev_dbg(hdev, "sock %p", sk); 1817 1818 if (cp->val != 0x00 && cp->val != 0x01) 1819 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BONDABLE, 1820 MGMT_STATUS_INVALID_PARAMS); 1821 1822 hci_dev_lock(hdev); 1823 1824 if (cp->val) 1825 changed = !hci_dev_test_and_set_flag(hdev, HCI_BONDABLE); 1826 else 1827 changed = hci_dev_test_and_clear_flag(hdev, HCI_BONDABLE); 1828 1829 err = send_settings_rsp(sk, MGMT_OP_SET_BONDABLE, hdev); 1830 if (err < 0) 1831 goto unlock; 1832 1833 if (changed) { 1834 /* In limited privacy mode the change of bondable mode 1835 * may affect the local advertising address. 1836 */ 1837 hci_update_discoverable(hdev); 1838 1839 err = new_settings(hdev, sk); 1840 } 1841 1842 unlock: 1843 hci_dev_unlock(hdev); 1844 return err; 1845 } 1846 1847 static int set_link_security(struct sock *sk, struct hci_dev *hdev, void *data, 1848 u16 len) 1849 { 1850 struct mgmt_mode *cp = data; 1851 struct mgmt_pending_cmd *cmd; 1852 u8 val, status; 1853 int err; 1854 1855 bt_dev_dbg(hdev, "sock %p", sk); 1856 1857 status = mgmt_bredr_support(hdev); 1858 if (status) 1859 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY, 1860 status); 1861 1862 if (cp->val != 0x00 && cp->val != 0x01) 1863 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY, 1864 MGMT_STATUS_INVALID_PARAMS); 1865 1866 hci_dev_lock(hdev); 1867 1868 if (!hdev_is_powered(hdev)) { 1869 bool changed = false; 1870 1871 if (!!cp->val != hci_dev_test_flag(hdev, HCI_LINK_SECURITY)) { 1872 hci_dev_change_flag(hdev, HCI_LINK_SECURITY); 1873 changed = true; 1874 } 1875 1876 err = send_settings_rsp(sk, MGMT_OP_SET_LINK_SECURITY, hdev); 1877 if (err < 0) 1878 goto failed; 1879 1880 if (changed) 1881 err = new_settings(hdev, sk); 1882 1883 goto failed; 1884 } 1885 1886 if (pending_find(MGMT_OP_SET_LINK_SECURITY, hdev)) { 1887 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY, 1888 MGMT_STATUS_BUSY); 1889 goto failed; 1890 } 1891 1892 val = !!cp->val; 1893 1894 if (test_bit(HCI_AUTH, &hdev->flags) == val) { 1895 err = send_settings_rsp(sk, MGMT_OP_SET_LINK_SECURITY, hdev); 1896 goto failed; 1897 } 1898 1899 cmd = mgmt_pending_add(sk, MGMT_OP_SET_LINK_SECURITY, hdev, data, len); 1900 if (!cmd) { 1901 err = -ENOMEM; 1902 goto failed; 1903 } 1904 1905 err = hci_send_cmd(hdev, HCI_OP_WRITE_AUTH_ENABLE, sizeof(val), &val); 1906 if (err < 0) { 1907 mgmt_pending_remove(cmd); 1908 goto failed; 1909 } 1910 1911 failed: 1912 hci_dev_unlock(hdev); 1913 return err; 1914 } 1915 1916 static void set_ssp_complete(struct hci_dev *hdev, void *data, int err) 1917 { 1918 struct cmd_lookup match = { NULL, hdev }; 1919 struct mgmt_pending_cmd *cmd = data; 1920 struct mgmt_mode *cp = cmd->param; 1921 u8 enable = cp->val; 1922 bool changed; 1923 1924 /* Make sure cmd still outstanding. */ 1925 if (cmd != pending_find(MGMT_OP_SET_SSP, hdev)) 1926 return; 1927 1928 if (err) { 1929 u8 mgmt_err = mgmt_status(err); 1930 1931 if (enable && hci_dev_test_and_clear_flag(hdev, 1932 HCI_SSP_ENABLED)) { 1933 hci_dev_clear_flag(hdev, HCI_HS_ENABLED); 1934 new_settings(hdev, NULL); 1935 } 1936 1937 mgmt_pending_foreach(MGMT_OP_SET_SSP, hdev, cmd_status_rsp, 1938 &mgmt_err); 1939 return; 1940 } 1941 1942 if (enable) { 1943 changed = !hci_dev_test_and_set_flag(hdev, HCI_SSP_ENABLED); 1944 } else { 1945 changed = hci_dev_test_and_clear_flag(hdev, HCI_SSP_ENABLED); 1946 1947 if (!changed) 1948 changed = hci_dev_test_and_clear_flag(hdev, 1949 HCI_HS_ENABLED); 1950 else 1951 hci_dev_clear_flag(hdev, HCI_HS_ENABLED); 1952 } 1953 1954 mgmt_pending_foreach(MGMT_OP_SET_SSP, hdev, settings_rsp, &match); 1955 1956 if (changed) 1957 new_settings(hdev, match.sk); 1958 1959 if (match.sk) 1960 sock_put(match.sk); 1961 1962 hci_update_eir_sync(hdev); 1963 } 1964 1965 static int set_ssp_sync(struct hci_dev *hdev, void *data) 1966 { 1967 struct mgmt_pending_cmd *cmd = data; 1968 struct mgmt_mode *cp = cmd->param; 1969 bool changed = false; 1970 int err; 1971 1972 if (cp->val) 1973 changed = !hci_dev_test_and_set_flag(hdev, HCI_SSP_ENABLED); 1974 1975 err = hci_write_ssp_mode_sync(hdev, cp->val); 1976 1977 if (!err && changed) 1978 hci_dev_clear_flag(hdev, HCI_SSP_ENABLED); 1979 1980 return err; 1981 } 1982 1983 static int set_ssp(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 1984 { 1985 struct mgmt_mode *cp = data; 1986 struct mgmt_pending_cmd *cmd; 1987 u8 status; 1988 int err; 1989 1990 bt_dev_dbg(hdev, "sock %p", sk); 1991 1992 status = mgmt_bredr_support(hdev); 1993 if (status) 1994 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, status); 1995 1996 if (!lmp_ssp_capable(hdev)) 1997 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 1998 MGMT_STATUS_NOT_SUPPORTED); 1999 2000 if (cp->val != 0x00 && cp->val != 0x01) 2001 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 2002 MGMT_STATUS_INVALID_PARAMS); 2003 2004 hci_dev_lock(hdev); 2005 2006 if (!hdev_is_powered(hdev)) { 2007 bool changed; 2008 2009 if (cp->val) { 2010 changed = !hci_dev_test_and_set_flag(hdev, 2011 HCI_SSP_ENABLED); 2012 } else { 2013 changed = hci_dev_test_and_clear_flag(hdev, 2014 HCI_SSP_ENABLED); 2015 if (!changed) 2016 changed = hci_dev_test_and_clear_flag(hdev, 2017 HCI_HS_ENABLED); 2018 else 2019 hci_dev_clear_flag(hdev, HCI_HS_ENABLED); 2020 } 2021 2022 err = send_settings_rsp(sk, MGMT_OP_SET_SSP, hdev); 2023 if (err < 0) 2024 goto failed; 2025 2026 if (changed) 2027 err = new_settings(hdev, sk); 2028 2029 goto failed; 2030 } 2031 2032 if (pending_find(MGMT_OP_SET_SSP, hdev)) { 2033 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 2034 MGMT_STATUS_BUSY); 2035 goto failed; 2036 } 2037 2038 if (!!cp->val == hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) { 2039 err = send_settings_rsp(sk, MGMT_OP_SET_SSP, hdev); 2040 goto failed; 2041 } 2042 2043 cmd = mgmt_pending_add(sk, MGMT_OP_SET_SSP, hdev, data, len); 2044 if (!cmd) 2045 err = -ENOMEM; 2046 else 2047 err = hci_cmd_sync_queue(hdev, set_ssp_sync, cmd, 2048 set_ssp_complete); 2049 2050 if (err < 0) { 2051 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, 2052 MGMT_STATUS_FAILED); 2053 2054 if (cmd) 2055 mgmt_pending_remove(cmd); 2056 } 2057 2058 failed: 2059 hci_dev_unlock(hdev); 2060 return err; 2061 } 2062 2063 static int set_hs(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2064 { 2065 struct mgmt_mode *cp = data; 2066 bool changed; 2067 u8 status; 2068 int err; 2069 2070 bt_dev_dbg(hdev, "sock %p", sk); 2071 2072 if (!IS_ENABLED(CONFIG_BT_HS)) 2073 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, 2074 MGMT_STATUS_NOT_SUPPORTED); 2075 2076 status = mgmt_bredr_support(hdev); 2077 if (status) 2078 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, status); 2079 2080 if (!lmp_ssp_capable(hdev)) 2081 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, 2082 MGMT_STATUS_NOT_SUPPORTED); 2083 2084 if (!hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) 2085 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, 2086 MGMT_STATUS_REJECTED); 2087 2088 if (cp->val != 0x00 && cp->val != 0x01) 2089 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, 2090 MGMT_STATUS_INVALID_PARAMS); 2091 2092 hci_dev_lock(hdev); 2093 2094 if (pending_find(MGMT_OP_SET_SSP, hdev)) { 2095 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, 2096 MGMT_STATUS_BUSY); 2097 goto unlock; 2098 } 2099 2100 if (cp->val) { 2101 changed = !hci_dev_test_and_set_flag(hdev, HCI_HS_ENABLED); 2102 } else { 2103 if (hdev_is_powered(hdev)) { 2104 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, 2105 MGMT_STATUS_REJECTED); 2106 goto unlock; 2107 } 2108 2109 changed = hci_dev_test_and_clear_flag(hdev, HCI_HS_ENABLED); 2110 } 2111 2112 err = send_settings_rsp(sk, MGMT_OP_SET_HS, hdev); 2113 if (err < 0) 2114 goto unlock; 2115 2116 if (changed) 2117 err = new_settings(hdev, sk); 2118 2119 unlock: 2120 hci_dev_unlock(hdev); 2121 return err; 2122 } 2123 2124 static void set_le_complete(struct hci_dev *hdev, void *data, int err) 2125 { 2126 struct cmd_lookup match = { NULL, hdev }; 2127 u8 status = mgmt_status(err); 2128 2129 bt_dev_dbg(hdev, "err %d", err); 2130 2131 if (status) { 2132 mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, cmd_status_rsp, 2133 &status); 2134 return; 2135 } 2136 2137 mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, settings_rsp, &match); 2138 2139 new_settings(hdev, match.sk); 2140 2141 if (match.sk) 2142 sock_put(match.sk); 2143 } 2144 2145 static int set_le_sync(struct hci_dev *hdev, void *data) 2146 { 2147 struct mgmt_pending_cmd *cmd = data; 2148 struct mgmt_mode *cp = cmd->param; 2149 u8 val = !!cp->val; 2150 int err; 2151 2152 if (!val) { 2153 hci_clear_adv_instance_sync(hdev, NULL, 0x00, true); 2154 2155 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 2156 hci_disable_advertising_sync(hdev); 2157 2158 if (ext_adv_capable(hdev)) 2159 hci_remove_ext_adv_instance_sync(hdev, 0, cmd->sk); 2160 } else { 2161 hci_dev_set_flag(hdev, HCI_LE_ENABLED); 2162 } 2163 2164 err = hci_write_le_host_supported_sync(hdev, val, 0); 2165 2166 /* Make sure the controller has a good default for 2167 * advertising data. Restrict the update to when LE 2168 * has actually been enabled. During power on, the 2169 * update in powered_update_hci will take care of it. 2170 */ 2171 if (!err && hci_dev_test_flag(hdev, HCI_LE_ENABLED)) { 2172 if (ext_adv_capable(hdev)) { 2173 int status; 2174 2175 status = hci_setup_ext_adv_instance_sync(hdev, 0x00); 2176 if (!status) 2177 hci_update_scan_rsp_data_sync(hdev, 0x00); 2178 } else { 2179 hci_update_adv_data_sync(hdev, 0x00); 2180 hci_update_scan_rsp_data_sync(hdev, 0x00); 2181 } 2182 2183 hci_update_passive_scan(hdev); 2184 } 2185 2186 return err; 2187 } 2188 2189 static void set_mesh_complete(struct hci_dev *hdev, void *data, int err) 2190 { 2191 struct mgmt_pending_cmd *cmd = data; 2192 u8 status = mgmt_status(err); 2193 struct sock *sk = cmd->sk; 2194 2195 if (status) { 2196 mgmt_pending_foreach(MGMT_OP_SET_MESH_RECEIVER, hdev, 2197 cmd_status_rsp, &status); 2198 return; 2199 } 2200 2201 mgmt_pending_remove(cmd); 2202 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 0, NULL, 0); 2203 } 2204 2205 static int set_mesh_sync(struct hci_dev *hdev, void *data) 2206 { 2207 struct mgmt_pending_cmd *cmd = data; 2208 struct mgmt_cp_set_mesh *cp = cmd->param; 2209 size_t len = cmd->param_len; 2210 2211 memset(hdev->mesh_ad_types, 0, sizeof(hdev->mesh_ad_types)); 2212 2213 if (cp->enable) 2214 hci_dev_set_flag(hdev, HCI_MESH); 2215 else 2216 hci_dev_clear_flag(hdev, HCI_MESH); 2217 2218 len -= sizeof(*cp); 2219 2220 /* If filters don't fit, forward all adv pkts */ 2221 if (len <= sizeof(hdev->mesh_ad_types)) 2222 memcpy(hdev->mesh_ad_types, cp->ad_types, len); 2223 2224 hci_update_passive_scan_sync(hdev); 2225 return 0; 2226 } 2227 2228 static int set_mesh(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2229 { 2230 struct mgmt_cp_set_mesh *cp = data; 2231 struct mgmt_pending_cmd *cmd; 2232 int err = 0; 2233 2234 bt_dev_dbg(hdev, "sock %p", sk); 2235 2236 if (!lmp_le_capable(hdev) || 2237 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2238 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 2239 MGMT_STATUS_NOT_SUPPORTED); 2240 2241 if (cp->enable != 0x00 && cp->enable != 0x01) 2242 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 2243 MGMT_STATUS_INVALID_PARAMS); 2244 2245 hci_dev_lock(hdev); 2246 2247 cmd = mgmt_pending_add(sk, MGMT_OP_SET_MESH_RECEIVER, hdev, data, len); 2248 if (!cmd) 2249 err = -ENOMEM; 2250 else 2251 err = hci_cmd_sync_queue(hdev, set_mesh_sync, cmd, 2252 set_mesh_complete); 2253 2254 if (err < 0) { 2255 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 2256 MGMT_STATUS_FAILED); 2257 2258 if (cmd) 2259 mgmt_pending_remove(cmd); 2260 } 2261 2262 hci_dev_unlock(hdev); 2263 return err; 2264 } 2265 2266 static void mesh_send_start_complete(struct hci_dev *hdev, void *data, int err) 2267 { 2268 struct mgmt_mesh_tx *mesh_tx = data; 2269 struct mgmt_cp_mesh_send *send = (void *)mesh_tx->param; 2270 unsigned long mesh_send_interval; 2271 u8 mgmt_err = mgmt_status(err); 2272 2273 /* Report any errors here, but don't report completion */ 2274 2275 if (mgmt_err) { 2276 hci_dev_clear_flag(hdev, HCI_MESH_SENDING); 2277 /* Send Complete Error Code for handle */ 2278 mesh_send_complete(hdev, mesh_tx, false); 2279 return; 2280 } 2281 2282 mesh_send_interval = msecs_to_jiffies((send->cnt) * 25); 2283 queue_delayed_work(hdev->req_workqueue, &hdev->mesh_send_done, 2284 mesh_send_interval); 2285 } 2286 2287 static int mesh_send_sync(struct hci_dev *hdev, void *data) 2288 { 2289 struct mgmt_mesh_tx *mesh_tx = data; 2290 struct mgmt_cp_mesh_send *send = (void *)mesh_tx->param; 2291 struct adv_info *adv, *next_instance; 2292 u8 instance = hdev->le_num_of_adv_sets + 1; 2293 u16 timeout, duration; 2294 int err = 0; 2295 2296 if (hdev->le_num_of_adv_sets <= hdev->adv_instance_cnt) 2297 return MGMT_STATUS_BUSY; 2298 2299 timeout = 1000; 2300 duration = send->cnt * INTERVAL_TO_MS(hdev->le_adv_max_interval); 2301 adv = hci_add_adv_instance(hdev, instance, 0, 2302 send->adv_data_len, send->adv_data, 2303 0, NULL, 2304 timeout, duration, 2305 HCI_ADV_TX_POWER_NO_PREFERENCE, 2306 hdev->le_adv_min_interval, 2307 hdev->le_adv_max_interval, 2308 mesh_tx->handle); 2309 2310 if (!IS_ERR(adv)) 2311 mesh_tx->instance = instance; 2312 else 2313 err = PTR_ERR(adv); 2314 2315 if (hdev->cur_adv_instance == instance) { 2316 /* If the currently advertised instance is being changed then 2317 * cancel the current advertising and schedule the next 2318 * instance. If there is only one instance then the overridden 2319 * advertising data will be visible right away. 2320 */ 2321 cancel_adv_timeout(hdev); 2322 2323 next_instance = hci_get_next_instance(hdev, instance); 2324 if (next_instance) 2325 instance = next_instance->instance; 2326 else 2327 instance = 0; 2328 } else if (hdev->adv_instance_timeout) { 2329 /* Immediately advertise the new instance if no other, or 2330 * let it go naturally from queue if ADV is already happening 2331 */ 2332 instance = 0; 2333 } 2334 2335 if (instance) 2336 return hci_schedule_adv_instance_sync(hdev, instance, true); 2337 2338 return err; 2339 } 2340 2341 static void send_count(struct mgmt_mesh_tx *mesh_tx, void *data) 2342 { 2343 struct mgmt_rp_mesh_read_features *rp = data; 2344 2345 if (rp->used_handles >= rp->max_handles) 2346 return; 2347 2348 rp->handles[rp->used_handles++] = mesh_tx->handle; 2349 } 2350 2351 static int mesh_features(struct sock *sk, struct hci_dev *hdev, 2352 void *data, u16 len) 2353 { 2354 struct mgmt_rp_mesh_read_features rp; 2355 2356 if (!lmp_le_capable(hdev) || 2357 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2358 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_READ_FEATURES, 2359 MGMT_STATUS_NOT_SUPPORTED); 2360 2361 memset(&rp, 0, sizeof(rp)); 2362 rp.index = cpu_to_le16(hdev->id); 2363 if (hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 2364 rp.max_handles = MESH_HANDLES_MAX; 2365 2366 hci_dev_lock(hdev); 2367 2368 if (rp.max_handles) 2369 mgmt_mesh_foreach(hdev, send_count, &rp, sk); 2370 2371 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_MESH_READ_FEATURES, 0, &rp, 2372 rp.used_handles + sizeof(rp) - MESH_HANDLES_MAX); 2373 2374 hci_dev_unlock(hdev); 2375 return 0; 2376 } 2377 2378 static int send_cancel(struct hci_dev *hdev, void *data) 2379 { 2380 struct mgmt_pending_cmd *cmd = data; 2381 struct mgmt_cp_mesh_send_cancel *cancel = (void *)cmd->param; 2382 struct mgmt_mesh_tx *mesh_tx; 2383 2384 if (!cancel->handle) { 2385 do { 2386 mesh_tx = mgmt_mesh_next(hdev, cmd->sk); 2387 2388 if (mesh_tx) 2389 mesh_send_complete(hdev, mesh_tx, false); 2390 } while (mesh_tx); 2391 } else { 2392 mesh_tx = mgmt_mesh_find(hdev, cancel->handle); 2393 2394 if (mesh_tx && mesh_tx->sk == cmd->sk) 2395 mesh_send_complete(hdev, mesh_tx, false); 2396 } 2397 2398 mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2399 0, NULL, 0); 2400 mgmt_pending_free(cmd); 2401 2402 return 0; 2403 } 2404 2405 static int mesh_send_cancel(struct sock *sk, struct hci_dev *hdev, 2406 void *data, u16 len) 2407 { 2408 struct mgmt_pending_cmd *cmd; 2409 int err; 2410 2411 if (!lmp_le_capable(hdev) || 2412 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2413 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2414 MGMT_STATUS_NOT_SUPPORTED); 2415 2416 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 2417 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2418 MGMT_STATUS_REJECTED); 2419 2420 hci_dev_lock(hdev); 2421 cmd = mgmt_pending_new(sk, MGMT_OP_MESH_SEND_CANCEL, hdev, data, len); 2422 if (!cmd) 2423 err = -ENOMEM; 2424 else 2425 err = hci_cmd_sync_queue(hdev, send_cancel, cmd, NULL); 2426 2427 if (err < 0) { 2428 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL, 2429 MGMT_STATUS_FAILED); 2430 2431 if (cmd) 2432 mgmt_pending_free(cmd); 2433 } 2434 2435 hci_dev_unlock(hdev); 2436 return err; 2437 } 2438 2439 static int mesh_send(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2440 { 2441 struct mgmt_mesh_tx *mesh_tx; 2442 struct mgmt_cp_mesh_send *send = data; 2443 struct mgmt_rp_mesh_read_features rp; 2444 bool sending; 2445 int err = 0; 2446 2447 if (!lmp_le_capable(hdev) || 2448 !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 2449 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2450 MGMT_STATUS_NOT_SUPPORTED); 2451 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) || 2452 len <= MGMT_MESH_SEND_SIZE || 2453 len > (MGMT_MESH_SEND_SIZE + 31)) 2454 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2455 MGMT_STATUS_REJECTED); 2456 2457 hci_dev_lock(hdev); 2458 2459 memset(&rp, 0, sizeof(rp)); 2460 rp.max_handles = MESH_HANDLES_MAX; 2461 2462 mgmt_mesh_foreach(hdev, send_count, &rp, sk); 2463 2464 if (rp.max_handles <= rp.used_handles) { 2465 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2466 MGMT_STATUS_BUSY); 2467 goto done; 2468 } 2469 2470 sending = hci_dev_test_flag(hdev, HCI_MESH_SENDING); 2471 mesh_tx = mgmt_mesh_add(sk, hdev, send, len); 2472 2473 if (!mesh_tx) 2474 err = -ENOMEM; 2475 else if (!sending) 2476 err = hci_cmd_sync_queue(hdev, mesh_send_sync, mesh_tx, 2477 mesh_send_start_complete); 2478 2479 if (err < 0) { 2480 bt_dev_err(hdev, "Send Mesh Failed %d", err); 2481 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND, 2482 MGMT_STATUS_FAILED); 2483 2484 if (mesh_tx) { 2485 if (sending) 2486 mgmt_mesh_remove(mesh_tx); 2487 } 2488 } else { 2489 hci_dev_set_flag(hdev, HCI_MESH_SENDING); 2490 2491 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_MESH_SEND, 0, 2492 &mesh_tx->handle, 1); 2493 } 2494 2495 done: 2496 hci_dev_unlock(hdev); 2497 return err; 2498 } 2499 2500 static int set_le(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2501 { 2502 struct mgmt_mode *cp = data; 2503 struct mgmt_pending_cmd *cmd; 2504 int err; 2505 u8 val, enabled; 2506 2507 bt_dev_dbg(hdev, "sock %p", sk); 2508 2509 if (!lmp_le_capable(hdev)) 2510 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2511 MGMT_STATUS_NOT_SUPPORTED); 2512 2513 if (cp->val != 0x00 && cp->val != 0x01) 2514 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2515 MGMT_STATUS_INVALID_PARAMS); 2516 2517 /* Bluetooth single mode LE only controllers or dual-mode 2518 * controllers configured as LE only devices, do not allow 2519 * switching LE off. These have either LE enabled explicitly 2520 * or BR/EDR has been previously switched off. 2521 * 2522 * When trying to enable an already enabled LE, then gracefully 2523 * send a positive response. Trying to disable it however will 2524 * result into rejection. 2525 */ 2526 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) { 2527 if (cp->val == 0x01) 2528 return send_settings_rsp(sk, MGMT_OP_SET_LE, hdev); 2529 2530 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2531 MGMT_STATUS_REJECTED); 2532 } 2533 2534 hci_dev_lock(hdev); 2535 2536 val = !!cp->val; 2537 enabled = lmp_host_le_capable(hdev); 2538 2539 if (!hdev_is_powered(hdev) || val == enabled) { 2540 bool changed = false; 2541 2542 if (val != hci_dev_test_flag(hdev, HCI_LE_ENABLED)) { 2543 hci_dev_change_flag(hdev, HCI_LE_ENABLED); 2544 changed = true; 2545 } 2546 2547 if (!val && hci_dev_test_flag(hdev, HCI_ADVERTISING)) { 2548 hci_dev_clear_flag(hdev, HCI_ADVERTISING); 2549 changed = true; 2550 } 2551 2552 err = send_settings_rsp(sk, MGMT_OP_SET_LE, hdev); 2553 if (err < 0) 2554 goto unlock; 2555 2556 if (changed) 2557 err = new_settings(hdev, sk); 2558 2559 goto unlock; 2560 } 2561 2562 if (pending_find(MGMT_OP_SET_LE, hdev) || 2563 pending_find(MGMT_OP_SET_ADVERTISING, hdev)) { 2564 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2565 MGMT_STATUS_BUSY); 2566 goto unlock; 2567 } 2568 2569 cmd = mgmt_pending_add(sk, MGMT_OP_SET_LE, hdev, data, len); 2570 if (!cmd) 2571 err = -ENOMEM; 2572 else 2573 err = hci_cmd_sync_queue(hdev, set_le_sync, cmd, 2574 set_le_complete); 2575 2576 if (err < 0) { 2577 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE, 2578 MGMT_STATUS_FAILED); 2579 2580 if (cmd) 2581 mgmt_pending_remove(cmd); 2582 } 2583 2584 unlock: 2585 hci_dev_unlock(hdev); 2586 return err; 2587 } 2588 2589 /* This is a helper function to test for pending mgmt commands that can 2590 * cause CoD or EIR HCI commands. We can only allow one such pending 2591 * mgmt command at a time since otherwise we cannot easily track what 2592 * the current values are, will be, and based on that calculate if a new 2593 * HCI command needs to be sent and if yes with what value. 2594 */ 2595 static bool pending_eir_or_class(struct hci_dev *hdev) 2596 { 2597 struct mgmt_pending_cmd *cmd; 2598 2599 list_for_each_entry(cmd, &hdev->mgmt_pending, list) { 2600 switch (cmd->opcode) { 2601 case MGMT_OP_ADD_UUID: 2602 case MGMT_OP_REMOVE_UUID: 2603 case MGMT_OP_SET_DEV_CLASS: 2604 case MGMT_OP_SET_POWERED: 2605 return true; 2606 } 2607 } 2608 2609 return false; 2610 } 2611 2612 static const u8 bluetooth_base_uuid[] = { 2613 0xfb, 0x34, 0x9b, 0x5f, 0x80, 0x00, 0x00, 0x80, 2614 0x00, 0x10, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 2615 }; 2616 2617 static u8 get_uuid_size(const u8 *uuid) 2618 { 2619 u32 val; 2620 2621 if (memcmp(uuid, bluetooth_base_uuid, 12)) 2622 return 128; 2623 2624 val = get_unaligned_le32(&uuid[12]); 2625 if (val > 0xffff) 2626 return 32; 2627 2628 return 16; 2629 } 2630 2631 static void mgmt_class_complete(struct hci_dev *hdev, void *data, int err) 2632 { 2633 struct mgmt_pending_cmd *cmd = data; 2634 2635 bt_dev_dbg(hdev, "err %d", err); 2636 2637 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 2638 mgmt_status(err), hdev->dev_class, 3); 2639 2640 mgmt_pending_free(cmd); 2641 } 2642 2643 static int add_uuid_sync(struct hci_dev *hdev, void *data) 2644 { 2645 int err; 2646 2647 err = hci_update_class_sync(hdev); 2648 if (err) 2649 return err; 2650 2651 return hci_update_eir_sync(hdev); 2652 } 2653 2654 static int add_uuid(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 2655 { 2656 struct mgmt_cp_add_uuid *cp = data; 2657 struct mgmt_pending_cmd *cmd; 2658 struct bt_uuid *uuid; 2659 int err; 2660 2661 bt_dev_dbg(hdev, "sock %p", sk); 2662 2663 hci_dev_lock(hdev); 2664 2665 if (pending_eir_or_class(hdev)) { 2666 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_UUID, 2667 MGMT_STATUS_BUSY); 2668 goto failed; 2669 } 2670 2671 uuid = kmalloc(sizeof(*uuid), GFP_KERNEL); 2672 if (!uuid) { 2673 err = -ENOMEM; 2674 goto failed; 2675 } 2676 2677 memcpy(uuid->uuid, cp->uuid, 16); 2678 uuid->svc_hint = cp->svc_hint; 2679 uuid->size = get_uuid_size(cp->uuid); 2680 2681 list_add_tail(&uuid->list, &hdev->uuids); 2682 2683 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_UUID, hdev, data, len); 2684 if (!cmd) { 2685 err = -ENOMEM; 2686 goto failed; 2687 } 2688 2689 err = hci_cmd_sync_queue(hdev, add_uuid_sync, cmd, mgmt_class_complete); 2690 if (err < 0) { 2691 mgmt_pending_free(cmd); 2692 goto failed; 2693 } 2694 2695 failed: 2696 hci_dev_unlock(hdev); 2697 return err; 2698 } 2699 2700 static bool enable_service_cache(struct hci_dev *hdev) 2701 { 2702 if (!hdev_is_powered(hdev)) 2703 return false; 2704 2705 if (!hci_dev_test_and_set_flag(hdev, HCI_SERVICE_CACHE)) { 2706 queue_delayed_work(hdev->workqueue, &hdev->service_cache, 2707 CACHE_TIMEOUT); 2708 return true; 2709 } 2710 2711 return false; 2712 } 2713 2714 static int remove_uuid_sync(struct hci_dev *hdev, void *data) 2715 { 2716 int err; 2717 2718 err = hci_update_class_sync(hdev); 2719 if (err) 2720 return err; 2721 2722 return hci_update_eir_sync(hdev); 2723 } 2724 2725 static int remove_uuid(struct sock *sk, struct hci_dev *hdev, void *data, 2726 u16 len) 2727 { 2728 struct mgmt_cp_remove_uuid *cp = data; 2729 struct mgmt_pending_cmd *cmd; 2730 struct bt_uuid *match, *tmp; 2731 static const u8 bt_uuid_any[] = { 2732 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 2733 }; 2734 int err, found; 2735 2736 bt_dev_dbg(hdev, "sock %p", sk); 2737 2738 hci_dev_lock(hdev); 2739 2740 if (pending_eir_or_class(hdev)) { 2741 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_UUID, 2742 MGMT_STATUS_BUSY); 2743 goto unlock; 2744 } 2745 2746 if (memcmp(cp->uuid, bt_uuid_any, 16) == 0) { 2747 hci_uuids_clear(hdev); 2748 2749 if (enable_service_cache(hdev)) { 2750 err = mgmt_cmd_complete(sk, hdev->id, 2751 MGMT_OP_REMOVE_UUID, 2752 0, hdev->dev_class, 3); 2753 goto unlock; 2754 } 2755 2756 goto update_class; 2757 } 2758 2759 found = 0; 2760 2761 list_for_each_entry_safe(match, tmp, &hdev->uuids, list) { 2762 if (memcmp(match->uuid, cp->uuid, 16) != 0) 2763 continue; 2764 2765 list_del(&match->list); 2766 kfree(match); 2767 found++; 2768 } 2769 2770 if (found == 0) { 2771 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_UUID, 2772 MGMT_STATUS_INVALID_PARAMS); 2773 goto unlock; 2774 } 2775 2776 update_class: 2777 cmd = mgmt_pending_new(sk, MGMT_OP_REMOVE_UUID, hdev, data, len); 2778 if (!cmd) { 2779 err = -ENOMEM; 2780 goto unlock; 2781 } 2782 2783 err = hci_cmd_sync_queue(hdev, remove_uuid_sync, cmd, 2784 mgmt_class_complete); 2785 if (err < 0) 2786 mgmt_pending_free(cmd); 2787 2788 unlock: 2789 hci_dev_unlock(hdev); 2790 return err; 2791 } 2792 2793 static int set_class_sync(struct hci_dev *hdev, void *data) 2794 { 2795 int err = 0; 2796 2797 if (hci_dev_test_and_clear_flag(hdev, HCI_SERVICE_CACHE)) { 2798 cancel_delayed_work_sync(&hdev->service_cache); 2799 err = hci_update_eir_sync(hdev); 2800 } 2801 2802 if (err) 2803 return err; 2804 2805 return hci_update_class_sync(hdev); 2806 } 2807 2808 static int set_dev_class(struct sock *sk, struct hci_dev *hdev, void *data, 2809 u16 len) 2810 { 2811 struct mgmt_cp_set_dev_class *cp = data; 2812 struct mgmt_pending_cmd *cmd; 2813 int err; 2814 2815 bt_dev_dbg(hdev, "sock %p", sk); 2816 2817 if (!lmp_bredr_capable(hdev)) 2818 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 2819 MGMT_STATUS_NOT_SUPPORTED); 2820 2821 hci_dev_lock(hdev); 2822 2823 if (pending_eir_or_class(hdev)) { 2824 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 2825 MGMT_STATUS_BUSY); 2826 goto unlock; 2827 } 2828 2829 if ((cp->minor & 0x03) != 0 || (cp->major & 0xe0) != 0) { 2830 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 2831 MGMT_STATUS_INVALID_PARAMS); 2832 goto unlock; 2833 } 2834 2835 hdev->major_class = cp->major; 2836 hdev->minor_class = cp->minor; 2837 2838 if (!hdev_is_powered(hdev)) { 2839 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 0, 2840 hdev->dev_class, 3); 2841 goto unlock; 2842 } 2843 2844 cmd = mgmt_pending_new(sk, MGMT_OP_SET_DEV_CLASS, hdev, data, len); 2845 if (!cmd) { 2846 err = -ENOMEM; 2847 goto unlock; 2848 } 2849 2850 err = hci_cmd_sync_queue(hdev, set_class_sync, cmd, 2851 mgmt_class_complete); 2852 if (err < 0) 2853 mgmt_pending_free(cmd); 2854 2855 unlock: 2856 hci_dev_unlock(hdev); 2857 return err; 2858 } 2859 2860 static int load_link_keys(struct sock *sk, struct hci_dev *hdev, void *data, 2861 u16 len) 2862 { 2863 struct mgmt_cp_load_link_keys *cp = data; 2864 const u16 max_key_count = ((U16_MAX - sizeof(*cp)) / 2865 sizeof(struct mgmt_link_key_info)); 2866 u16 key_count, expected_len; 2867 bool changed; 2868 int i; 2869 2870 bt_dev_dbg(hdev, "sock %p", sk); 2871 2872 if (!lmp_bredr_capable(hdev)) 2873 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2874 MGMT_STATUS_NOT_SUPPORTED); 2875 2876 key_count = __le16_to_cpu(cp->key_count); 2877 if (key_count > max_key_count) { 2878 bt_dev_err(hdev, "load_link_keys: too big key_count value %u", 2879 key_count); 2880 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2881 MGMT_STATUS_INVALID_PARAMS); 2882 } 2883 2884 expected_len = struct_size(cp, keys, key_count); 2885 if (expected_len != len) { 2886 bt_dev_err(hdev, "load_link_keys: expected %u bytes, got %u bytes", 2887 expected_len, len); 2888 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2889 MGMT_STATUS_INVALID_PARAMS); 2890 } 2891 2892 if (cp->debug_keys != 0x00 && cp->debug_keys != 0x01) 2893 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 2894 MGMT_STATUS_INVALID_PARAMS); 2895 2896 bt_dev_dbg(hdev, "debug_keys %u key_count %u", cp->debug_keys, 2897 key_count); 2898 2899 for (i = 0; i < key_count; i++) { 2900 struct mgmt_link_key_info *key = &cp->keys[i]; 2901 2902 /* Considering SMP over BREDR/LE, there is no need to check addr_type */ 2903 if (key->type > 0x08) 2904 return mgmt_cmd_status(sk, hdev->id, 2905 MGMT_OP_LOAD_LINK_KEYS, 2906 MGMT_STATUS_INVALID_PARAMS); 2907 } 2908 2909 hci_dev_lock(hdev); 2910 2911 hci_link_keys_clear(hdev); 2912 2913 if (cp->debug_keys) 2914 changed = !hci_dev_test_and_set_flag(hdev, HCI_KEEP_DEBUG_KEYS); 2915 else 2916 changed = hci_dev_test_and_clear_flag(hdev, 2917 HCI_KEEP_DEBUG_KEYS); 2918 2919 if (changed) 2920 new_settings(hdev, NULL); 2921 2922 for (i = 0; i < key_count; i++) { 2923 struct mgmt_link_key_info *key = &cp->keys[i]; 2924 2925 if (hci_is_blocked_key(hdev, 2926 HCI_BLOCKED_KEY_TYPE_LINKKEY, 2927 key->val)) { 2928 bt_dev_warn(hdev, "Skipping blocked link key for %pMR", 2929 &key->addr.bdaddr); 2930 continue; 2931 } 2932 2933 /* Always ignore debug keys and require a new pairing if 2934 * the user wants to use them. 2935 */ 2936 if (key->type == HCI_LK_DEBUG_COMBINATION) 2937 continue; 2938 2939 hci_add_link_key(hdev, NULL, &key->addr.bdaddr, key->val, 2940 key->type, key->pin_len, NULL); 2941 } 2942 2943 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 0, NULL, 0); 2944 2945 hci_dev_unlock(hdev); 2946 2947 return 0; 2948 } 2949 2950 static int device_unpaired(struct hci_dev *hdev, bdaddr_t *bdaddr, 2951 u8 addr_type, struct sock *skip_sk) 2952 { 2953 struct mgmt_ev_device_unpaired ev; 2954 2955 bacpy(&ev.addr.bdaddr, bdaddr); 2956 ev.addr.type = addr_type; 2957 2958 return mgmt_event(MGMT_EV_DEVICE_UNPAIRED, hdev, &ev, sizeof(ev), 2959 skip_sk); 2960 } 2961 2962 static void unpair_device_complete(struct hci_dev *hdev, void *data, int err) 2963 { 2964 struct mgmt_pending_cmd *cmd = data; 2965 struct mgmt_cp_unpair_device *cp = cmd->param; 2966 2967 if (!err) 2968 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, cmd->sk); 2969 2970 cmd->cmd_complete(cmd, err); 2971 mgmt_pending_free(cmd); 2972 } 2973 2974 static int unpair_device_sync(struct hci_dev *hdev, void *data) 2975 { 2976 struct mgmt_pending_cmd *cmd = data; 2977 struct mgmt_cp_unpair_device *cp = cmd->param; 2978 struct hci_conn *conn; 2979 2980 if (cp->addr.type == BDADDR_BREDR) 2981 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 2982 &cp->addr.bdaddr); 2983 else 2984 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr, 2985 le_addr_type(cp->addr.type)); 2986 2987 if (!conn) 2988 return 0; 2989 2990 return hci_abort_conn_sync(hdev, conn, HCI_ERROR_REMOTE_USER_TERM); 2991 } 2992 2993 static int unpair_device(struct sock *sk, struct hci_dev *hdev, void *data, 2994 u16 len) 2995 { 2996 struct mgmt_cp_unpair_device *cp = data; 2997 struct mgmt_rp_unpair_device rp; 2998 struct hci_conn_params *params; 2999 struct mgmt_pending_cmd *cmd; 3000 struct hci_conn *conn; 3001 u8 addr_type; 3002 int err; 3003 3004 memset(&rp, 0, sizeof(rp)); 3005 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 3006 rp.addr.type = cp->addr.type; 3007 3008 if (!bdaddr_type_is_valid(cp->addr.type)) 3009 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 3010 MGMT_STATUS_INVALID_PARAMS, 3011 &rp, sizeof(rp)); 3012 3013 if (cp->disconnect != 0x00 && cp->disconnect != 0x01) 3014 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 3015 MGMT_STATUS_INVALID_PARAMS, 3016 &rp, sizeof(rp)); 3017 3018 hci_dev_lock(hdev); 3019 3020 if (!hdev_is_powered(hdev)) { 3021 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 3022 MGMT_STATUS_NOT_POWERED, &rp, 3023 sizeof(rp)); 3024 goto unlock; 3025 } 3026 3027 if (cp->addr.type == BDADDR_BREDR) { 3028 /* If disconnection is requested, then look up the 3029 * connection. If the remote device is connected, it 3030 * will be later used to terminate the link. 3031 * 3032 * Setting it to NULL explicitly will cause no 3033 * termination of the link. 3034 */ 3035 if (cp->disconnect) 3036 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 3037 &cp->addr.bdaddr); 3038 else 3039 conn = NULL; 3040 3041 err = hci_remove_link_key(hdev, &cp->addr.bdaddr); 3042 if (err < 0) { 3043 err = mgmt_cmd_complete(sk, hdev->id, 3044 MGMT_OP_UNPAIR_DEVICE, 3045 MGMT_STATUS_NOT_PAIRED, &rp, 3046 sizeof(rp)); 3047 goto unlock; 3048 } 3049 3050 goto done; 3051 } 3052 3053 /* LE address type */ 3054 addr_type = le_addr_type(cp->addr.type); 3055 3056 /* Abort any ongoing SMP pairing. Removes ltk and irk if they exist. */ 3057 err = smp_cancel_and_remove_pairing(hdev, &cp->addr.bdaddr, addr_type); 3058 if (err < 0) { 3059 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 3060 MGMT_STATUS_NOT_PAIRED, &rp, 3061 sizeof(rp)); 3062 goto unlock; 3063 } 3064 3065 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr, addr_type); 3066 if (!conn) { 3067 hci_conn_params_del(hdev, &cp->addr.bdaddr, addr_type); 3068 goto done; 3069 } 3070 3071 3072 /* Defer clearing up the connection parameters until closing to 3073 * give a chance of keeping them if a repairing happens. 3074 */ 3075 set_bit(HCI_CONN_PARAM_REMOVAL_PEND, &conn->flags); 3076 3077 /* Disable auto-connection parameters if present */ 3078 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, addr_type); 3079 if (params) { 3080 if (params->explicit_connect) 3081 params->auto_connect = HCI_AUTO_CONN_EXPLICIT; 3082 else 3083 params->auto_connect = HCI_AUTO_CONN_DISABLED; 3084 } 3085 3086 /* If disconnection is not requested, then clear the connection 3087 * variable so that the link is not terminated. 3088 */ 3089 if (!cp->disconnect) 3090 conn = NULL; 3091 3092 done: 3093 /* If the connection variable is set, then termination of the 3094 * link is requested. 3095 */ 3096 if (!conn) { 3097 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 0, 3098 &rp, sizeof(rp)); 3099 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, sk); 3100 goto unlock; 3101 } 3102 3103 cmd = mgmt_pending_new(sk, MGMT_OP_UNPAIR_DEVICE, hdev, cp, 3104 sizeof(*cp)); 3105 if (!cmd) { 3106 err = -ENOMEM; 3107 goto unlock; 3108 } 3109 3110 cmd->cmd_complete = addr_cmd_complete; 3111 3112 err = hci_cmd_sync_queue(hdev, unpair_device_sync, cmd, 3113 unpair_device_complete); 3114 if (err < 0) 3115 mgmt_pending_free(cmd); 3116 3117 unlock: 3118 hci_dev_unlock(hdev); 3119 return err; 3120 } 3121 3122 static int disconnect(struct sock *sk, struct hci_dev *hdev, void *data, 3123 u16 len) 3124 { 3125 struct mgmt_cp_disconnect *cp = data; 3126 struct mgmt_rp_disconnect rp; 3127 struct mgmt_pending_cmd *cmd; 3128 struct hci_conn *conn; 3129 int err; 3130 3131 bt_dev_dbg(hdev, "sock %p", sk); 3132 3133 memset(&rp, 0, sizeof(rp)); 3134 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 3135 rp.addr.type = cp->addr.type; 3136 3137 if (!bdaddr_type_is_valid(cp->addr.type)) 3138 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT, 3139 MGMT_STATUS_INVALID_PARAMS, 3140 &rp, sizeof(rp)); 3141 3142 hci_dev_lock(hdev); 3143 3144 if (!test_bit(HCI_UP, &hdev->flags)) { 3145 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT, 3146 MGMT_STATUS_NOT_POWERED, &rp, 3147 sizeof(rp)); 3148 goto failed; 3149 } 3150 3151 if (pending_find(MGMT_OP_DISCONNECT, hdev)) { 3152 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT, 3153 MGMT_STATUS_BUSY, &rp, sizeof(rp)); 3154 goto failed; 3155 } 3156 3157 if (cp->addr.type == BDADDR_BREDR) 3158 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 3159 &cp->addr.bdaddr); 3160 else 3161 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr, 3162 le_addr_type(cp->addr.type)); 3163 3164 if (!conn || conn->state == BT_OPEN || conn->state == BT_CLOSED) { 3165 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT, 3166 MGMT_STATUS_NOT_CONNECTED, &rp, 3167 sizeof(rp)); 3168 goto failed; 3169 } 3170 3171 cmd = mgmt_pending_add(sk, MGMT_OP_DISCONNECT, hdev, data, len); 3172 if (!cmd) { 3173 err = -ENOMEM; 3174 goto failed; 3175 } 3176 3177 cmd->cmd_complete = generic_cmd_complete; 3178 3179 err = hci_disconnect(conn, HCI_ERROR_REMOTE_USER_TERM); 3180 if (err < 0) 3181 mgmt_pending_remove(cmd); 3182 3183 failed: 3184 hci_dev_unlock(hdev); 3185 return err; 3186 } 3187 3188 static u8 link_to_bdaddr(u8 link_type, u8 addr_type) 3189 { 3190 switch (link_type) { 3191 case LE_LINK: 3192 switch (addr_type) { 3193 case ADDR_LE_DEV_PUBLIC: 3194 return BDADDR_LE_PUBLIC; 3195 3196 default: 3197 /* Fallback to LE Random address type */ 3198 return BDADDR_LE_RANDOM; 3199 } 3200 3201 default: 3202 /* Fallback to BR/EDR type */ 3203 return BDADDR_BREDR; 3204 } 3205 } 3206 3207 static int get_connections(struct sock *sk, struct hci_dev *hdev, void *data, 3208 u16 data_len) 3209 { 3210 struct mgmt_rp_get_connections *rp; 3211 struct hci_conn *c; 3212 int err; 3213 u16 i; 3214 3215 bt_dev_dbg(hdev, "sock %p", sk); 3216 3217 hci_dev_lock(hdev); 3218 3219 if (!hdev_is_powered(hdev)) { 3220 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_CONNECTIONS, 3221 MGMT_STATUS_NOT_POWERED); 3222 goto unlock; 3223 } 3224 3225 i = 0; 3226 list_for_each_entry(c, &hdev->conn_hash.list, list) { 3227 if (test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags)) 3228 i++; 3229 } 3230 3231 rp = kmalloc(struct_size(rp, addr, i), GFP_KERNEL); 3232 if (!rp) { 3233 err = -ENOMEM; 3234 goto unlock; 3235 } 3236 3237 i = 0; 3238 list_for_each_entry(c, &hdev->conn_hash.list, list) { 3239 if (!test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags)) 3240 continue; 3241 bacpy(&rp->addr[i].bdaddr, &c->dst); 3242 rp->addr[i].type = link_to_bdaddr(c->type, c->dst_type); 3243 if (c->type == SCO_LINK || c->type == ESCO_LINK) 3244 continue; 3245 i++; 3246 } 3247 3248 rp->conn_count = cpu_to_le16(i); 3249 3250 /* Recalculate length in case of filtered SCO connections, etc */ 3251 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONNECTIONS, 0, rp, 3252 struct_size(rp, addr, i)); 3253 3254 kfree(rp); 3255 3256 unlock: 3257 hci_dev_unlock(hdev); 3258 return err; 3259 } 3260 3261 static int send_pin_code_neg_reply(struct sock *sk, struct hci_dev *hdev, 3262 struct mgmt_cp_pin_code_neg_reply *cp) 3263 { 3264 struct mgmt_pending_cmd *cmd; 3265 int err; 3266 3267 cmd = mgmt_pending_add(sk, MGMT_OP_PIN_CODE_NEG_REPLY, hdev, cp, 3268 sizeof(*cp)); 3269 if (!cmd) 3270 return -ENOMEM; 3271 3272 cmd->cmd_complete = addr_cmd_complete; 3273 3274 err = hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY, 3275 sizeof(cp->addr.bdaddr), &cp->addr.bdaddr); 3276 if (err < 0) 3277 mgmt_pending_remove(cmd); 3278 3279 return err; 3280 } 3281 3282 static int pin_code_reply(struct sock *sk, struct hci_dev *hdev, void *data, 3283 u16 len) 3284 { 3285 struct hci_conn *conn; 3286 struct mgmt_cp_pin_code_reply *cp = data; 3287 struct hci_cp_pin_code_reply reply; 3288 struct mgmt_pending_cmd *cmd; 3289 int err; 3290 3291 bt_dev_dbg(hdev, "sock %p", sk); 3292 3293 hci_dev_lock(hdev); 3294 3295 if (!hdev_is_powered(hdev)) { 3296 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, 3297 MGMT_STATUS_NOT_POWERED); 3298 goto failed; 3299 } 3300 3301 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr); 3302 if (!conn) { 3303 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, 3304 MGMT_STATUS_NOT_CONNECTED); 3305 goto failed; 3306 } 3307 3308 if (conn->pending_sec_level == BT_SECURITY_HIGH && cp->pin_len != 16) { 3309 struct mgmt_cp_pin_code_neg_reply ncp; 3310 3311 memcpy(&ncp.addr, &cp->addr, sizeof(ncp.addr)); 3312 3313 bt_dev_err(hdev, "PIN code is not 16 bytes long"); 3314 3315 err = send_pin_code_neg_reply(sk, hdev, &ncp); 3316 if (err >= 0) 3317 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY, 3318 MGMT_STATUS_INVALID_PARAMS); 3319 3320 goto failed; 3321 } 3322 3323 cmd = mgmt_pending_add(sk, MGMT_OP_PIN_CODE_REPLY, hdev, data, len); 3324 if (!cmd) { 3325 err = -ENOMEM; 3326 goto failed; 3327 } 3328 3329 cmd->cmd_complete = addr_cmd_complete; 3330 3331 bacpy(&reply.bdaddr, &cp->addr.bdaddr); 3332 reply.pin_len = cp->pin_len; 3333 memcpy(reply.pin_code, cp->pin_code, sizeof(reply.pin_code)); 3334 3335 err = hci_send_cmd(hdev, HCI_OP_PIN_CODE_REPLY, sizeof(reply), &reply); 3336 if (err < 0) 3337 mgmt_pending_remove(cmd); 3338 3339 failed: 3340 hci_dev_unlock(hdev); 3341 return err; 3342 } 3343 3344 static int set_io_capability(struct sock *sk, struct hci_dev *hdev, void *data, 3345 u16 len) 3346 { 3347 struct mgmt_cp_set_io_capability *cp = data; 3348 3349 bt_dev_dbg(hdev, "sock %p", sk); 3350 3351 if (cp->io_capability > SMP_IO_KEYBOARD_DISPLAY) 3352 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_IO_CAPABILITY, 3353 MGMT_STATUS_INVALID_PARAMS); 3354 3355 hci_dev_lock(hdev); 3356 3357 hdev->io_capability = cp->io_capability; 3358 3359 bt_dev_dbg(hdev, "IO capability set to 0x%02x", hdev->io_capability); 3360 3361 hci_dev_unlock(hdev); 3362 3363 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_IO_CAPABILITY, 0, 3364 NULL, 0); 3365 } 3366 3367 static struct mgmt_pending_cmd *find_pairing(struct hci_conn *conn) 3368 { 3369 struct hci_dev *hdev = conn->hdev; 3370 struct mgmt_pending_cmd *cmd; 3371 3372 list_for_each_entry(cmd, &hdev->mgmt_pending, list) { 3373 if (cmd->opcode != MGMT_OP_PAIR_DEVICE) 3374 continue; 3375 3376 if (cmd->user_data != conn) 3377 continue; 3378 3379 return cmd; 3380 } 3381 3382 return NULL; 3383 } 3384 3385 static int pairing_complete(struct mgmt_pending_cmd *cmd, u8 status) 3386 { 3387 struct mgmt_rp_pair_device rp; 3388 struct hci_conn *conn = cmd->user_data; 3389 int err; 3390 3391 bacpy(&rp.addr.bdaddr, &conn->dst); 3392 rp.addr.type = link_to_bdaddr(conn->type, conn->dst_type); 3393 3394 err = mgmt_cmd_complete(cmd->sk, cmd->index, MGMT_OP_PAIR_DEVICE, 3395 status, &rp, sizeof(rp)); 3396 3397 /* So we don't get further callbacks for this connection */ 3398 conn->connect_cfm_cb = NULL; 3399 conn->security_cfm_cb = NULL; 3400 conn->disconn_cfm_cb = NULL; 3401 3402 hci_conn_drop(conn); 3403 3404 /* The device is paired so there is no need to remove 3405 * its connection parameters anymore. 3406 */ 3407 clear_bit(HCI_CONN_PARAM_REMOVAL_PEND, &conn->flags); 3408 3409 hci_conn_put(conn); 3410 3411 return err; 3412 } 3413 3414 void mgmt_smp_complete(struct hci_conn *conn, bool complete) 3415 { 3416 u8 status = complete ? MGMT_STATUS_SUCCESS : MGMT_STATUS_FAILED; 3417 struct mgmt_pending_cmd *cmd; 3418 3419 cmd = find_pairing(conn); 3420 if (cmd) { 3421 cmd->cmd_complete(cmd, status); 3422 mgmt_pending_remove(cmd); 3423 } 3424 } 3425 3426 static void pairing_complete_cb(struct hci_conn *conn, u8 status) 3427 { 3428 struct mgmt_pending_cmd *cmd; 3429 3430 BT_DBG("status %u", status); 3431 3432 cmd = find_pairing(conn); 3433 if (!cmd) { 3434 BT_DBG("Unable to find a pending command"); 3435 return; 3436 } 3437 3438 cmd->cmd_complete(cmd, mgmt_status(status)); 3439 mgmt_pending_remove(cmd); 3440 } 3441 3442 static void le_pairing_complete_cb(struct hci_conn *conn, u8 status) 3443 { 3444 struct mgmt_pending_cmd *cmd; 3445 3446 BT_DBG("status %u", status); 3447 3448 if (!status) 3449 return; 3450 3451 cmd = find_pairing(conn); 3452 if (!cmd) { 3453 BT_DBG("Unable to find a pending command"); 3454 return; 3455 } 3456 3457 cmd->cmd_complete(cmd, mgmt_status(status)); 3458 mgmt_pending_remove(cmd); 3459 } 3460 3461 static int pair_device(struct sock *sk, struct hci_dev *hdev, void *data, 3462 u16 len) 3463 { 3464 struct mgmt_cp_pair_device *cp = data; 3465 struct mgmt_rp_pair_device rp; 3466 struct mgmt_pending_cmd *cmd; 3467 u8 sec_level, auth_type; 3468 struct hci_conn *conn; 3469 int err; 3470 3471 bt_dev_dbg(hdev, "sock %p", sk); 3472 3473 memset(&rp, 0, sizeof(rp)); 3474 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 3475 rp.addr.type = cp->addr.type; 3476 3477 if (!bdaddr_type_is_valid(cp->addr.type)) 3478 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3479 MGMT_STATUS_INVALID_PARAMS, 3480 &rp, sizeof(rp)); 3481 3482 if (cp->io_cap > SMP_IO_KEYBOARD_DISPLAY) 3483 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3484 MGMT_STATUS_INVALID_PARAMS, 3485 &rp, sizeof(rp)); 3486 3487 hci_dev_lock(hdev); 3488 3489 if (!hdev_is_powered(hdev)) { 3490 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3491 MGMT_STATUS_NOT_POWERED, &rp, 3492 sizeof(rp)); 3493 goto unlock; 3494 } 3495 3496 if (hci_bdaddr_is_paired(hdev, &cp->addr.bdaddr, cp->addr.type)) { 3497 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3498 MGMT_STATUS_ALREADY_PAIRED, &rp, 3499 sizeof(rp)); 3500 goto unlock; 3501 } 3502 3503 sec_level = BT_SECURITY_MEDIUM; 3504 auth_type = HCI_AT_DEDICATED_BONDING; 3505 3506 if (cp->addr.type == BDADDR_BREDR) { 3507 conn = hci_connect_acl(hdev, &cp->addr.bdaddr, sec_level, 3508 auth_type, CONN_REASON_PAIR_DEVICE); 3509 } else { 3510 u8 addr_type = le_addr_type(cp->addr.type); 3511 struct hci_conn_params *p; 3512 3513 /* When pairing a new device, it is expected to remember 3514 * this device for future connections. Adding the connection 3515 * parameter information ahead of time allows tracking 3516 * of the peripheral preferred values and will speed up any 3517 * further connection establishment. 3518 * 3519 * If connection parameters already exist, then they 3520 * will be kept and this function does nothing. 3521 */ 3522 p = hci_conn_params_add(hdev, &cp->addr.bdaddr, addr_type); 3523 3524 if (p->auto_connect == HCI_AUTO_CONN_EXPLICIT) 3525 p->auto_connect = HCI_AUTO_CONN_DISABLED; 3526 3527 conn = hci_connect_le_scan(hdev, &cp->addr.bdaddr, addr_type, 3528 sec_level, HCI_LE_CONN_TIMEOUT, 3529 CONN_REASON_PAIR_DEVICE); 3530 } 3531 3532 if (IS_ERR(conn)) { 3533 int status; 3534 3535 if (PTR_ERR(conn) == -EBUSY) 3536 status = MGMT_STATUS_BUSY; 3537 else if (PTR_ERR(conn) == -EOPNOTSUPP) 3538 status = MGMT_STATUS_NOT_SUPPORTED; 3539 else if (PTR_ERR(conn) == -ECONNREFUSED) 3540 status = MGMT_STATUS_REJECTED; 3541 else 3542 status = MGMT_STATUS_CONNECT_FAILED; 3543 3544 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3545 status, &rp, sizeof(rp)); 3546 goto unlock; 3547 } 3548 3549 if (conn->connect_cfm_cb) { 3550 hci_conn_drop(conn); 3551 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE, 3552 MGMT_STATUS_BUSY, &rp, sizeof(rp)); 3553 goto unlock; 3554 } 3555 3556 cmd = mgmt_pending_add(sk, MGMT_OP_PAIR_DEVICE, hdev, data, len); 3557 if (!cmd) { 3558 err = -ENOMEM; 3559 hci_conn_drop(conn); 3560 goto unlock; 3561 } 3562 3563 cmd->cmd_complete = pairing_complete; 3564 3565 /* For LE, just connecting isn't a proof that the pairing finished */ 3566 if (cp->addr.type == BDADDR_BREDR) { 3567 conn->connect_cfm_cb = pairing_complete_cb; 3568 conn->security_cfm_cb = pairing_complete_cb; 3569 conn->disconn_cfm_cb = pairing_complete_cb; 3570 } else { 3571 conn->connect_cfm_cb = le_pairing_complete_cb; 3572 conn->security_cfm_cb = le_pairing_complete_cb; 3573 conn->disconn_cfm_cb = le_pairing_complete_cb; 3574 } 3575 3576 conn->io_capability = cp->io_cap; 3577 cmd->user_data = hci_conn_get(conn); 3578 3579 if ((conn->state == BT_CONNECTED || conn->state == BT_CONFIG) && 3580 hci_conn_security(conn, sec_level, auth_type, true)) { 3581 cmd->cmd_complete(cmd, 0); 3582 mgmt_pending_remove(cmd); 3583 } 3584 3585 err = 0; 3586 3587 unlock: 3588 hci_dev_unlock(hdev); 3589 return err; 3590 } 3591 3592 static int cancel_pair_device(struct sock *sk, struct hci_dev *hdev, void *data, 3593 u16 len) 3594 { 3595 struct mgmt_addr_info *addr = data; 3596 struct mgmt_pending_cmd *cmd; 3597 struct hci_conn *conn; 3598 int err; 3599 3600 bt_dev_dbg(hdev, "sock %p", sk); 3601 3602 hci_dev_lock(hdev); 3603 3604 if (!hdev_is_powered(hdev)) { 3605 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 3606 MGMT_STATUS_NOT_POWERED); 3607 goto unlock; 3608 } 3609 3610 cmd = pending_find(MGMT_OP_PAIR_DEVICE, hdev); 3611 if (!cmd) { 3612 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 3613 MGMT_STATUS_INVALID_PARAMS); 3614 goto unlock; 3615 } 3616 3617 conn = cmd->user_data; 3618 3619 if (bacmp(&addr->bdaddr, &conn->dst) != 0) { 3620 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 3621 MGMT_STATUS_INVALID_PARAMS); 3622 goto unlock; 3623 } 3624 3625 cmd->cmd_complete(cmd, MGMT_STATUS_CANCELLED); 3626 mgmt_pending_remove(cmd); 3627 3628 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 0, 3629 addr, sizeof(*addr)); 3630 3631 /* Since user doesn't want to proceed with the connection, abort any 3632 * ongoing pairing and then terminate the link if it was created 3633 * because of the pair device action. 3634 */ 3635 if (addr->type == BDADDR_BREDR) 3636 hci_remove_link_key(hdev, &addr->bdaddr); 3637 else 3638 smp_cancel_and_remove_pairing(hdev, &addr->bdaddr, 3639 le_addr_type(addr->type)); 3640 3641 if (conn->conn_reason == CONN_REASON_PAIR_DEVICE) 3642 hci_abort_conn(conn, HCI_ERROR_REMOTE_USER_TERM); 3643 3644 unlock: 3645 hci_dev_unlock(hdev); 3646 return err; 3647 } 3648 3649 static int user_pairing_resp(struct sock *sk, struct hci_dev *hdev, 3650 struct mgmt_addr_info *addr, u16 mgmt_op, 3651 u16 hci_op, __le32 passkey) 3652 { 3653 struct mgmt_pending_cmd *cmd; 3654 struct hci_conn *conn; 3655 int err; 3656 3657 hci_dev_lock(hdev); 3658 3659 if (!hdev_is_powered(hdev)) { 3660 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3661 MGMT_STATUS_NOT_POWERED, addr, 3662 sizeof(*addr)); 3663 goto done; 3664 } 3665 3666 if (addr->type == BDADDR_BREDR) 3667 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &addr->bdaddr); 3668 else 3669 conn = hci_conn_hash_lookup_le(hdev, &addr->bdaddr, 3670 le_addr_type(addr->type)); 3671 3672 if (!conn) { 3673 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3674 MGMT_STATUS_NOT_CONNECTED, addr, 3675 sizeof(*addr)); 3676 goto done; 3677 } 3678 3679 if (addr->type == BDADDR_LE_PUBLIC || addr->type == BDADDR_LE_RANDOM) { 3680 err = smp_user_confirm_reply(conn, mgmt_op, passkey); 3681 if (!err) 3682 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3683 MGMT_STATUS_SUCCESS, addr, 3684 sizeof(*addr)); 3685 else 3686 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op, 3687 MGMT_STATUS_FAILED, addr, 3688 sizeof(*addr)); 3689 3690 goto done; 3691 } 3692 3693 cmd = mgmt_pending_add(sk, mgmt_op, hdev, addr, sizeof(*addr)); 3694 if (!cmd) { 3695 err = -ENOMEM; 3696 goto done; 3697 } 3698 3699 cmd->cmd_complete = addr_cmd_complete; 3700 3701 /* Continue with pairing via HCI */ 3702 if (hci_op == HCI_OP_USER_PASSKEY_REPLY) { 3703 struct hci_cp_user_passkey_reply cp; 3704 3705 bacpy(&cp.bdaddr, &addr->bdaddr); 3706 cp.passkey = passkey; 3707 err = hci_send_cmd(hdev, hci_op, sizeof(cp), &cp); 3708 } else 3709 err = hci_send_cmd(hdev, hci_op, sizeof(addr->bdaddr), 3710 &addr->bdaddr); 3711 3712 if (err < 0) 3713 mgmt_pending_remove(cmd); 3714 3715 done: 3716 hci_dev_unlock(hdev); 3717 return err; 3718 } 3719 3720 static int pin_code_neg_reply(struct sock *sk, struct hci_dev *hdev, 3721 void *data, u16 len) 3722 { 3723 struct mgmt_cp_pin_code_neg_reply *cp = data; 3724 3725 bt_dev_dbg(hdev, "sock %p", sk); 3726 3727 return user_pairing_resp(sk, hdev, &cp->addr, 3728 MGMT_OP_PIN_CODE_NEG_REPLY, 3729 HCI_OP_PIN_CODE_NEG_REPLY, 0); 3730 } 3731 3732 static int user_confirm_reply(struct sock *sk, struct hci_dev *hdev, void *data, 3733 u16 len) 3734 { 3735 struct mgmt_cp_user_confirm_reply *cp = data; 3736 3737 bt_dev_dbg(hdev, "sock %p", sk); 3738 3739 if (len != sizeof(*cp)) 3740 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_USER_CONFIRM_REPLY, 3741 MGMT_STATUS_INVALID_PARAMS); 3742 3743 return user_pairing_resp(sk, hdev, &cp->addr, 3744 MGMT_OP_USER_CONFIRM_REPLY, 3745 HCI_OP_USER_CONFIRM_REPLY, 0); 3746 } 3747 3748 static int user_confirm_neg_reply(struct sock *sk, struct hci_dev *hdev, 3749 void *data, u16 len) 3750 { 3751 struct mgmt_cp_user_confirm_neg_reply *cp = data; 3752 3753 bt_dev_dbg(hdev, "sock %p", sk); 3754 3755 return user_pairing_resp(sk, hdev, &cp->addr, 3756 MGMT_OP_USER_CONFIRM_NEG_REPLY, 3757 HCI_OP_USER_CONFIRM_NEG_REPLY, 0); 3758 } 3759 3760 static int user_passkey_reply(struct sock *sk, struct hci_dev *hdev, void *data, 3761 u16 len) 3762 { 3763 struct mgmt_cp_user_passkey_reply *cp = data; 3764 3765 bt_dev_dbg(hdev, "sock %p", sk); 3766 3767 return user_pairing_resp(sk, hdev, &cp->addr, 3768 MGMT_OP_USER_PASSKEY_REPLY, 3769 HCI_OP_USER_PASSKEY_REPLY, cp->passkey); 3770 } 3771 3772 static int user_passkey_neg_reply(struct sock *sk, struct hci_dev *hdev, 3773 void *data, u16 len) 3774 { 3775 struct mgmt_cp_user_passkey_neg_reply *cp = data; 3776 3777 bt_dev_dbg(hdev, "sock %p", sk); 3778 3779 return user_pairing_resp(sk, hdev, &cp->addr, 3780 MGMT_OP_USER_PASSKEY_NEG_REPLY, 3781 HCI_OP_USER_PASSKEY_NEG_REPLY, 0); 3782 } 3783 3784 static int adv_expire_sync(struct hci_dev *hdev, u32 flags) 3785 { 3786 struct adv_info *adv_instance; 3787 3788 adv_instance = hci_find_adv_instance(hdev, hdev->cur_adv_instance); 3789 if (!adv_instance) 3790 return 0; 3791 3792 /* stop if current instance doesn't need to be changed */ 3793 if (!(adv_instance->flags & flags)) 3794 return 0; 3795 3796 cancel_adv_timeout(hdev); 3797 3798 adv_instance = hci_get_next_instance(hdev, adv_instance->instance); 3799 if (!adv_instance) 3800 return 0; 3801 3802 hci_schedule_adv_instance_sync(hdev, adv_instance->instance, true); 3803 3804 return 0; 3805 } 3806 3807 static int name_changed_sync(struct hci_dev *hdev, void *data) 3808 { 3809 return adv_expire_sync(hdev, MGMT_ADV_FLAG_LOCAL_NAME); 3810 } 3811 3812 static void set_name_complete(struct hci_dev *hdev, void *data, int err) 3813 { 3814 struct mgmt_pending_cmd *cmd = data; 3815 struct mgmt_cp_set_local_name *cp = cmd->param; 3816 u8 status = mgmt_status(err); 3817 3818 bt_dev_dbg(hdev, "err %d", err); 3819 3820 if (cmd != pending_find(MGMT_OP_SET_LOCAL_NAME, hdev)) 3821 return; 3822 3823 if (status) { 3824 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 3825 status); 3826 } else { 3827 mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0, 3828 cp, sizeof(*cp)); 3829 3830 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 3831 hci_cmd_sync_queue(hdev, name_changed_sync, NULL, NULL); 3832 } 3833 3834 mgmt_pending_remove(cmd); 3835 } 3836 3837 static int set_name_sync(struct hci_dev *hdev, void *data) 3838 { 3839 if (lmp_bredr_capable(hdev)) { 3840 hci_update_name_sync(hdev); 3841 hci_update_eir_sync(hdev); 3842 } 3843 3844 /* The name is stored in the scan response data and so 3845 * no need to update the advertising data here. 3846 */ 3847 if (lmp_le_capable(hdev) && hci_dev_test_flag(hdev, HCI_ADVERTISING)) 3848 hci_update_scan_rsp_data_sync(hdev, hdev->cur_adv_instance); 3849 3850 return 0; 3851 } 3852 3853 static int set_local_name(struct sock *sk, struct hci_dev *hdev, void *data, 3854 u16 len) 3855 { 3856 struct mgmt_cp_set_local_name *cp = data; 3857 struct mgmt_pending_cmd *cmd; 3858 int err; 3859 3860 bt_dev_dbg(hdev, "sock %p", sk); 3861 3862 hci_dev_lock(hdev); 3863 3864 /* If the old values are the same as the new ones just return a 3865 * direct command complete event. 3866 */ 3867 if (!memcmp(hdev->dev_name, cp->name, sizeof(hdev->dev_name)) && 3868 !memcmp(hdev->short_name, cp->short_name, 3869 sizeof(hdev->short_name))) { 3870 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0, 3871 data, len); 3872 goto failed; 3873 } 3874 3875 memcpy(hdev->short_name, cp->short_name, sizeof(hdev->short_name)); 3876 3877 if (!hdev_is_powered(hdev)) { 3878 memcpy(hdev->dev_name, cp->name, sizeof(hdev->dev_name)); 3879 3880 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0, 3881 data, len); 3882 if (err < 0) 3883 goto failed; 3884 3885 err = mgmt_limited_event(MGMT_EV_LOCAL_NAME_CHANGED, hdev, data, 3886 len, HCI_MGMT_LOCAL_NAME_EVENTS, sk); 3887 ext_info_changed(hdev, sk); 3888 3889 goto failed; 3890 } 3891 3892 cmd = mgmt_pending_add(sk, MGMT_OP_SET_LOCAL_NAME, hdev, data, len); 3893 if (!cmd) 3894 err = -ENOMEM; 3895 else 3896 err = hci_cmd_sync_queue(hdev, set_name_sync, cmd, 3897 set_name_complete); 3898 3899 if (err < 0) { 3900 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 3901 MGMT_STATUS_FAILED); 3902 3903 if (cmd) 3904 mgmt_pending_remove(cmd); 3905 3906 goto failed; 3907 } 3908 3909 memcpy(hdev->dev_name, cp->name, sizeof(hdev->dev_name)); 3910 3911 failed: 3912 hci_dev_unlock(hdev); 3913 return err; 3914 } 3915 3916 static int appearance_changed_sync(struct hci_dev *hdev, void *data) 3917 { 3918 return adv_expire_sync(hdev, MGMT_ADV_FLAG_APPEARANCE); 3919 } 3920 3921 static int set_appearance(struct sock *sk, struct hci_dev *hdev, void *data, 3922 u16 len) 3923 { 3924 struct mgmt_cp_set_appearance *cp = data; 3925 u16 appearance; 3926 int err; 3927 3928 bt_dev_dbg(hdev, "sock %p", sk); 3929 3930 if (!lmp_le_capable(hdev)) 3931 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_APPEARANCE, 3932 MGMT_STATUS_NOT_SUPPORTED); 3933 3934 appearance = le16_to_cpu(cp->appearance); 3935 3936 hci_dev_lock(hdev); 3937 3938 if (hdev->appearance != appearance) { 3939 hdev->appearance = appearance; 3940 3941 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 3942 hci_cmd_sync_queue(hdev, appearance_changed_sync, NULL, 3943 NULL); 3944 3945 ext_info_changed(hdev, sk); 3946 } 3947 3948 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_APPEARANCE, 0, NULL, 3949 0); 3950 3951 hci_dev_unlock(hdev); 3952 3953 return err; 3954 } 3955 3956 static int get_phy_configuration(struct sock *sk, struct hci_dev *hdev, 3957 void *data, u16 len) 3958 { 3959 struct mgmt_rp_get_phy_configuration rp; 3960 3961 bt_dev_dbg(hdev, "sock %p", sk); 3962 3963 hci_dev_lock(hdev); 3964 3965 memset(&rp, 0, sizeof(rp)); 3966 3967 rp.supported_phys = cpu_to_le32(get_supported_phys(hdev)); 3968 rp.selected_phys = cpu_to_le32(get_selected_phys(hdev)); 3969 rp.configurable_phys = cpu_to_le32(get_configurable_phys(hdev)); 3970 3971 hci_dev_unlock(hdev); 3972 3973 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_PHY_CONFIGURATION, 0, 3974 &rp, sizeof(rp)); 3975 } 3976 3977 int mgmt_phy_configuration_changed(struct hci_dev *hdev, struct sock *skip) 3978 { 3979 struct mgmt_ev_phy_configuration_changed ev; 3980 3981 memset(&ev, 0, sizeof(ev)); 3982 3983 ev.selected_phys = cpu_to_le32(get_selected_phys(hdev)); 3984 3985 return mgmt_event(MGMT_EV_PHY_CONFIGURATION_CHANGED, hdev, &ev, 3986 sizeof(ev), skip); 3987 } 3988 3989 static void set_default_phy_complete(struct hci_dev *hdev, void *data, int err) 3990 { 3991 struct mgmt_pending_cmd *cmd = data; 3992 struct sk_buff *skb = cmd->skb; 3993 u8 status = mgmt_status(err); 3994 3995 if (cmd != pending_find(MGMT_OP_SET_PHY_CONFIGURATION, hdev)) 3996 return; 3997 3998 if (!status) { 3999 if (!skb) 4000 status = MGMT_STATUS_FAILED; 4001 else if (IS_ERR(skb)) 4002 status = mgmt_status(PTR_ERR(skb)); 4003 else 4004 status = mgmt_status(skb->data[0]); 4005 } 4006 4007 bt_dev_dbg(hdev, "status %d", status); 4008 4009 if (status) { 4010 mgmt_cmd_status(cmd->sk, hdev->id, 4011 MGMT_OP_SET_PHY_CONFIGURATION, status); 4012 } else { 4013 mgmt_cmd_complete(cmd->sk, hdev->id, 4014 MGMT_OP_SET_PHY_CONFIGURATION, 0, 4015 NULL, 0); 4016 4017 mgmt_phy_configuration_changed(hdev, cmd->sk); 4018 } 4019 4020 if (skb && !IS_ERR(skb)) 4021 kfree_skb(skb); 4022 4023 mgmt_pending_remove(cmd); 4024 } 4025 4026 static int set_default_phy_sync(struct hci_dev *hdev, void *data) 4027 { 4028 struct mgmt_pending_cmd *cmd = data; 4029 struct mgmt_cp_set_phy_configuration *cp = cmd->param; 4030 struct hci_cp_le_set_default_phy cp_phy; 4031 u32 selected_phys = __le32_to_cpu(cp->selected_phys); 4032 4033 memset(&cp_phy, 0, sizeof(cp_phy)); 4034 4035 if (!(selected_phys & MGMT_PHY_LE_TX_MASK)) 4036 cp_phy.all_phys |= 0x01; 4037 4038 if (!(selected_phys & MGMT_PHY_LE_RX_MASK)) 4039 cp_phy.all_phys |= 0x02; 4040 4041 if (selected_phys & MGMT_PHY_LE_1M_TX) 4042 cp_phy.tx_phys |= HCI_LE_SET_PHY_1M; 4043 4044 if (selected_phys & MGMT_PHY_LE_2M_TX) 4045 cp_phy.tx_phys |= HCI_LE_SET_PHY_2M; 4046 4047 if (selected_phys & MGMT_PHY_LE_CODED_TX) 4048 cp_phy.tx_phys |= HCI_LE_SET_PHY_CODED; 4049 4050 if (selected_phys & MGMT_PHY_LE_1M_RX) 4051 cp_phy.rx_phys |= HCI_LE_SET_PHY_1M; 4052 4053 if (selected_phys & MGMT_PHY_LE_2M_RX) 4054 cp_phy.rx_phys |= HCI_LE_SET_PHY_2M; 4055 4056 if (selected_phys & MGMT_PHY_LE_CODED_RX) 4057 cp_phy.rx_phys |= HCI_LE_SET_PHY_CODED; 4058 4059 cmd->skb = __hci_cmd_sync(hdev, HCI_OP_LE_SET_DEFAULT_PHY, 4060 sizeof(cp_phy), &cp_phy, HCI_CMD_TIMEOUT); 4061 4062 return 0; 4063 } 4064 4065 static int set_phy_configuration(struct sock *sk, struct hci_dev *hdev, 4066 void *data, u16 len) 4067 { 4068 struct mgmt_cp_set_phy_configuration *cp = data; 4069 struct mgmt_pending_cmd *cmd; 4070 u32 selected_phys, configurable_phys, supported_phys, unconfigure_phys; 4071 u16 pkt_type = (HCI_DH1 | HCI_DM1); 4072 bool changed = false; 4073 int err; 4074 4075 bt_dev_dbg(hdev, "sock %p", sk); 4076 4077 configurable_phys = get_configurable_phys(hdev); 4078 supported_phys = get_supported_phys(hdev); 4079 selected_phys = __le32_to_cpu(cp->selected_phys); 4080 4081 if (selected_phys & ~supported_phys) 4082 return mgmt_cmd_status(sk, hdev->id, 4083 MGMT_OP_SET_PHY_CONFIGURATION, 4084 MGMT_STATUS_INVALID_PARAMS); 4085 4086 unconfigure_phys = supported_phys & ~configurable_phys; 4087 4088 if ((selected_phys & unconfigure_phys) != unconfigure_phys) 4089 return mgmt_cmd_status(sk, hdev->id, 4090 MGMT_OP_SET_PHY_CONFIGURATION, 4091 MGMT_STATUS_INVALID_PARAMS); 4092 4093 if (selected_phys == get_selected_phys(hdev)) 4094 return mgmt_cmd_complete(sk, hdev->id, 4095 MGMT_OP_SET_PHY_CONFIGURATION, 4096 0, NULL, 0); 4097 4098 hci_dev_lock(hdev); 4099 4100 if (!hdev_is_powered(hdev)) { 4101 err = mgmt_cmd_status(sk, hdev->id, 4102 MGMT_OP_SET_PHY_CONFIGURATION, 4103 MGMT_STATUS_REJECTED); 4104 goto unlock; 4105 } 4106 4107 if (pending_find(MGMT_OP_SET_PHY_CONFIGURATION, hdev)) { 4108 err = mgmt_cmd_status(sk, hdev->id, 4109 MGMT_OP_SET_PHY_CONFIGURATION, 4110 MGMT_STATUS_BUSY); 4111 goto unlock; 4112 } 4113 4114 if (selected_phys & MGMT_PHY_BR_1M_3SLOT) 4115 pkt_type |= (HCI_DH3 | HCI_DM3); 4116 else 4117 pkt_type &= ~(HCI_DH3 | HCI_DM3); 4118 4119 if (selected_phys & MGMT_PHY_BR_1M_5SLOT) 4120 pkt_type |= (HCI_DH5 | HCI_DM5); 4121 else 4122 pkt_type &= ~(HCI_DH5 | HCI_DM5); 4123 4124 if (selected_phys & MGMT_PHY_EDR_2M_1SLOT) 4125 pkt_type &= ~HCI_2DH1; 4126 else 4127 pkt_type |= HCI_2DH1; 4128 4129 if (selected_phys & MGMT_PHY_EDR_2M_3SLOT) 4130 pkt_type &= ~HCI_2DH3; 4131 else 4132 pkt_type |= HCI_2DH3; 4133 4134 if (selected_phys & MGMT_PHY_EDR_2M_5SLOT) 4135 pkt_type &= ~HCI_2DH5; 4136 else 4137 pkt_type |= HCI_2DH5; 4138 4139 if (selected_phys & MGMT_PHY_EDR_3M_1SLOT) 4140 pkt_type &= ~HCI_3DH1; 4141 else 4142 pkt_type |= HCI_3DH1; 4143 4144 if (selected_phys & MGMT_PHY_EDR_3M_3SLOT) 4145 pkt_type &= ~HCI_3DH3; 4146 else 4147 pkt_type |= HCI_3DH3; 4148 4149 if (selected_phys & MGMT_PHY_EDR_3M_5SLOT) 4150 pkt_type &= ~HCI_3DH5; 4151 else 4152 pkt_type |= HCI_3DH5; 4153 4154 if (pkt_type != hdev->pkt_type) { 4155 hdev->pkt_type = pkt_type; 4156 changed = true; 4157 } 4158 4159 if ((selected_phys & MGMT_PHY_LE_MASK) == 4160 (get_selected_phys(hdev) & MGMT_PHY_LE_MASK)) { 4161 if (changed) 4162 mgmt_phy_configuration_changed(hdev, sk); 4163 4164 err = mgmt_cmd_complete(sk, hdev->id, 4165 MGMT_OP_SET_PHY_CONFIGURATION, 4166 0, NULL, 0); 4167 4168 goto unlock; 4169 } 4170 4171 cmd = mgmt_pending_add(sk, MGMT_OP_SET_PHY_CONFIGURATION, hdev, data, 4172 len); 4173 if (!cmd) 4174 err = -ENOMEM; 4175 else 4176 err = hci_cmd_sync_queue(hdev, set_default_phy_sync, cmd, 4177 set_default_phy_complete); 4178 4179 if (err < 0) { 4180 err = mgmt_cmd_status(sk, hdev->id, 4181 MGMT_OP_SET_PHY_CONFIGURATION, 4182 MGMT_STATUS_FAILED); 4183 4184 if (cmd) 4185 mgmt_pending_remove(cmd); 4186 } 4187 4188 unlock: 4189 hci_dev_unlock(hdev); 4190 4191 return err; 4192 } 4193 4194 static int set_blocked_keys(struct sock *sk, struct hci_dev *hdev, void *data, 4195 u16 len) 4196 { 4197 int err = MGMT_STATUS_SUCCESS; 4198 struct mgmt_cp_set_blocked_keys *keys = data; 4199 const u16 max_key_count = ((U16_MAX - sizeof(*keys)) / 4200 sizeof(struct mgmt_blocked_key_info)); 4201 u16 key_count, expected_len; 4202 int i; 4203 4204 bt_dev_dbg(hdev, "sock %p", sk); 4205 4206 key_count = __le16_to_cpu(keys->key_count); 4207 if (key_count > max_key_count) { 4208 bt_dev_err(hdev, "too big key_count value %u", key_count); 4209 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS, 4210 MGMT_STATUS_INVALID_PARAMS); 4211 } 4212 4213 expected_len = struct_size(keys, keys, key_count); 4214 if (expected_len != len) { 4215 bt_dev_err(hdev, "expected %u bytes, got %u bytes", 4216 expected_len, len); 4217 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS, 4218 MGMT_STATUS_INVALID_PARAMS); 4219 } 4220 4221 hci_dev_lock(hdev); 4222 4223 hci_blocked_keys_clear(hdev); 4224 4225 for (i = 0; i < key_count; ++i) { 4226 struct blocked_key *b = kzalloc(sizeof(*b), GFP_KERNEL); 4227 4228 if (!b) { 4229 err = MGMT_STATUS_NO_RESOURCES; 4230 break; 4231 } 4232 4233 b->type = keys->keys[i].type; 4234 memcpy(b->val, keys->keys[i].val, sizeof(b->val)); 4235 list_add_rcu(&b->list, &hdev->blocked_keys); 4236 } 4237 hci_dev_unlock(hdev); 4238 4239 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS, 4240 err, NULL, 0); 4241 } 4242 4243 static int set_wideband_speech(struct sock *sk, struct hci_dev *hdev, 4244 void *data, u16 len) 4245 { 4246 struct mgmt_mode *cp = data; 4247 int err; 4248 bool changed = false; 4249 4250 bt_dev_dbg(hdev, "sock %p", sk); 4251 4252 if (!test_bit(HCI_QUIRK_WIDEBAND_SPEECH_SUPPORTED, &hdev->quirks)) 4253 return mgmt_cmd_status(sk, hdev->id, 4254 MGMT_OP_SET_WIDEBAND_SPEECH, 4255 MGMT_STATUS_NOT_SUPPORTED); 4256 4257 if (cp->val != 0x00 && cp->val != 0x01) 4258 return mgmt_cmd_status(sk, hdev->id, 4259 MGMT_OP_SET_WIDEBAND_SPEECH, 4260 MGMT_STATUS_INVALID_PARAMS); 4261 4262 hci_dev_lock(hdev); 4263 4264 if (hdev_is_powered(hdev) && 4265 !!cp->val != hci_dev_test_flag(hdev, 4266 HCI_WIDEBAND_SPEECH_ENABLED)) { 4267 err = mgmt_cmd_status(sk, hdev->id, 4268 MGMT_OP_SET_WIDEBAND_SPEECH, 4269 MGMT_STATUS_REJECTED); 4270 goto unlock; 4271 } 4272 4273 if (cp->val) 4274 changed = !hci_dev_test_and_set_flag(hdev, 4275 HCI_WIDEBAND_SPEECH_ENABLED); 4276 else 4277 changed = hci_dev_test_and_clear_flag(hdev, 4278 HCI_WIDEBAND_SPEECH_ENABLED); 4279 4280 err = send_settings_rsp(sk, MGMT_OP_SET_WIDEBAND_SPEECH, hdev); 4281 if (err < 0) 4282 goto unlock; 4283 4284 if (changed) 4285 err = new_settings(hdev, sk); 4286 4287 unlock: 4288 hci_dev_unlock(hdev); 4289 return err; 4290 } 4291 4292 static int read_controller_cap(struct sock *sk, struct hci_dev *hdev, 4293 void *data, u16 data_len) 4294 { 4295 char buf[20]; 4296 struct mgmt_rp_read_controller_cap *rp = (void *)buf; 4297 u16 cap_len = 0; 4298 u8 flags = 0; 4299 u8 tx_power_range[2]; 4300 4301 bt_dev_dbg(hdev, "sock %p", sk); 4302 4303 memset(&buf, 0, sizeof(buf)); 4304 4305 hci_dev_lock(hdev); 4306 4307 /* When the Read Simple Pairing Options command is supported, then 4308 * the remote public key validation is supported. 4309 * 4310 * Alternatively, when Microsoft extensions are available, they can 4311 * indicate support for public key validation as well. 4312 */ 4313 if ((hdev->commands[41] & 0x08) || msft_curve_validity(hdev)) 4314 flags |= 0x01; /* Remote public key validation (BR/EDR) */ 4315 4316 flags |= 0x02; /* Remote public key validation (LE) */ 4317 4318 /* When the Read Encryption Key Size command is supported, then the 4319 * encryption key size is enforced. 4320 */ 4321 if (hdev->commands[20] & 0x10) 4322 flags |= 0x04; /* Encryption key size enforcement (BR/EDR) */ 4323 4324 flags |= 0x08; /* Encryption key size enforcement (LE) */ 4325 4326 cap_len = eir_append_data(rp->cap, cap_len, MGMT_CAP_SEC_FLAGS, 4327 &flags, 1); 4328 4329 /* When the Read Simple Pairing Options command is supported, then 4330 * also max encryption key size information is provided. 4331 */ 4332 if (hdev->commands[41] & 0x08) 4333 cap_len = eir_append_le16(rp->cap, cap_len, 4334 MGMT_CAP_MAX_ENC_KEY_SIZE, 4335 hdev->max_enc_key_size); 4336 4337 cap_len = eir_append_le16(rp->cap, cap_len, 4338 MGMT_CAP_SMP_MAX_ENC_KEY_SIZE, 4339 SMP_MAX_ENC_KEY_SIZE); 4340 4341 /* Append the min/max LE tx power parameters if we were able to fetch 4342 * it from the controller 4343 */ 4344 if (hdev->commands[38] & 0x80) { 4345 memcpy(&tx_power_range[0], &hdev->min_le_tx_power, 1); 4346 memcpy(&tx_power_range[1], &hdev->max_le_tx_power, 1); 4347 cap_len = eir_append_data(rp->cap, cap_len, MGMT_CAP_LE_TX_PWR, 4348 tx_power_range, 2); 4349 } 4350 4351 rp->cap_len = cpu_to_le16(cap_len); 4352 4353 hci_dev_unlock(hdev); 4354 4355 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_CONTROLLER_CAP, 0, 4356 rp, sizeof(*rp) + cap_len); 4357 } 4358 4359 #ifdef CONFIG_BT_FEATURE_DEBUG 4360 /* d4992530-b9ec-469f-ab01-6c481c47da1c */ 4361 static const u8 debug_uuid[16] = { 4362 0x1c, 0xda, 0x47, 0x1c, 0x48, 0x6c, 0x01, 0xab, 4363 0x9f, 0x46, 0xec, 0xb9, 0x30, 0x25, 0x99, 0xd4, 4364 }; 4365 #endif 4366 4367 /* 330859bc-7506-492d-9370-9a6f0614037f */ 4368 static const u8 quality_report_uuid[16] = { 4369 0x7f, 0x03, 0x14, 0x06, 0x6f, 0x9a, 0x70, 0x93, 4370 0x2d, 0x49, 0x06, 0x75, 0xbc, 0x59, 0x08, 0x33, 4371 }; 4372 4373 /* a6695ace-ee7f-4fb9-881a-5fac66c629af */ 4374 static const u8 offload_codecs_uuid[16] = { 4375 0xaf, 0x29, 0xc6, 0x66, 0xac, 0x5f, 0x1a, 0x88, 4376 0xb9, 0x4f, 0x7f, 0xee, 0xce, 0x5a, 0x69, 0xa6, 4377 }; 4378 4379 /* 671b10b5-42c0-4696-9227-eb28d1b049d6 */ 4380 static const u8 le_simultaneous_roles_uuid[16] = { 4381 0xd6, 0x49, 0xb0, 0xd1, 0x28, 0xeb, 0x27, 0x92, 4382 0x96, 0x46, 0xc0, 0x42, 0xb5, 0x10, 0x1b, 0x67, 4383 }; 4384 4385 /* 15c0a148-c273-11ea-b3de-0242ac130004 */ 4386 static const u8 rpa_resolution_uuid[16] = { 4387 0x04, 0x00, 0x13, 0xac, 0x42, 0x02, 0xde, 0xb3, 4388 0xea, 0x11, 0x73, 0xc2, 0x48, 0xa1, 0xc0, 0x15, 4389 }; 4390 4391 /* 6fbaf188-05e0-496a-9885-d6ddfdb4e03e */ 4392 static const u8 iso_socket_uuid[16] = { 4393 0x3e, 0xe0, 0xb4, 0xfd, 0xdd, 0xd6, 0x85, 0x98, 4394 0x6a, 0x49, 0xe0, 0x05, 0x88, 0xf1, 0xba, 0x6f, 4395 }; 4396 4397 /* 2ce463d7-7a03-4d8d-bf05-5f24e8f36e76 */ 4398 static const u8 mgmt_mesh_uuid[16] = { 4399 0x76, 0x6e, 0xf3, 0xe8, 0x24, 0x5f, 0x05, 0xbf, 4400 0x8d, 0x4d, 0x03, 0x7a, 0xd7, 0x63, 0xe4, 0x2c, 4401 }; 4402 4403 static int read_exp_features_info(struct sock *sk, struct hci_dev *hdev, 4404 void *data, u16 data_len) 4405 { 4406 struct mgmt_rp_read_exp_features_info *rp; 4407 size_t len; 4408 u16 idx = 0; 4409 u32 flags; 4410 int status; 4411 4412 bt_dev_dbg(hdev, "sock %p", sk); 4413 4414 /* Enough space for 7 features */ 4415 len = sizeof(*rp) + (sizeof(rp->features[0]) * 7); 4416 rp = kzalloc(len, GFP_KERNEL); 4417 if (!rp) 4418 return -ENOMEM; 4419 4420 #ifdef CONFIG_BT_FEATURE_DEBUG 4421 if (!hdev) { 4422 flags = bt_dbg_get() ? BIT(0) : 0; 4423 4424 memcpy(rp->features[idx].uuid, debug_uuid, 16); 4425 rp->features[idx].flags = cpu_to_le32(flags); 4426 idx++; 4427 } 4428 #endif 4429 4430 if (hdev && hci_dev_le_state_simultaneous(hdev)) { 4431 if (hci_dev_test_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES)) 4432 flags = BIT(0); 4433 else 4434 flags = 0; 4435 4436 memcpy(rp->features[idx].uuid, le_simultaneous_roles_uuid, 16); 4437 rp->features[idx].flags = cpu_to_le32(flags); 4438 idx++; 4439 } 4440 4441 if (hdev && ll_privacy_capable(hdev)) { 4442 if (hci_dev_test_flag(hdev, HCI_ENABLE_LL_PRIVACY)) 4443 flags = BIT(0) | BIT(1); 4444 else 4445 flags = BIT(1); 4446 4447 memcpy(rp->features[idx].uuid, rpa_resolution_uuid, 16); 4448 rp->features[idx].flags = cpu_to_le32(flags); 4449 idx++; 4450 } 4451 4452 if (hdev && (aosp_has_quality_report(hdev) || 4453 hdev->set_quality_report)) { 4454 if (hci_dev_test_flag(hdev, HCI_QUALITY_REPORT)) 4455 flags = BIT(0); 4456 else 4457 flags = 0; 4458 4459 memcpy(rp->features[idx].uuid, quality_report_uuid, 16); 4460 rp->features[idx].flags = cpu_to_le32(flags); 4461 idx++; 4462 } 4463 4464 if (hdev && hdev->get_data_path_id) { 4465 if (hci_dev_test_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED)) 4466 flags = BIT(0); 4467 else 4468 flags = 0; 4469 4470 memcpy(rp->features[idx].uuid, offload_codecs_uuid, 16); 4471 rp->features[idx].flags = cpu_to_le32(flags); 4472 idx++; 4473 } 4474 4475 if (IS_ENABLED(CONFIG_BT_LE)) { 4476 flags = iso_enabled() ? BIT(0) : 0; 4477 memcpy(rp->features[idx].uuid, iso_socket_uuid, 16); 4478 rp->features[idx].flags = cpu_to_le32(flags); 4479 idx++; 4480 } 4481 4482 if (hdev && lmp_le_capable(hdev)) { 4483 if (hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL)) 4484 flags = BIT(0); 4485 else 4486 flags = 0; 4487 4488 memcpy(rp->features[idx].uuid, mgmt_mesh_uuid, 16); 4489 rp->features[idx].flags = cpu_to_le32(flags); 4490 idx++; 4491 } 4492 4493 rp->feature_count = cpu_to_le16(idx); 4494 4495 /* After reading the experimental features information, enable 4496 * the events to update client on any future change. 4497 */ 4498 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4499 4500 status = mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE, 4501 MGMT_OP_READ_EXP_FEATURES_INFO, 4502 0, rp, sizeof(*rp) + (20 * idx)); 4503 4504 kfree(rp); 4505 return status; 4506 } 4507 4508 static int exp_ll_privacy_feature_changed(bool enabled, struct hci_dev *hdev, 4509 struct sock *skip) 4510 { 4511 struct mgmt_ev_exp_feature_changed ev; 4512 4513 memset(&ev, 0, sizeof(ev)); 4514 memcpy(ev.uuid, rpa_resolution_uuid, 16); 4515 ev.flags = cpu_to_le32((enabled ? BIT(0) : 0) | BIT(1)); 4516 4517 // Do we need to be atomic with the conn_flags? 4518 if (enabled && privacy_mode_capable(hdev)) 4519 hdev->conn_flags |= HCI_CONN_FLAG_DEVICE_PRIVACY; 4520 else 4521 hdev->conn_flags &= ~HCI_CONN_FLAG_DEVICE_PRIVACY; 4522 4523 return mgmt_limited_event(MGMT_EV_EXP_FEATURE_CHANGED, hdev, 4524 &ev, sizeof(ev), 4525 HCI_MGMT_EXP_FEATURE_EVENTS, skip); 4526 4527 } 4528 4529 static int exp_feature_changed(struct hci_dev *hdev, const u8 *uuid, 4530 bool enabled, struct sock *skip) 4531 { 4532 struct mgmt_ev_exp_feature_changed ev; 4533 4534 memset(&ev, 0, sizeof(ev)); 4535 memcpy(ev.uuid, uuid, 16); 4536 ev.flags = cpu_to_le32(enabled ? BIT(0) : 0); 4537 4538 return mgmt_limited_event(MGMT_EV_EXP_FEATURE_CHANGED, hdev, 4539 &ev, sizeof(ev), 4540 HCI_MGMT_EXP_FEATURE_EVENTS, skip); 4541 } 4542 4543 #define EXP_FEAT(_uuid, _set_func) \ 4544 { \ 4545 .uuid = _uuid, \ 4546 .set_func = _set_func, \ 4547 } 4548 4549 /* The zero key uuid is special. Multiple exp features are set through it. */ 4550 static int set_zero_key_func(struct sock *sk, struct hci_dev *hdev, 4551 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4552 { 4553 struct mgmt_rp_set_exp_feature rp; 4554 4555 memset(rp.uuid, 0, 16); 4556 rp.flags = cpu_to_le32(0); 4557 4558 #ifdef CONFIG_BT_FEATURE_DEBUG 4559 if (!hdev) { 4560 bool changed = bt_dbg_get(); 4561 4562 bt_dbg_set(false); 4563 4564 if (changed) 4565 exp_feature_changed(NULL, ZERO_KEY, false, sk); 4566 } 4567 #endif 4568 4569 if (hdev && use_ll_privacy(hdev) && !hdev_is_powered(hdev)) { 4570 bool changed; 4571 4572 changed = hci_dev_test_and_clear_flag(hdev, 4573 HCI_ENABLE_LL_PRIVACY); 4574 if (changed) 4575 exp_feature_changed(hdev, rpa_resolution_uuid, false, 4576 sk); 4577 } 4578 4579 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4580 4581 return mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE, 4582 MGMT_OP_SET_EXP_FEATURE, 0, 4583 &rp, sizeof(rp)); 4584 } 4585 4586 #ifdef CONFIG_BT_FEATURE_DEBUG 4587 static int set_debug_func(struct sock *sk, struct hci_dev *hdev, 4588 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4589 { 4590 struct mgmt_rp_set_exp_feature rp; 4591 4592 bool val, changed; 4593 int err; 4594 4595 /* Command requires to use the non-controller index */ 4596 if (hdev) 4597 return mgmt_cmd_status(sk, hdev->id, 4598 MGMT_OP_SET_EXP_FEATURE, 4599 MGMT_STATUS_INVALID_INDEX); 4600 4601 /* Parameters are limited to a single octet */ 4602 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4603 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4604 MGMT_OP_SET_EXP_FEATURE, 4605 MGMT_STATUS_INVALID_PARAMS); 4606 4607 /* Only boolean on/off is supported */ 4608 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4609 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4610 MGMT_OP_SET_EXP_FEATURE, 4611 MGMT_STATUS_INVALID_PARAMS); 4612 4613 val = !!cp->param[0]; 4614 changed = val ? !bt_dbg_get() : bt_dbg_get(); 4615 bt_dbg_set(val); 4616 4617 memcpy(rp.uuid, debug_uuid, 16); 4618 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4619 4620 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4621 4622 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 4623 MGMT_OP_SET_EXP_FEATURE, 0, 4624 &rp, sizeof(rp)); 4625 4626 if (changed) 4627 exp_feature_changed(hdev, debug_uuid, val, sk); 4628 4629 return err; 4630 } 4631 #endif 4632 4633 static int set_mgmt_mesh_func(struct sock *sk, struct hci_dev *hdev, 4634 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4635 { 4636 struct mgmt_rp_set_exp_feature rp; 4637 bool val, changed; 4638 int err; 4639 4640 /* Command requires to use the controller index */ 4641 if (!hdev) 4642 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4643 MGMT_OP_SET_EXP_FEATURE, 4644 MGMT_STATUS_INVALID_INDEX); 4645 4646 /* Parameters are limited to a single octet */ 4647 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4648 return mgmt_cmd_status(sk, hdev->id, 4649 MGMT_OP_SET_EXP_FEATURE, 4650 MGMT_STATUS_INVALID_PARAMS); 4651 4652 /* Only boolean on/off is supported */ 4653 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4654 return mgmt_cmd_status(sk, hdev->id, 4655 MGMT_OP_SET_EXP_FEATURE, 4656 MGMT_STATUS_INVALID_PARAMS); 4657 4658 val = !!cp->param[0]; 4659 4660 if (val) { 4661 changed = !hci_dev_test_and_set_flag(hdev, 4662 HCI_MESH_EXPERIMENTAL); 4663 } else { 4664 hci_dev_clear_flag(hdev, HCI_MESH); 4665 changed = hci_dev_test_and_clear_flag(hdev, 4666 HCI_MESH_EXPERIMENTAL); 4667 } 4668 4669 memcpy(rp.uuid, mgmt_mesh_uuid, 16); 4670 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4671 4672 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4673 4674 err = mgmt_cmd_complete(sk, hdev->id, 4675 MGMT_OP_SET_EXP_FEATURE, 0, 4676 &rp, sizeof(rp)); 4677 4678 if (changed) 4679 exp_feature_changed(hdev, mgmt_mesh_uuid, val, sk); 4680 4681 return err; 4682 } 4683 4684 static int set_rpa_resolution_func(struct sock *sk, struct hci_dev *hdev, 4685 struct mgmt_cp_set_exp_feature *cp, 4686 u16 data_len) 4687 { 4688 struct mgmt_rp_set_exp_feature rp; 4689 bool val, changed; 4690 int err; 4691 u32 flags; 4692 4693 /* Command requires to use the controller index */ 4694 if (!hdev) 4695 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4696 MGMT_OP_SET_EXP_FEATURE, 4697 MGMT_STATUS_INVALID_INDEX); 4698 4699 /* Changes can only be made when controller is powered down */ 4700 if (hdev_is_powered(hdev)) 4701 return mgmt_cmd_status(sk, hdev->id, 4702 MGMT_OP_SET_EXP_FEATURE, 4703 MGMT_STATUS_REJECTED); 4704 4705 /* Parameters are limited to a single octet */ 4706 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4707 return mgmt_cmd_status(sk, hdev->id, 4708 MGMT_OP_SET_EXP_FEATURE, 4709 MGMT_STATUS_INVALID_PARAMS); 4710 4711 /* Only boolean on/off is supported */ 4712 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4713 return mgmt_cmd_status(sk, hdev->id, 4714 MGMT_OP_SET_EXP_FEATURE, 4715 MGMT_STATUS_INVALID_PARAMS); 4716 4717 val = !!cp->param[0]; 4718 4719 if (val) { 4720 changed = !hci_dev_test_and_set_flag(hdev, 4721 HCI_ENABLE_LL_PRIVACY); 4722 hci_dev_clear_flag(hdev, HCI_ADVERTISING); 4723 4724 /* Enable LL privacy + supported settings changed */ 4725 flags = BIT(0) | BIT(1); 4726 } else { 4727 changed = hci_dev_test_and_clear_flag(hdev, 4728 HCI_ENABLE_LL_PRIVACY); 4729 4730 /* Disable LL privacy + supported settings changed */ 4731 flags = BIT(1); 4732 } 4733 4734 memcpy(rp.uuid, rpa_resolution_uuid, 16); 4735 rp.flags = cpu_to_le32(flags); 4736 4737 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4738 4739 err = mgmt_cmd_complete(sk, hdev->id, 4740 MGMT_OP_SET_EXP_FEATURE, 0, 4741 &rp, sizeof(rp)); 4742 4743 if (changed) 4744 exp_ll_privacy_feature_changed(val, hdev, sk); 4745 4746 return err; 4747 } 4748 4749 static int set_quality_report_func(struct sock *sk, struct hci_dev *hdev, 4750 struct mgmt_cp_set_exp_feature *cp, 4751 u16 data_len) 4752 { 4753 struct mgmt_rp_set_exp_feature rp; 4754 bool val, changed; 4755 int err; 4756 4757 /* Command requires to use a valid controller index */ 4758 if (!hdev) 4759 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4760 MGMT_OP_SET_EXP_FEATURE, 4761 MGMT_STATUS_INVALID_INDEX); 4762 4763 /* Parameters are limited to a single octet */ 4764 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4765 return mgmt_cmd_status(sk, hdev->id, 4766 MGMT_OP_SET_EXP_FEATURE, 4767 MGMT_STATUS_INVALID_PARAMS); 4768 4769 /* Only boolean on/off is supported */ 4770 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4771 return mgmt_cmd_status(sk, hdev->id, 4772 MGMT_OP_SET_EXP_FEATURE, 4773 MGMT_STATUS_INVALID_PARAMS); 4774 4775 hci_req_sync_lock(hdev); 4776 4777 val = !!cp->param[0]; 4778 changed = (val != hci_dev_test_flag(hdev, HCI_QUALITY_REPORT)); 4779 4780 if (!aosp_has_quality_report(hdev) && !hdev->set_quality_report) { 4781 err = mgmt_cmd_status(sk, hdev->id, 4782 MGMT_OP_SET_EXP_FEATURE, 4783 MGMT_STATUS_NOT_SUPPORTED); 4784 goto unlock_quality_report; 4785 } 4786 4787 if (changed) { 4788 if (hdev->set_quality_report) 4789 err = hdev->set_quality_report(hdev, val); 4790 else 4791 err = aosp_set_quality_report(hdev, val); 4792 4793 if (err) { 4794 err = mgmt_cmd_status(sk, hdev->id, 4795 MGMT_OP_SET_EXP_FEATURE, 4796 MGMT_STATUS_FAILED); 4797 goto unlock_quality_report; 4798 } 4799 4800 if (val) 4801 hci_dev_set_flag(hdev, HCI_QUALITY_REPORT); 4802 else 4803 hci_dev_clear_flag(hdev, HCI_QUALITY_REPORT); 4804 } 4805 4806 bt_dev_dbg(hdev, "quality report enable %d changed %d", val, changed); 4807 4808 memcpy(rp.uuid, quality_report_uuid, 16); 4809 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4810 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4811 4812 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_EXP_FEATURE, 0, 4813 &rp, sizeof(rp)); 4814 4815 if (changed) 4816 exp_feature_changed(hdev, quality_report_uuid, val, sk); 4817 4818 unlock_quality_report: 4819 hci_req_sync_unlock(hdev); 4820 return err; 4821 } 4822 4823 static int set_offload_codec_func(struct sock *sk, struct hci_dev *hdev, 4824 struct mgmt_cp_set_exp_feature *cp, 4825 u16 data_len) 4826 { 4827 bool val, changed; 4828 int err; 4829 struct mgmt_rp_set_exp_feature rp; 4830 4831 /* Command requires to use a valid controller index */ 4832 if (!hdev) 4833 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4834 MGMT_OP_SET_EXP_FEATURE, 4835 MGMT_STATUS_INVALID_INDEX); 4836 4837 /* Parameters are limited to a single octet */ 4838 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4839 return mgmt_cmd_status(sk, hdev->id, 4840 MGMT_OP_SET_EXP_FEATURE, 4841 MGMT_STATUS_INVALID_PARAMS); 4842 4843 /* Only boolean on/off is supported */ 4844 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4845 return mgmt_cmd_status(sk, hdev->id, 4846 MGMT_OP_SET_EXP_FEATURE, 4847 MGMT_STATUS_INVALID_PARAMS); 4848 4849 val = !!cp->param[0]; 4850 changed = (val != hci_dev_test_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED)); 4851 4852 if (!hdev->get_data_path_id) { 4853 return mgmt_cmd_status(sk, hdev->id, 4854 MGMT_OP_SET_EXP_FEATURE, 4855 MGMT_STATUS_NOT_SUPPORTED); 4856 } 4857 4858 if (changed) { 4859 if (val) 4860 hci_dev_set_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED); 4861 else 4862 hci_dev_clear_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED); 4863 } 4864 4865 bt_dev_info(hdev, "offload codecs enable %d changed %d", 4866 val, changed); 4867 4868 memcpy(rp.uuid, offload_codecs_uuid, 16); 4869 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4870 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4871 err = mgmt_cmd_complete(sk, hdev->id, 4872 MGMT_OP_SET_EXP_FEATURE, 0, 4873 &rp, sizeof(rp)); 4874 4875 if (changed) 4876 exp_feature_changed(hdev, offload_codecs_uuid, val, sk); 4877 4878 return err; 4879 } 4880 4881 static int set_le_simultaneous_roles_func(struct sock *sk, struct hci_dev *hdev, 4882 struct mgmt_cp_set_exp_feature *cp, 4883 u16 data_len) 4884 { 4885 bool val, changed; 4886 int err; 4887 struct mgmt_rp_set_exp_feature rp; 4888 4889 /* Command requires to use a valid controller index */ 4890 if (!hdev) 4891 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4892 MGMT_OP_SET_EXP_FEATURE, 4893 MGMT_STATUS_INVALID_INDEX); 4894 4895 /* Parameters are limited to a single octet */ 4896 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4897 return mgmt_cmd_status(sk, hdev->id, 4898 MGMT_OP_SET_EXP_FEATURE, 4899 MGMT_STATUS_INVALID_PARAMS); 4900 4901 /* Only boolean on/off is supported */ 4902 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4903 return mgmt_cmd_status(sk, hdev->id, 4904 MGMT_OP_SET_EXP_FEATURE, 4905 MGMT_STATUS_INVALID_PARAMS); 4906 4907 val = !!cp->param[0]; 4908 changed = (val != hci_dev_test_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES)); 4909 4910 if (!hci_dev_le_state_simultaneous(hdev)) { 4911 return mgmt_cmd_status(sk, hdev->id, 4912 MGMT_OP_SET_EXP_FEATURE, 4913 MGMT_STATUS_NOT_SUPPORTED); 4914 } 4915 4916 if (changed) { 4917 if (val) 4918 hci_dev_set_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES); 4919 else 4920 hci_dev_clear_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES); 4921 } 4922 4923 bt_dev_info(hdev, "LE simultaneous roles enable %d changed %d", 4924 val, changed); 4925 4926 memcpy(rp.uuid, le_simultaneous_roles_uuid, 16); 4927 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4928 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4929 err = mgmt_cmd_complete(sk, hdev->id, 4930 MGMT_OP_SET_EXP_FEATURE, 0, 4931 &rp, sizeof(rp)); 4932 4933 if (changed) 4934 exp_feature_changed(hdev, le_simultaneous_roles_uuid, val, sk); 4935 4936 return err; 4937 } 4938 4939 #ifdef CONFIG_BT_LE 4940 static int set_iso_socket_func(struct sock *sk, struct hci_dev *hdev, 4941 struct mgmt_cp_set_exp_feature *cp, u16 data_len) 4942 { 4943 struct mgmt_rp_set_exp_feature rp; 4944 bool val, changed = false; 4945 int err; 4946 4947 /* Command requires to use the non-controller index */ 4948 if (hdev) 4949 return mgmt_cmd_status(sk, hdev->id, 4950 MGMT_OP_SET_EXP_FEATURE, 4951 MGMT_STATUS_INVALID_INDEX); 4952 4953 /* Parameters are limited to a single octet */ 4954 if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) 4955 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4956 MGMT_OP_SET_EXP_FEATURE, 4957 MGMT_STATUS_INVALID_PARAMS); 4958 4959 /* Only boolean on/off is supported */ 4960 if (cp->param[0] != 0x00 && cp->param[0] != 0x01) 4961 return mgmt_cmd_status(sk, MGMT_INDEX_NONE, 4962 MGMT_OP_SET_EXP_FEATURE, 4963 MGMT_STATUS_INVALID_PARAMS); 4964 4965 val = cp->param[0] ? true : false; 4966 if (val) 4967 err = iso_init(); 4968 else 4969 err = iso_exit(); 4970 4971 if (!err) 4972 changed = true; 4973 4974 memcpy(rp.uuid, iso_socket_uuid, 16); 4975 rp.flags = cpu_to_le32(val ? BIT(0) : 0); 4976 4977 hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); 4978 4979 err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, 4980 MGMT_OP_SET_EXP_FEATURE, 0, 4981 &rp, sizeof(rp)); 4982 4983 if (changed) 4984 exp_feature_changed(hdev, iso_socket_uuid, val, sk); 4985 4986 return err; 4987 } 4988 #endif 4989 4990 static const struct mgmt_exp_feature { 4991 const u8 *uuid; 4992 int (*set_func)(struct sock *sk, struct hci_dev *hdev, 4993 struct mgmt_cp_set_exp_feature *cp, u16 data_len); 4994 } exp_features[] = { 4995 EXP_FEAT(ZERO_KEY, set_zero_key_func), 4996 #ifdef CONFIG_BT_FEATURE_DEBUG 4997 EXP_FEAT(debug_uuid, set_debug_func), 4998 #endif 4999 EXP_FEAT(mgmt_mesh_uuid, set_mgmt_mesh_func), 5000 EXP_FEAT(rpa_resolution_uuid, set_rpa_resolution_func), 5001 EXP_FEAT(quality_report_uuid, set_quality_report_func), 5002 EXP_FEAT(offload_codecs_uuid, set_offload_codec_func), 5003 EXP_FEAT(le_simultaneous_roles_uuid, set_le_simultaneous_roles_func), 5004 #ifdef CONFIG_BT_LE 5005 EXP_FEAT(iso_socket_uuid, set_iso_socket_func), 5006 #endif 5007 5008 /* end with a null feature */ 5009 EXP_FEAT(NULL, NULL) 5010 }; 5011 5012 static int set_exp_feature(struct sock *sk, struct hci_dev *hdev, 5013 void *data, u16 data_len) 5014 { 5015 struct mgmt_cp_set_exp_feature *cp = data; 5016 size_t i = 0; 5017 5018 bt_dev_dbg(hdev, "sock %p", sk); 5019 5020 for (i = 0; exp_features[i].uuid; i++) { 5021 if (!memcmp(cp->uuid, exp_features[i].uuid, 16)) 5022 return exp_features[i].set_func(sk, hdev, cp, data_len); 5023 } 5024 5025 return mgmt_cmd_status(sk, hdev ? hdev->id : MGMT_INDEX_NONE, 5026 MGMT_OP_SET_EXP_FEATURE, 5027 MGMT_STATUS_NOT_SUPPORTED); 5028 } 5029 5030 static u32 get_params_flags(struct hci_dev *hdev, 5031 struct hci_conn_params *params) 5032 { 5033 u32 flags = hdev->conn_flags; 5034 5035 /* Devices using RPAs can only be programmed in the acceptlist if 5036 * LL Privacy has been enable otherwise they cannot mark 5037 * HCI_CONN_FLAG_REMOTE_WAKEUP. 5038 */ 5039 if ((flags & HCI_CONN_FLAG_REMOTE_WAKEUP) && !use_ll_privacy(hdev) && 5040 hci_find_irk_by_addr(hdev, ¶ms->addr, params->addr_type)) 5041 flags &= ~HCI_CONN_FLAG_REMOTE_WAKEUP; 5042 5043 return flags; 5044 } 5045 5046 static int get_device_flags(struct sock *sk, struct hci_dev *hdev, void *data, 5047 u16 data_len) 5048 { 5049 struct mgmt_cp_get_device_flags *cp = data; 5050 struct mgmt_rp_get_device_flags rp; 5051 struct bdaddr_list_with_flags *br_params; 5052 struct hci_conn_params *params; 5053 u32 supported_flags; 5054 u32 current_flags = 0; 5055 u8 status = MGMT_STATUS_INVALID_PARAMS; 5056 5057 bt_dev_dbg(hdev, "Get device flags %pMR (type 0x%x)\n", 5058 &cp->addr.bdaddr, cp->addr.type); 5059 5060 hci_dev_lock(hdev); 5061 5062 supported_flags = hdev->conn_flags; 5063 5064 memset(&rp, 0, sizeof(rp)); 5065 5066 if (cp->addr.type == BDADDR_BREDR) { 5067 br_params = hci_bdaddr_list_lookup_with_flags(&hdev->accept_list, 5068 &cp->addr.bdaddr, 5069 cp->addr.type); 5070 if (!br_params) 5071 goto done; 5072 5073 current_flags = br_params->flags; 5074 } else { 5075 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 5076 le_addr_type(cp->addr.type)); 5077 if (!params) 5078 goto done; 5079 5080 supported_flags = get_params_flags(hdev, params); 5081 current_flags = params->flags; 5082 } 5083 5084 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 5085 rp.addr.type = cp->addr.type; 5086 rp.supported_flags = cpu_to_le32(supported_flags); 5087 rp.current_flags = cpu_to_le32(current_flags); 5088 5089 status = MGMT_STATUS_SUCCESS; 5090 5091 done: 5092 hci_dev_unlock(hdev); 5093 5094 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_DEVICE_FLAGS, status, 5095 &rp, sizeof(rp)); 5096 } 5097 5098 static void device_flags_changed(struct sock *sk, struct hci_dev *hdev, 5099 bdaddr_t *bdaddr, u8 bdaddr_type, 5100 u32 supported_flags, u32 current_flags) 5101 { 5102 struct mgmt_ev_device_flags_changed ev; 5103 5104 bacpy(&ev.addr.bdaddr, bdaddr); 5105 ev.addr.type = bdaddr_type; 5106 ev.supported_flags = cpu_to_le32(supported_flags); 5107 ev.current_flags = cpu_to_le32(current_flags); 5108 5109 mgmt_event(MGMT_EV_DEVICE_FLAGS_CHANGED, hdev, &ev, sizeof(ev), sk); 5110 } 5111 5112 static int set_device_flags(struct sock *sk, struct hci_dev *hdev, void *data, 5113 u16 len) 5114 { 5115 struct mgmt_cp_set_device_flags *cp = data; 5116 struct bdaddr_list_with_flags *br_params; 5117 struct hci_conn_params *params; 5118 u8 status = MGMT_STATUS_INVALID_PARAMS; 5119 u32 supported_flags; 5120 u32 current_flags = __le32_to_cpu(cp->current_flags); 5121 5122 bt_dev_dbg(hdev, "Set device flags %pMR (type 0x%x) = 0x%x", 5123 &cp->addr.bdaddr, cp->addr.type, current_flags); 5124 5125 // We should take hci_dev_lock() early, I think.. conn_flags can change 5126 supported_flags = hdev->conn_flags; 5127 5128 if ((supported_flags | current_flags) != supported_flags) { 5129 bt_dev_warn(hdev, "Bad flag given (0x%x) vs supported (0x%0x)", 5130 current_flags, supported_flags); 5131 goto done; 5132 } 5133 5134 hci_dev_lock(hdev); 5135 5136 if (cp->addr.type == BDADDR_BREDR) { 5137 br_params = hci_bdaddr_list_lookup_with_flags(&hdev->accept_list, 5138 &cp->addr.bdaddr, 5139 cp->addr.type); 5140 5141 if (br_params) { 5142 br_params->flags = current_flags; 5143 status = MGMT_STATUS_SUCCESS; 5144 } else { 5145 bt_dev_warn(hdev, "No such BR/EDR device %pMR (0x%x)", 5146 &cp->addr.bdaddr, cp->addr.type); 5147 } 5148 5149 goto unlock; 5150 } 5151 5152 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 5153 le_addr_type(cp->addr.type)); 5154 if (!params) { 5155 bt_dev_warn(hdev, "No such LE device %pMR (0x%x)", 5156 &cp->addr.bdaddr, le_addr_type(cp->addr.type)); 5157 goto unlock; 5158 } 5159 5160 supported_flags = get_params_flags(hdev, params); 5161 5162 if ((supported_flags | current_flags) != supported_flags) { 5163 bt_dev_warn(hdev, "Bad flag given (0x%x) vs supported (0x%0x)", 5164 current_flags, supported_flags); 5165 goto unlock; 5166 } 5167 5168 WRITE_ONCE(params->flags, current_flags); 5169 status = MGMT_STATUS_SUCCESS; 5170 5171 /* Update passive scan if HCI_CONN_FLAG_DEVICE_PRIVACY 5172 * has been set. 5173 */ 5174 if (params->flags & HCI_CONN_FLAG_DEVICE_PRIVACY) 5175 hci_update_passive_scan(hdev); 5176 5177 unlock: 5178 hci_dev_unlock(hdev); 5179 5180 done: 5181 if (status == MGMT_STATUS_SUCCESS) 5182 device_flags_changed(sk, hdev, &cp->addr.bdaddr, cp->addr.type, 5183 supported_flags, current_flags); 5184 5185 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEVICE_FLAGS, status, 5186 &cp->addr, sizeof(cp->addr)); 5187 } 5188 5189 static void mgmt_adv_monitor_added(struct sock *sk, struct hci_dev *hdev, 5190 u16 handle) 5191 { 5192 struct mgmt_ev_adv_monitor_added ev; 5193 5194 ev.monitor_handle = cpu_to_le16(handle); 5195 5196 mgmt_event(MGMT_EV_ADV_MONITOR_ADDED, hdev, &ev, sizeof(ev), sk); 5197 } 5198 5199 void mgmt_adv_monitor_removed(struct hci_dev *hdev, u16 handle) 5200 { 5201 struct mgmt_ev_adv_monitor_removed ev; 5202 struct mgmt_pending_cmd *cmd; 5203 struct sock *sk_skip = NULL; 5204 struct mgmt_cp_remove_adv_monitor *cp; 5205 5206 cmd = pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev); 5207 if (cmd) { 5208 cp = cmd->param; 5209 5210 if (cp->monitor_handle) 5211 sk_skip = cmd->sk; 5212 } 5213 5214 ev.monitor_handle = cpu_to_le16(handle); 5215 5216 mgmt_event(MGMT_EV_ADV_MONITOR_REMOVED, hdev, &ev, sizeof(ev), sk_skip); 5217 } 5218 5219 static int read_adv_mon_features(struct sock *sk, struct hci_dev *hdev, 5220 void *data, u16 len) 5221 { 5222 struct adv_monitor *monitor = NULL; 5223 struct mgmt_rp_read_adv_monitor_features *rp = NULL; 5224 int handle, err; 5225 size_t rp_size = 0; 5226 __u32 supported = 0; 5227 __u32 enabled = 0; 5228 __u16 num_handles = 0; 5229 __u16 handles[HCI_MAX_ADV_MONITOR_NUM_HANDLES]; 5230 5231 BT_DBG("request for %s", hdev->name); 5232 5233 hci_dev_lock(hdev); 5234 5235 if (msft_monitor_supported(hdev)) 5236 supported |= MGMT_ADV_MONITOR_FEATURE_MASK_OR_PATTERNS; 5237 5238 idr_for_each_entry(&hdev->adv_monitors_idr, monitor, handle) 5239 handles[num_handles++] = monitor->handle; 5240 5241 hci_dev_unlock(hdev); 5242 5243 rp_size = sizeof(*rp) + (num_handles * sizeof(u16)); 5244 rp = kmalloc(rp_size, GFP_KERNEL); 5245 if (!rp) 5246 return -ENOMEM; 5247 5248 /* All supported features are currently enabled */ 5249 enabled = supported; 5250 5251 rp->supported_features = cpu_to_le32(supported); 5252 rp->enabled_features = cpu_to_le32(enabled); 5253 rp->max_num_handles = cpu_to_le16(HCI_MAX_ADV_MONITOR_NUM_HANDLES); 5254 rp->max_num_patterns = HCI_MAX_ADV_MONITOR_NUM_PATTERNS; 5255 rp->num_handles = cpu_to_le16(num_handles); 5256 if (num_handles) 5257 memcpy(&rp->handles, &handles, (num_handles * sizeof(u16))); 5258 5259 err = mgmt_cmd_complete(sk, hdev->id, 5260 MGMT_OP_READ_ADV_MONITOR_FEATURES, 5261 MGMT_STATUS_SUCCESS, rp, rp_size); 5262 5263 kfree(rp); 5264 5265 return err; 5266 } 5267 5268 static void mgmt_add_adv_patterns_monitor_complete(struct hci_dev *hdev, 5269 void *data, int status) 5270 { 5271 struct mgmt_rp_add_adv_patterns_monitor rp; 5272 struct mgmt_pending_cmd *cmd = data; 5273 struct adv_monitor *monitor = cmd->user_data; 5274 5275 hci_dev_lock(hdev); 5276 5277 rp.monitor_handle = cpu_to_le16(monitor->handle); 5278 5279 if (!status) { 5280 mgmt_adv_monitor_added(cmd->sk, hdev, monitor->handle); 5281 hdev->adv_monitors_cnt++; 5282 if (monitor->state == ADV_MONITOR_STATE_NOT_REGISTERED) 5283 monitor->state = ADV_MONITOR_STATE_REGISTERED; 5284 hci_update_passive_scan(hdev); 5285 } 5286 5287 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 5288 mgmt_status(status), &rp, sizeof(rp)); 5289 mgmt_pending_remove(cmd); 5290 5291 hci_dev_unlock(hdev); 5292 bt_dev_dbg(hdev, "add monitor %d complete, status %d", 5293 rp.monitor_handle, status); 5294 } 5295 5296 static int mgmt_add_adv_patterns_monitor_sync(struct hci_dev *hdev, void *data) 5297 { 5298 struct mgmt_pending_cmd *cmd = data; 5299 struct adv_monitor *monitor = cmd->user_data; 5300 5301 return hci_add_adv_monitor(hdev, monitor); 5302 } 5303 5304 static int __add_adv_patterns_monitor(struct sock *sk, struct hci_dev *hdev, 5305 struct adv_monitor *m, u8 status, 5306 void *data, u16 len, u16 op) 5307 { 5308 struct mgmt_pending_cmd *cmd; 5309 int err; 5310 5311 hci_dev_lock(hdev); 5312 5313 if (status) 5314 goto unlock; 5315 5316 if (pending_find(MGMT_OP_SET_LE, hdev) || 5317 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR, hdev) || 5318 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, hdev) || 5319 pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev)) { 5320 status = MGMT_STATUS_BUSY; 5321 goto unlock; 5322 } 5323 5324 cmd = mgmt_pending_add(sk, op, hdev, data, len); 5325 if (!cmd) { 5326 status = MGMT_STATUS_NO_RESOURCES; 5327 goto unlock; 5328 } 5329 5330 cmd->user_data = m; 5331 err = hci_cmd_sync_queue(hdev, mgmt_add_adv_patterns_monitor_sync, cmd, 5332 mgmt_add_adv_patterns_monitor_complete); 5333 if (err) { 5334 if (err == -ENOMEM) 5335 status = MGMT_STATUS_NO_RESOURCES; 5336 else 5337 status = MGMT_STATUS_FAILED; 5338 5339 goto unlock; 5340 } 5341 5342 hci_dev_unlock(hdev); 5343 5344 return 0; 5345 5346 unlock: 5347 hci_free_adv_monitor(hdev, m); 5348 hci_dev_unlock(hdev); 5349 return mgmt_cmd_status(sk, hdev->id, op, status); 5350 } 5351 5352 static void parse_adv_monitor_rssi(struct adv_monitor *m, 5353 struct mgmt_adv_rssi_thresholds *rssi) 5354 { 5355 if (rssi) { 5356 m->rssi.low_threshold = rssi->low_threshold; 5357 m->rssi.low_threshold_timeout = 5358 __le16_to_cpu(rssi->low_threshold_timeout); 5359 m->rssi.high_threshold = rssi->high_threshold; 5360 m->rssi.high_threshold_timeout = 5361 __le16_to_cpu(rssi->high_threshold_timeout); 5362 m->rssi.sampling_period = rssi->sampling_period; 5363 } else { 5364 /* Default values. These numbers are the least constricting 5365 * parameters for MSFT API to work, so it behaves as if there 5366 * are no rssi parameter to consider. May need to be changed 5367 * if other API are to be supported. 5368 */ 5369 m->rssi.low_threshold = -127; 5370 m->rssi.low_threshold_timeout = 60; 5371 m->rssi.high_threshold = -127; 5372 m->rssi.high_threshold_timeout = 0; 5373 m->rssi.sampling_period = 0; 5374 } 5375 } 5376 5377 static u8 parse_adv_monitor_pattern(struct adv_monitor *m, u8 pattern_count, 5378 struct mgmt_adv_pattern *patterns) 5379 { 5380 u8 offset = 0, length = 0; 5381 struct adv_pattern *p = NULL; 5382 int i; 5383 5384 for (i = 0; i < pattern_count; i++) { 5385 offset = patterns[i].offset; 5386 length = patterns[i].length; 5387 if (offset >= HCI_MAX_EXT_AD_LENGTH || 5388 length > HCI_MAX_EXT_AD_LENGTH || 5389 (offset + length) > HCI_MAX_EXT_AD_LENGTH) 5390 return MGMT_STATUS_INVALID_PARAMS; 5391 5392 p = kmalloc(sizeof(*p), GFP_KERNEL); 5393 if (!p) 5394 return MGMT_STATUS_NO_RESOURCES; 5395 5396 p->ad_type = patterns[i].ad_type; 5397 p->offset = patterns[i].offset; 5398 p->length = patterns[i].length; 5399 memcpy(p->value, patterns[i].value, p->length); 5400 5401 INIT_LIST_HEAD(&p->list); 5402 list_add(&p->list, &m->patterns); 5403 } 5404 5405 return MGMT_STATUS_SUCCESS; 5406 } 5407 5408 static int add_adv_patterns_monitor(struct sock *sk, struct hci_dev *hdev, 5409 void *data, u16 len) 5410 { 5411 struct mgmt_cp_add_adv_patterns_monitor *cp = data; 5412 struct adv_monitor *m = NULL; 5413 u8 status = MGMT_STATUS_SUCCESS; 5414 size_t expected_size = sizeof(*cp); 5415 5416 BT_DBG("request for %s", hdev->name); 5417 5418 if (len <= sizeof(*cp)) { 5419 status = MGMT_STATUS_INVALID_PARAMS; 5420 goto done; 5421 } 5422 5423 expected_size += cp->pattern_count * sizeof(struct mgmt_adv_pattern); 5424 if (len != expected_size) { 5425 status = MGMT_STATUS_INVALID_PARAMS; 5426 goto done; 5427 } 5428 5429 m = kzalloc(sizeof(*m), GFP_KERNEL); 5430 if (!m) { 5431 status = MGMT_STATUS_NO_RESOURCES; 5432 goto done; 5433 } 5434 5435 INIT_LIST_HEAD(&m->patterns); 5436 5437 parse_adv_monitor_rssi(m, NULL); 5438 status = parse_adv_monitor_pattern(m, cp->pattern_count, cp->patterns); 5439 5440 done: 5441 return __add_adv_patterns_monitor(sk, hdev, m, status, data, len, 5442 MGMT_OP_ADD_ADV_PATTERNS_MONITOR); 5443 } 5444 5445 static int add_adv_patterns_monitor_rssi(struct sock *sk, struct hci_dev *hdev, 5446 void *data, u16 len) 5447 { 5448 struct mgmt_cp_add_adv_patterns_monitor_rssi *cp = data; 5449 struct adv_monitor *m = NULL; 5450 u8 status = MGMT_STATUS_SUCCESS; 5451 size_t expected_size = sizeof(*cp); 5452 5453 BT_DBG("request for %s", hdev->name); 5454 5455 if (len <= sizeof(*cp)) { 5456 status = MGMT_STATUS_INVALID_PARAMS; 5457 goto done; 5458 } 5459 5460 expected_size += cp->pattern_count * sizeof(struct mgmt_adv_pattern); 5461 if (len != expected_size) { 5462 status = MGMT_STATUS_INVALID_PARAMS; 5463 goto done; 5464 } 5465 5466 m = kzalloc(sizeof(*m), GFP_KERNEL); 5467 if (!m) { 5468 status = MGMT_STATUS_NO_RESOURCES; 5469 goto done; 5470 } 5471 5472 INIT_LIST_HEAD(&m->patterns); 5473 5474 parse_adv_monitor_rssi(m, &cp->rssi); 5475 status = parse_adv_monitor_pattern(m, cp->pattern_count, cp->patterns); 5476 5477 done: 5478 return __add_adv_patterns_monitor(sk, hdev, m, status, data, len, 5479 MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI); 5480 } 5481 5482 static void mgmt_remove_adv_monitor_complete(struct hci_dev *hdev, 5483 void *data, int status) 5484 { 5485 struct mgmt_rp_remove_adv_monitor rp; 5486 struct mgmt_pending_cmd *cmd = data; 5487 struct mgmt_cp_remove_adv_monitor *cp = cmd->param; 5488 5489 hci_dev_lock(hdev); 5490 5491 rp.monitor_handle = cp->monitor_handle; 5492 5493 if (!status) 5494 hci_update_passive_scan(hdev); 5495 5496 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 5497 mgmt_status(status), &rp, sizeof(rp)); 5498 mgmt_pending_remove(cmd); 5499 5500 hci_dev_unlock(hdev); 5501 bt_dev_dbg(hdev, "remove monitor %d complete, status %d", 5502 rp.monitor_handle, status); 5503 } 5504 5505 static int mgmt_remove_adv_monitor_sync(struct hci_dev *hdev, void *data) 5506 { 5507 struct mgmt_pending_cmd *cmd = data; 5508 struct mgmt_cp_remove_adv_monitor *cp = cmd->param; 5509 u16 handle = __le16_to_cpu(cp->monitor_handle); 5510 5511 if (!handle) 5512 return hci_remove_all_adv_monitor(hdev); 5513 5514 return hci_remove_single_adv_monitor(hdev, handle); 5515 } 5516 5517 static int remove_adv_monitor(struct sock *sk, struct hci_dev *hdev, 5518 void *data, u16 len) 5519 { 5520 struct mgmt_pending_cmd *cmd; 5521 int err, status; 5522 5523 hci_dev_lock(hdev); 5524 5525 if (pending_find(MGMT_OP_SET_LE, hdev) || 5526 pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev) || 5527 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR, hdev) || 5528 pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, hdev)) { 5529 status = MGMT_STATUS_BUSY; 5530 goto unlock; 5531 } 5532 5533 cmd = mgmt_pending_add(sk, MGMT_OP_REMOVE_ADV_MONITOR, hdev, data, len); 5534 if (!cmd) { 5535 status = MGMT_STATUS_NO_RESOURCES; 5536 goto unlock; 5537 } 5538 5539 err = hci_cmd_sync_queue(hdev, mgmt_remove_adv_monitor_sync, cmd, 5540 mgmt_remove_adv_monitor_complete); 5541 5542 if (err) { 5543 mgmt_pending_remove(cmd); 5544 5545 if (err == -ENOMEM) 5546 status = MGMT_STATUS_NO_RESOURCES; 5547 else 5548 status = MGMT_STATUS_FAILED; 5549 5550 goto unlock; 5551 } 5552 5553 hci_dev_unlock(hdev); 5554 5555 return 0; 5556 5557 unlock: 5558 hci_dev_unlock(hdev); 5559 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADV_MONITOR, 5560 status); 5561 } 5562 5563 static void read_local_oob_data_complete(struct hci_dev *hdev, void *data, int err) 5564 { 5565 struct mgmt_rp_read_local_oob_data mgmt_rp; 5566 size_t rp_size = sizeof(mgmt_rp); 5567 struct mgmt_pending_cmd *cmd = data; 5568 struct sk_buff *skb = cmd->skb; 5569 u8 status = mgmt_status(err); 5570 5571 if (!status) { 5572 if (!skb) 5573 status = MGMT_STATUS_FAILED; 5574 else if (IS_ERR(skb)) 5575 status = mgmt_status(PTR_ERR(skb)); 5576 else 5577 status = mgmt_status(skb->data[0]); 5578 } 5579 5580 bt_dev_dbg(hdev, "status %d", status); 5581 5582 if (status) { 5583 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, status); 5584 goto remove; 5585 } 5586 5587 memset(&mgmt_rp, 0, sizeof(mgmt_rp)); 5588 5589 if (!bredr_sc_enabled(hdev)) { 5590 struct hci_rp_read_local_oob_data *rp = (void *) skb->data; 5591 5592 if (skb->len < sizeof(*rp)) { 5593 mgmt_cmd_status(cmd->sk, hdev->id, 5594 MGMT_OP_READ_LOCAL_OOB_DATA, 5595 MGMT_STATUS_FAILED); 5596 goto remove; 5597 } 5598 5599 memcpy(mgmt_rp.hash192, rp->hash, sizeof(rp->hash)); 5600 memcpy(mgmt_rp.rand192, rp->rand, sizeof(rp->rand)); 5601 5602 rp_size -= sizeof(mgmt_rp.hash256) + sizeof(mgmt_rp.rand256); 5603 } else { 5604 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data; 5605 5606 if (skb->len < sizeof(*rp)) { 5607 mgmt_cmd_status(cmd->sk, hdev->id, 5608 MGMT_OP_READ_LOCAL_OOB_DATA, 5609 MGMT_STATUS_FAILED); 5610 goto remove; 5611 } 5612 5613 memcpy(mgmt_rp.hash192, rp->hash192, sizeof(rp->hash192)); 5614 memcpy(mgmt_rp.rand192, rp->rand192, sizeof(rp->rand192)); 5615 5616 memcpy(mgmt_rp.hash256, rp->hash256, sizeof(rp->hash256)); 5617 memcpy(mgmt_rp.rand256, rp->rand256, sizeof(rp->rand256)); 5618 } 5619 5620 mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5621 MGMT_STATUS_SUCCESS, &mgmt_rp, rp_size); 5622 5623 remove: 5624 if (skb && !IS_ERR(skb)) 5625 kfree_skb(skb); 5626 5627 mgmt_pending_free(cmd); 5628 } 5629 5630 static int read_local_oob_data_sync(struct hci_dev *hdev, void *data) 5631 { 5632 struct mgmt_pending_cmd *cmd = data; 5633 5634 if (bredr_sc_enabled(hdev)) 5635 cmd->skb = hci_read_local_oob_data_sync(hdev, true, cmd->sk); 5636 else 5637 cmd->skb = hci_read_local_oob_data_sync(hdev, false, cmd->sk); 5638 5639 if (IS_ERR(cmd->skb)) 5640 return PTR_ERR(cmd->skb); 5641 else 5642 return 0; 5643 } 5644 5645 static int read_local_oob_data(struct sock *sk, struct hci_dev *hdev, 5646 void *data, u16 data_len) 5647 { 5648 struct mgmt_pending_cmd *cmd; 5649 int err; 5650 5651 bt_dev_dbg(hdev, "sock %p", sk); 5652 5653 hci_dev_lock(hdev); 5654 5655 if (!hdev_is_powered(hdev)) { 5656 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5657 MGMT_STATUS_NOT_POWERED); 5658 goto unlock; 5659 } 5660 5661 if (!lmp_ssp_capable(hdev)) { 5662 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5663 MGMT_STATUS_NOT_SUPPORTED); 5664 goto unlock; 5665 } 5666 5667 cmd = mgmt_pending_new(sk, MGMT_OP_READ_LOCAL_OOB_DATA, hdev, NULL, 0); 5668 if (!cmd) 5669 err = -ENOMEM; 5670 else 5671 err = hci_cmd_sync_queue(hdev, read_local_oob_data_sync, cmd, 5672 read_local_oob_data_complete); 5673 5674 if (err < 0) { 5675 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, 5676 MGMT_STATUS_FAILED); 5677 5678 if (cmd) 5679 mgmt_pending_free(cmd); 5680 } 5681 5682 unlock: 5683 hci_dev_unlock(hdev); 5684 return err; 5685 } 5686 5687 static int add_remote_oob_data(struct sock *sk, struct hci_dev *hdev, 5688 void *data, u16 len) 5689 { 5690 struct mgmt_addr_info *addr = data; 5691 int err; 5692 5693 bt_dev_dbg(hdev, "sock %p", sk); 5694 5695 if (!bdaddr_type_is_valid(addr->type)) 5696 return mgmt_cmd_complete(sk, hdev->id, 5697 MGMT_OP_ADD_REMOTE_OOB_DATA, 5698 MGMT_STATUS_INVALID_PARAMS, 5699 addr, sizeof(*addr)); 5700 5701 hci_dev_lock(hdev); 5702 5703 if (len == MGMT_ADD_REMOTE_OOB_DATA_SIZE) { 5704 struct mgmt_cp_add_remote_oob_data *cp = data; 5705 u8 status; 5706 5707 if (cp->addr.type != BDADDR_BREDR) { 5708 err = mgmt_cmd_complete(sk, hdev->id, 5709 MGMT_OP_ADD_REMOTE_OOB_DATA, 5710 MGMT_STATUS_INVALID_PARAMS, 5711 &cp->addr, sizeof(cp->addr)); 5712 goto unlock; 5713 } 5714 5715 err = hci_add_remote_oob_data(hdev, &cp->addr.bdaddr, 5716 cp->addr.type, cp->hash, 5717 cp->rand, NULL, NULL); 5718 if (err < 0) 5719 status = MGMT_STATUS_FAILED; 5720 else 5721 status = MGMT_STATUS_SUCCESS; 5722 5723 err = mgmt_cmd_complete(sk, hdev->id, 5724 MGMT_OP_ADD_REMOTE_OOB_DATA, status, 5725 &cp->addr, sizeof(cp->addr)); 5726 } else if (len == MGMT_ADD_REMOTE_OOB_EXT_DATA_SIZE) { 5727 struct mgmt_cp_add_remote_oob_ext_data *cp = data; 5728 u8 *rand192, *hash192, *rand256, *hash256; 5729 u8 status; 5730 5731 if (bdaddr_type_is_le(cp->addr.type)) { 5732 /* Enforce zero-valued 192-bit parameters as 5733 * long as legacy SMP OOB isn't implemented. 5734 */ 5735 if (memcmp(cp->rand192, ZERO_KEY, 16) || 5736 memcmp(cp->hash192, ZERO_KEY, 16)) { 5737 err = mgmt_cmd_complete(sk, hdev->id, 5738 MGMT_OP_ADD_REMOTE_OOB_DATA, 5739 MGMT_STATUS_INVALID_PARAMS, 5740 addr, sizeof(*addr)); 5741 goto unlock; 5742 } 5743 5744 rand192 = NULL; 5745 hash192 = NULL; 5746 } else { 5747 /* In case one of the P-192 values is set to zero, 5748 * then just disable OOB data for P-192. 5749 */ 5750 if (!memcmp(cp->rand192, ZERO_KEY, 16) || 5751 !memcmp(cp->hash192, ZERO_KEY, 16)) { 5752 rand192 = NULL; 5753 hash192 = NULL; 5754 } else { 5755 rand192 = cp->rand192; 5756 hash192 = cp->hash192; 5757 } 5758 } 5759 5760 /* In case one of the P-256 values is set to zero, then just 5761 * disable OOB data for P-256. 5762 */ 5763 if (!memcmp(cp->rand256, ZERO_KEY, 16) || 5764 !memcmp(cp->hash256, ZERO_KEY, 16)) { 5765 rand256 = NULL; 5766 hash256 = NULL; 5767 } else { 5768 rand256 = cp->rand256; 5769 hash256 = cp->hash256; 5770 } 5771 5772 err = hci_add_remote_oob_data(hdev, &cp->addr.bdaddr, 5773 cp->addr.type, hash192, rand192, 5774 hash256, rand256); 5775 if (err < 0) 5776 status = MGMT_STATUS_FAILED; 5777 else 5778 status = MGMT_STATUS_SUCCESS; 5779 5780 err = mgmt_cmd_complete(sk, hdev->id, 5781 MGMT_OP_ADD_REMOTE_OOB_DATA, 5782 status, &cp->addr, sizeof(cp->addr)); 5783 } else { 5784 bt_dev_err(hdev, "add_remote_oob_data: invalid len of %u bytes", 5785 len); 5786 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_REMOTE_OOB_DATA, 5787 MGMT_STATUS_INVALID_PARAMS); 5788 } 5789 5790 unlock: 5791 hci_dev_unlock(hdev); 5792 return err; 5793 } 5794 5795 static int remove_remote_oob_data(struct sock *sk, struct hci_dev *hdev, 5796 void *data, u16 len) 5797 { 5798 struct mgmt_cp_remove_remote_oob_data *cp = data; 5799 u8 status; 5800 int err; 5801 5802 bt_dev_dbg(hdev, "sock %p", sk); 5803 5804 if (cp->addr.type != BDADDR_BREDR) 5805 return mgmt_cmd_complete(sk, hdev->id, 5806 MGMT_OP_REMOVE_REMOTE_OOB_DATA, 5807 MGMT_STATUS_INVALID_PARAMS, 5808 &cp->addr, sizeof(cp->addr)); 5809 5810 hci_dev_lock(hdev); 5811 5812 if (!bacmp(&cp->addr.bdaddr, BDADDR_ANY)) { 5813 hci_remote_oob_data_clear(hdev); 5814 status = MGMT_STATUS_SUCCESS; 5815 goto done; 5816 } 5817 5818 err = hci_remove_remote_oob_data(hdev, &cp->addr.bdaddr, cp->addr.type); 5819 if (err < 0) 5820 status = MGMT_STATUS_INVALID_PARAMS; 5821 else 5822 status = MGMT_STATUS_SUCCESS; 5823 5824 done: 5825 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_REMOVE_REMOTE_OOB_DATA, 5826 status, &cp->addr, sizeof(cp->addr)); 5827 5828 hci_dev_unlock(hdev); 5829 return err; 5830 } 5831 5832 void mgmt_start_discovery_complete(struct hci_dev *hdev, u8 status) 5833 { 5834 struct mgmt_pending_cmd *cmd; 5835 5836 bt_dev_dbg(hdev, "status %u", status); 5837 5838 hci_dev_lock(hdev); 5839 5840 cmd = pending_find(MGMT_OP_START_DISCOVERY, hdev); 5841 if (!cmd) 5842 cmd = pending_find(MGMT_OP_START_SERVICE_DISCOVERY, hdev); 5843 5844 if (!cmd) 5845 cmd = pending_find(MGMT_OP_START_LIMITED_DISCOVERY, hdev); 5846 5847 if (cmd) { 5848 cmd->cmd_complete(cmd, mgmt_status(status)); 5849 mgmt_pending_remove(cmd); 5850 } 5851 5852 hci_dev_unlock(hdev); 5853 } 5854 5855 static bool discovery_type_is_valid(struct hci_dev *hdev, uint8_t type, 5856 uint8_t *mgmt_status) 5857 { 5858 switch (type) { 5859 case DISCOV_TYPE_LE: 5860 *mgmt_status = mgmt_le_support(hdev); 5861 if (*mgmt_status) 5862 return false; 5863 break; 5864 case DISCOV_TYPE_INTERLEAVED: 5865 *mgmt_status = mgmt_le_support(hdev); 5866 if (*mgmt_status) 5867 return false; 5868 fallthrough; 5869 case DISCOV_TYPE_BREDR: 5870 *mgmt_status = mgmt_bredr_support(hdev); 5871 if (*mgmt_status) 5872 return false; 5873 break; 5874 default: 5875 *mgmt_status = MGMT_STATUS_INVALID_PARAMS; 5876 return false; 5877 } 5878 5879 return true; 5880 } 5881 5882 static void start_discovery_complete(struct hci_dev *hdev, void *data, int err) 5883 { 5884 struct mgmt_pending_cmd *cmd = data; 5885 5886 if (cmd != pending_find(MGMT_OP_START_DISCOVERY, hdev) && 5887 cmd != pending_find(MGMT_OP_START_LIMITED_DISCOVERY, hdev) && 5888 cmd != pending_find(MGMT_OP_START_SERVICE_DISCOVERY, hdev)) 5889 return; 5890 5891 bt_dev_dbg(hdev, "err %d", err); 5892 5893 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, mgmt_status(err), 5894 cmd->param, 1); 5895 mgmt_pending_remove(cmd); 5896 5897 hci_discovery_set_state(hdev, err ? DISCOVERY_STOPPED: 5898 DISCOVERY_FINDING); 5899 } 5900 5901 static int start_discovery_sync(struct hci_dev *hdev, void *data) 5902 { 5903 return hci_start_discovery_sync(hdev); 5904 } 5905 5906 static int start_discovery_internal(struct sock *sk, struct hci_dev *hdev, 5907 u16 op, void *data, u16 len) 5908 { 5909 struct mgmt_cp_start_discovery *cp = data; 5910 struct mgmt_pending_cmd *cmd; 5911 u8 status; 5912 int err; 5913 5914 bt_dev_dbg(hdev, "sock %p", sk); 5915 5916 hci_dev_lock(hdev); 5917 5918 if (!hdev_is_powered(hdev)) { 5919 err = mgmt_cmd_complete(sk, hdev->id, op, 5920 MGMT_STATUS_NOT_POWERED, 5921 &cp->type, sizeof(cp->type)); 5922 goto failed; 5923 } 5924 5925 if (hdev->discovery.state != DISCOVERY_STOPPED || 5926 hci_dev_test_flag(hdev, HCI_PERIODIC_INQ)) { 5927 err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY, 5928 &cp->type, sizeof(cp->type)); 5929 goto failed; 5930 } 5931 5932 if (!discovery_type_is_valid(hdev, cp->type, &status)) { 5933 err = mgmt_cmd_complete(sk, hdev->id, op, status, 5934 &cp->type, sizeof(cp->type)); 5935 goto failed; 5936 } 5937 5938 /* Can't start discovery when it is paused */ 5939 if (hdev->discovery_paused) { 5940 err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY, 5941 &cp->type, sizeof(cp->type)); 5942 goto failed; 5943 } 5944 5945 /* Clear the discovery filter first to free any previously 5946 * allocated memory for the UUID list. 5947 */ 5948 hci_discovery_filter_clear(hdev); 5949 5950 hdev->discovery.type = cp->type; 5951 hdev->discovery.report_invalid_rssi = false; 5952 if (op == MGMT_OP_START_LIMITED_DISCOVERY) 5953 hdev->discovery.limited = true; 5954 else 5955 hdev->discovery.limited = false; 5956 5957 cmd = mgmt_pending_add(sk, op, hdev, data, len); 5958 if (!cmd) { 5959 err = -ENOMEM; 5960 goto failed; 5961 } 5962 5963 err = hci_cmd_sync_queue(hdev, start_discovery_sync, cmd, 5964 start_discovery_complete); 5965 if (err < 0) { 5966 mgmt_pending_remove(cmd); 5967 goto failed; 5968 } 5969 5970 hci_discovery_set_state(hdev, DISCOVERY_STARTING); 5971 5972 failed: 5973 hci_dev_unlock(hdev); 5974 return err; 5975 } 5976 5977 static int start_discovery(struct sock *sk, struct hci_dev *hdev, 5978 void *data, u16 len) 5979 { 5980 return start_discovery_internal(sk, hdev, MGMT_OP_START_DISCOVERY, 5981 data, len); 5982 } 5983 5984 static int start_limited_discovery(struct sock *sk, struct hci_dev *hdev, 5985 void *data, u16 len) 5986 { 5987 return start_discovery_internal(sk, hdev, 5988 MGMT_OP_START_LIMITED_DISCOVERY, 5989 data, len); 5990 } 5991 5992 static int start_service_discovery(struct sock *sk, struct hci_dev *hdev, 5993 void *data, u16 len) 5994 { 5995 struct mgmt_cp_start_service_discovery *cp = data; 5996 struct mgmt_pending_cmd *cmd; 5997 const u16 max_uuid_count = ((U16_MAX - sizeof(*cp)) / 16); 5998 u16 uuid_count, expected_len; 5999 u8 status; 6000 int err; 6001 6002 bt_dev_dbg(hdev, "sock %p", sk); 6003 6004 hci_dev_lock(hdev); 6005 6006 if (!hdev_is_powered(hdev)) { 6007 err = mgmt_cmd_complete(sk, hdev->id, 6008 MGMT_OP_START_SERVICE_DISCOVERY, 6009 MGMT_STATUS_NOT_POWERED, 6010 &cp->type, sizeof(cp->type)); 6011 goto failed; 6012 } 6013 6014 if (hdev->discovery.state != DISCOVERY_STOPPED || 6015 hci_dev_test_flag(hdev, HCI_PERIODIC_INQ)) { 6016 err = mgmt_cmd_complete(sk, hdev->id, 6017 MGMT_OP_START_SERVICE_DISCOVERY, 6018 MGMT_STATUS_BUSY, &cp->type, 6019 sizeof(cp->type)); 6020 goto failed; 6021 } 6022 6023 if (hdev->discovery_paused) { 6024 err = mgmt_cmd_complete(sk, hdev->id, 6025 MGMT_OP_START_SERVICE_DISCOVERY, 6026 MGMT_STATUS_BUSY, &cp->type, 6027 sizeof(cp->type)); 6028 goto failed; 6029 } 6030 6031 uuid_count = __le16_to_cpu(cp->uuid_count); 6032 if (uuid_count > max_uuid_count) { 6033 bt_dev_err(hdev, "service_discovery: too big uuid_count value %u", 6034 uuid_count); 6035 err = mgmt_cmd_complete(sk, hdev->id, 6036 MGMT_OP_START_SERVICE_DISCOVERY, 6037 MGMT_STATUS_INVALID_PARAMS, &cp->type, 6038 sizeof(cp->type)); 6039 goto failed; 6040 } 6041 6042 expected_len = sizeof(*cp) + uuid_count * 16; 6043 if (expected_len != len) { 6044 bt_dev_err(hdev, "service_discovery: expected %u bytes, got %u bytes", 6045 expected_len, len); 6046 err = mgmt_cmd_complete(sk, hdev->id, 6047 MGMT_OP_START_SERVICE_DISCOVERY, 6048 MGMT_STATUS_INVALID_PARAMS, &cp->type, 6049 sizeof(cp->type)); 6050 goto failed; 6051 } 6052 6053 if (!discovery_type_is_valid(hdev, cp->type, &status)) { 6054 err = mgmt_cmd_complete(sk, hdev->id, 6055 MGMT_OP_START_SERVICE_DISCOVERY, 6056 status, &cp->type, sizeof(cp->type)); 6057 goto failed; 6058 } 6059 6060 cmd = mgmt_pending_add(sk, MGMT_OP_START_SERVICE_DISCOVERY, 6061 hdev, data, len); 6062 if (!cmd) { 6063 err = -ENOMEM; 6064 goto failed; 6065 } 6066 6067 /* Clear the discovery filter first to free any previously 6068 * allocated memory for the UUID list. 6069 */ 6070 hci_discovery_filter_clear(hdev); 6071 6072 hdev->discovery.result_filtering = true; 6073 hdev->discovery.type = cp->type; 6074 hdev->discovery.rssi = cp->rssi; 6075 hdev->discovery.uuid_count = uuid_count; 6076 6077 if (uuid_count > 0) { 6078 hdev->discovery.uuids = kmemdup(cp->uuids, uuid_count * 16, 6079 GFP_KERNEL); 6080 if (!hdev->discovery.uuids) { 6081 err = mgmt_cmd_complete(sk, hdev->id, 6082 MGMT_OP_START_SERVICE_DISCOVERY, 6083 MGMT_STATUS_FAILED, 6084 &cp->type, sizeof(cp->type)); 6085 mgmt_pending_remove(cmd); 6086 goto failed; 6087 } 6088 } 6089 6090 err = hci_cmd_sync_queue(hdev, start_discovery_sync, cmd, 6091 start_discovery_complete); 6092 if (err < 0) { 6093 mgmt_pending_remove(cmd); 6094 goto failed; 6095 } 6096 6097 hci_discovery_set_state(hdev, DISCOVERY_STARTING); 6098 6099 failed: 6100 hci_dev_unlock(hdev); 6101 return err; 6102 } 6103 6104 void mgmt_stop_discovery_complete(struct hci_dev *hdev, u8 status) 6105 { 6106 struct mgmt_pending_cmd *cmd; 6107 6108 bt_dev_dbg(hdev, "status %u", status); 6109 6110 hci_dev_lock(hdev); 6111 6112 cmd = pending_find(MGMT_OP_STOP_DISCOVERY, hdev); 6113 if (cmd) { 6114 cmd->cmd_complete(cmd, mgmt_status(status)); 6115 mgmt_pending_remove(cmd); 6116 } 6117 6118 hci_dev_unlock(hdev); 6119 } 6120 6121 static void stop_discovery_complete(struct hci_dev *hdev, void *data, int err) 6122 { 6123 struct mgmt_pending_cmd *cmd = data; 6124 6125 if (cmd != pending_find(MGMT_OP_STOP_DISCOVERY, hdev)) 6126 return; 6127 6128 bt_dev_dbg(hdev, "err %d", err); 6129 6130 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, mgmt_status(err), 6131 cmd->param, 1); 6132 mgmt_pending_remove(cmd); 6133 6134 if (!err) 6135 hci_discovery_set_state(hdev, DISCOVERY_STOPPED); 6136 } 6137 6138 static int stop_discovery_sync(struct hci_dev *hdev, void *data) 6139 { 6140 return hci_stop_discovery_sync(hdev); 6141 } 6142 6143 static int stop_discovery(struct sock *sk, struct hci_dev *hdev, void *data, 6144 u16 len) 6145 { 6146 struct mgmt_cp_stop_discovery *mgmt_cp = data; 6147 struct mgmt_pending_cmd *cmd; 6148 int err; 6149 6150 bt_dev_dbg(hdev, "sock %p", sk); 6151 6152 hci_dev_lock(hdev); 6153 6154 if (!hci_discovery_active(hdev)) { 6155 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_STOP_DISCOVERY, 6156 MGMT_STATUS_REJECTED, &mgmt_cp->type, 6157 sizeof(mgmt_cp->type)); 6158 goto unlock; 6159 } 6160 6161 if (hdev->discovery.type != mgmt_cp->type) { 6162 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_STOP_DISCOVERY, 6163 MGMT_STATUS_INVALID_PARAMS, 6164 &mgmt_cp->type, sizeof(mgmt_cp->type)); 6165 goto unlock; 6166 } 6167 6168 cmd = mgmt_pending_add(sk, MGMT_OP_STOP_DISCOVERY, hdev, data, len); 6169 if (!cmd) { 6170 err = -ENOMEM; 6171 goto unlock; 6172 } 6173 6174 err = hci_cmd_sync_queue(hdev, stop_discovery_sync, cmd, 6175 stop_discovery_complete); 6176 if (err < 0) { 6177 mgmt_pending_remove(cmd); 6178 goto unlock; 6179 } 6180 6181 hci_discovery_set_state(hdev, DISCOVERY_STOPPING); 6182 6183 unlock: 6184 hci_dev_unlock(hdev); 6185 return err; 6186 } 6187 6188 static int confirm_name(struct sock *sk, struct hci_dev *hdev, void *data, 6189 u16 len) 6190 { 6191 struct mgmt_cp_confirm_name *cp = data; 6192 struct inquiry_entry *e; 6193 int err; 6194 6195 bt_dev_dbg(hdev, "sock %p", sk); 6196 6197 hci_dev_lock(hdev); 6198 6199 if (!hci_discovery_active(hdev)) { 6200 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME, 6201 MGMT_STATUS_FAILED, &cp->addr, 6202 sizeof(cp->addr)); 6203 goto failed; 6204 } 6205 6206 e = hci_inquiry_cache_lookup_unknown(hdev, &cp->addr.bdaddr); 6207 if (!e) { 6208 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME, 6209 MGMT_STATUS_INVALID_PARAMS, &cp->addr, 6210 sizeof(cp->addr)); 6211 goto failed; 6212 } 6213 6214 if (cp->name_known) { 6215 e->name_state = NAME_KNOWN; 6216 list_del(&e->list); 6217 } else { 6218 e->name_state = NAME_NEEDED; 6219 hci_inquiry_cache_update_resolve(hdev, e); 6220 } 6221 6222 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME, 0, 6223 &cp->addr, sizeof(cp->addr)); 6224 6225 failed: 6226 hci_dev_unlock(hdev); 6227 return err; 6228 } 6229 6230 static int block_device(struct sock *sk, struct hci_dev *hdev, void *data, 6231 u16 len) 6232 { 6233 struct mgmt_cp_block_device *cp = data; 6234 u8 status; 6235 int err; 6236 6237 bt_dev_dbg(hdev, "sock %p", sk); 6238 6239 if (!bdaddr_type_is_valid(cp->addr.type)) 6240 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_BLOCK_DEVICE, 6241 MGMT_STATUS_INVALID_PARAMS, 6242 &cp->addr, sizeof(cp->addr)); 6243 6244 hci_dev_lock(hdev); 6245 6246 err = hci_bdaddr_list_add(&hdev->reject_list, &cp->addr.bdaddr, 6247 cp->addr.type); 6248 if (err < 0) { 6249 status = MGMT_STATUS_FAILED; 6250 goto done; 6251 } 6252 6253 mgmt_event(MGMT_EV_DEVICE_BLOCKED, hdev, &cp->addr, sizeof(cp->addr), 6254 sk); 6255 status = MGMT_STATUS_SUCCESS; 6256 6257 done: 6258 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_BLOCK_DEVICE, status, 6259 &cp->addr, sizeof(cp->addr)); 6260 6261 hci_dev_unlock(hdev); 6262 6263 return err; 6264 } 6265 6266 static int unblock_device(struct sock *sk, struct hci_dev *hdev, void *data, 6267 u16 len) 6268 { 6269 struct mgmt_cp_unblock_device *cp = data; 6270 u8 status; 6271 int err; 6272 6273 bt_dev_dbg(hdev, "sock %p", sk); 6274 6275 if (!bdaddr_type_is_valid(cp->addr.type)) 6276 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNBLOCK_DEVICE, 6277 MGMT_STATUS_INVALID_PARAMS, 6278 &cp->addr, sizeof(cp->addr)); 6279 6280 hci_dev_lock(hdev); 6281 6282 err = hci_bdaddr_list_del(&hdev->reject_list, &cp->addr.bdaddr, 6283 cp->addr.type); 6284 if (err < 0) { 6285 status = MGMT_STATUS_INVALID_PARAMS; 6286 goto done; 6287 } 6288 6289 mgmt_event(MGMT_EV_DEVICE_UNBLOCKED, hdev, &cp->addr, sizeof(cp->addr), 6290 sk); 6291 status = MGMT_STATUS_SUCCESS; 6292 6293 done: 6294 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNBLOCK_DEVICE, status, 6295 &cp->addr, sizeof(cp->addr)); 6296 6297 hci_dev_unlock(hdev); 6298 6299 return err; 6300 } 6301 6302 static int set_device_id_sync(struct hci_dev *hdev, void *data) 6303 { 6304 return hci_update_eir_sync(hdev); 6305 } 6306 6307 static int set_device_id(struct sock *sk, struct hci_dev *hdev, void *data, 6308 u16 len) 6309 { 6310 struct mgmt_cp_set_device_id *cp = data; 6311 int err; 6312 __u16 source; 6313 6314 bt_dev_dbg(hdev, "sock %p", sk); 6315 6316 source = __le16_to_cpu(cp->source); 6317 6318 if (source > 0x0002) 6319 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEVICE_ID, 6320 MGMT_STATUS_INVALID_PARAMS); 6321 6322 hci_dev_lock(hdev); 6323 6324 hdev->devid_source = source; 6325 hdev->devid_vendor = __le16_to_cpu(cp->vendor); 6326 hdev->devid_product = __le16_to_cpu(cp->product); 6327 hdev->devid_version = __le16_to_cpu(cp->version); 6328 6329 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEVICE_ID, 0, 6330 NULL, 0); 6331 6332 hci_cmd_sync_queue(hdev, set_device_id_sync, NULL, NULL); 6333 6334 hci_dev_unlock(hdev); 6335 6336 return err; 6337 } 6338 6339 static void enable_advertising_instance(struct hci_dev *hdev, int err) 6340 { 6341 if (err) 6342 bt_dev_err(hdev, "failed to re-configure advertising %d", err); 6343 else 6344 bt_dev_dbg(hdev, "status %d", err); 6345 } 6346 6347 static void set_advertising_complete(struct hci_dev *hdev, void *data, int err) 6348 { 6349 struct cmd_lookup match = { NULL, hdev }; 6350 u8 instance; 6351 struct adv_info *adv_instance; 6352 u8 status = mgmt_status(err); 6353 6354 if (status) { 6355 mgmt_pending_foreach(MGMT_OP_SET_ADVERTISING, hdev, 6356 cmd_status_rsp, &status); 6357 return; 6358 } 6359 6360 if (hci_dev_test_flag(hdev, HCI_LE_ADV)) 6361 hci_dev_set_flag(hdev, HCI_ADVERTISING); 6362 else 6363 hci_dev_clear_flag(hdev, HCI_ADVERTISING); 6364 6365 mgmt_pending_foreach(MGMT_OP_SET_ADVERTISING, hdev, settings_rsp, 6366 &match); 6367 6368 new_settings(hdev, match.sk); 6369 6370 if (match.sk) 6371 sock_put(match.sk); 6372 6373 /* If "Set Advertising" was just disabled and instance advertising was 6374 * set up earlier, then re-enable multi-instance advertising. 6375 */ 6376 if (hci_dev_test_flag(hdev, HCI_ADVERTISING) || 6377 list_empty(&hdev->adv_instances)) 6378 return; 6379 6380 instance = hdev->cur_adv_instance; 6381 if (!instance) { 6382 adv_instance = list_first_entry_or_null(&hdev->adv_instances, 6383 struct adv_info, list); 6384 if (!adv_instance) 6385 return; 6386 6387 instance = adv_instance->instance; 6388 } 6389 6390 err = hci_schedule_adv_instance_sync(hdev, instance, true); 6391 6392 enable_advertising_instance(hdev, err); 6393 } 6394 6395 static int set_adv_sync(struct hci_dev *hdev, void *data) 6396 { 6397 struct mgmt_pending_cmd *cmd = data; 6398 struct mgmt_mode *cp = cmd->param; 6399 u8 val = !!cp->val; 6400 6401 if (cp->val == 0x02) 6402 hci_dev_set_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6403 else 6404 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6405 6406 cancel_adv_timeout(hdev); 6407 6408 if (val) { 6409 /* Switch to instance "0" for the Set Advertising setting. 6410 * We cannot use update_[adv|scan_rsp]_data() here as the 6411 * HCI_ADVERTISING flag is not yet set. 6412 */ 6413 hdev->cur_adv_instance = 0x00; 6414 6415 if (ext_adv_capable(hdev)) { 6416 hci_start_ext_adv_sync(hdev, 0x00); 6417 } else { 6418 hci_update_adv_data_sync(hdev, 0x00); 6419 hci_update_scan_rsp_data_sync(hdev, 0x00); 6420 hci_enable_advertising_sync(hdev); 6421 } 6422 } else { 6423 hci_disable_advertising_sync(hdev); 6424 } 6425 6426 return 0; 6427 } 6428 6429 static int set_advertising(struct sock *sk, struct hci_dev *hdev, void *data, 6430 u16 len) 6431 { 6432 struct mgmt_mode *cp = data; 6433 struct mgmt_pending_cmd *cmd; 6434 u8 val, status; 6435 int err; 6436 6437 bt_dev_dbg(hdev, "sock %p", sk); 6438 6439 status = mgmt_le_support(hdev); 6440 if (status) 6441 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6442 status); 6443 6444 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 6445 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6446 MGMT_STATUS_INVALID_PARAMS); 6447 6448 if (hdev->advertising_paused) 6449 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6450 MGMT_STATUS_BUSY); 6451 6452 hci_dev_lock(hdev); 6453 6454 val = !!cp->val; 6455 6456 /* The following conditions are ones which mean that we should 6457 * not do any HCI communication but directly send a mgmt 6458 * response to user space (after toggling the flag if 6459 * necessary). 6460 */ 6461 if (!hdev_is_powered(hdev) || 6462 (val == hci_dev_test_flag(hdev, HCI_ADVERTISING) && 6463 (cp->val == 0x02) == hci_dev_test_flag(hdev, HCI_ADVERTISING_CONNECTABLE)) || 6464 hci_dev_test_flag(hdev, HCI_MESH) || 6465 hci_conn_num(hdev, LE_LINK) > 0 || 6466 (hci_dev_test_flag(hdev, HCI_LE_SCAN) && 6467 hdev->le_scan_type == LE_SCAN_ACTIVE)) { 6468 bool changed; 6469 6470 if (cp->val) { 6471 hdev->cur_adv_instance = 0x00; 6472 changed = !hci_dev_test_and_set_flag(hdev, HCI_ADVERTISING); 6473 if (cp->val == 0x02) 6474 hci_dev_set_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6475 else 6476 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6477 } else { 6478 changed = hci_dev_test_and_clear_flag(hdev, HCI_ADVERTISING); 6479 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE); 6480 } 6481 6482 err = send_settings_rsp(sk, MGMT_OP_SET_ADVERTISING, hdev); 6483 if (err < 0) 6484 goto unlock; 6485 6486 if (changed) 6487 err = new_settings(hdev, sk); 6488 6489 goto unlock; 6490 } 6491 6492 if (pending_find(MGMT_OP_SET_ADVERTISING, hdev) || 6493 pending_find(MGMT_OP_SET_LE, hdev)) { 6494 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING, 6495 MGMT_STATUS_BUSY); 6496 goto unlock; 6497 } 6498 6499 cmd = mgmt_pending_add(sk, MGMT_OP_SET_ADVERTISING, hdev, data, len); 6500 if (!cmd) 6501 err = -ENOMEM; 6502 else 6503 err = hci_cmd_sync_queue(hdev, set_adv_sync, cmd, 6504 set_advertising_complete); 6505 6506 if (err < 0 && cmd) 6507 mgmt_pending_remove(cmd); 6508 6509 unlock: 6510 hci_dev_unlock(hdev); 6511 return err; 6512 } 6513 6514 static int set_static_address(struct sock *sk, struct hci_dev *hdev, 6515 void *data, u16 len) 6516 { 6517 struct mgmt_cp_set_static_address *cp = data; 6518 int err; 6519 6520 bt_dev_dbg(hdev, "sock %p", sk); 6521 6522 if (!lmp_le_capable(hdev)) 6523 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_STATIC_ADDRESS, 6524 MGMT_STATUS_NOT_SUPPORTED); 6525 6526 if (hdev_is_powered(hdev)) 6527 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_STATIC_ADDRESS, 6528 MGMT_STATUS_REJECTED); 6529 6530 if (bacmp(&cp->bdaddr, BDADDR_ANY)) { 6531 if (!bacmp(&cp->bdaddr, BDADDR_NONE)) 6532 return mgmt_cmd_status(sk, hdev->id, 6533 MGMT_OP_SET_STATIC_ADDRESS, 6534 MGMT_STATUS_INVALID_PARAMS); 6535 6536 /* Two most significant bits shall be set */ 6537 if ((cp->bdaddr.b[5] & 0xc0) != 0xc0) 6538 return mgmt_cmd_status(sk, hdev->id, 6539 MGMT_OP_SET_STATIC_ADDRESS, 6540 MGMT_STATUS_INVALID_PARAMS); 6541 } 6542 6543 hci_dev_lock(hdev); 6544 6545 bacpy(&hdev->static_addr, &cp->bdaddr); 6546 6547 err = send_settings_rsp(sk, MGMT_OP_SET_STATIC_ADDRESS, hdev); 6548 if (err < 0) 6549 goto unlock; 6550 6551 err = new_settings(hdev, sk); 6552 6553 unlock: 6554 hci_dev_unlock(hdev); 6555 return err; 6556 } 6557 6558 static int set_scan_params(struct sock *sk, struct hci_dev *hdev, 6559 void *data, u16 len) 6560 { 6561 struct mgmt_cp_set_scan_params *cp = data; 6562 __u16 interval, window; 6563 int err; 6564 6565 bt_dev_dbg(hdev, "sock %p", sk); 6566 6567 if (!lmp_le_capable(hdev)) 6568 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6569 MGMT_STATUS_NOT_SUPPORTED); 6570 6571 interval = __le16_to_cpu(cp->interval); 6572 6573 if (interval < 0x0004 || interval > 0x4000) 6574 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6575 MGMT_STATUS_INVALID_PARAMS); 6576 6577 window = __le16_to_cpu(cp->window); 6578 6579 if (window < 0x0004 || window > 0x4000) 6580 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6581 MGMT_STATUS_INVALID_PARAMS); 6582 6583 if (window > interval) 6584 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 6585 MGMT_STATUS_INVALID_PARAMS); 6586 6587 hci_dev_lock(hdev); 6588 6589 hdev->le_scan_interval = interval; 6590 hdev->le_scan_window = window; 6591 6592 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 0, 6593 NULL, 0); 6594 6595 /* If background scan is running, restart it so new parameters are 6596 * loaded. 6597 */ 6598 if (hci_dev_test_flag(hdev, HCI_LE_SCAN) && 6599 hdev->discovery.state == DISCOVERY_STOPPED) 6600 hci_update_passive_scan(hdev); 6601 6602 hci_dev_unlock(hdev); 6603 6604 return err; 6605 } 6606 6607 static void fast_connectable_complete(struct hci_dev *hdev, void *data, int err) 6608 { 6609 struct mgmt_pending_cmd *cmd = data; 6610 6611 bt_dev_dbg(hdev, "err %d", err); 6612 6613 if (err) { 6614 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_FAST_CONNECTABLE, 6615 mgmt_status(err)); 6616 } else { 6617 struct mgmt_mode *cp = cmd->param; 6618 6619 if (cp->val) 6620 hci_dev_set_flag(hdev, HCI_FAST_CONNECTABLE); 6621 else 6622 hci_dev_clear_flag(hdev, HCI_FAST_CONNECTABLE); 6623 6624 send_settings_rsp(cmd->sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev); 6625 new_settings(hdev, cmd->sk); 6626 } 6627 6628 mgmt_pending_free(cmd); 6629 } 6630 6631 static int write_fast_connectable_sync(struct hci_dev *hdev, void *data) 6632 { 6633 struct mgmt_pending_cmd *cmd = data; 6634 struct mgmt_mode *cp = cmd->param; 6635 6636 return hci_write_fast_connectable_sync(hdev, cp->val); 6637 } 6638 6639 static int set_fast_connectable(struct sock *sk, struct hci_dev *hdev, 6640 void *data, u16 len) 6641 { 6642 struct mgmt_mode *cp = data; 6643 struct mgmt_pending_cmd *cmd; 6644 int err; 6645 6646 bt_dev_dbg(hdev, "sock %p", sk); 6647 6648 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) || 6649 hdev->hci_ver < BLUETOOTH_VER_1_2) 6650 return mgmt_cmd_status(sk, hdev->id, 6651 MGMT_OP_SET_FAST_CONNECTABLE, 6652 MGMT_STATUS_NOT_SUPPORTED); 6653 6654 if (cp->val != 0x00 && cp->val != 0x01) 6655 return mgmt_cmd_status(sk, hdev->id, 6656 MGMT_OP_SET_FAST_CONNECTABLE, 6657 MGMT_STATUS_INVALID_PARAMS); 6658 6659 hci_dev_lock(hdev); 6660 6661 if (!!cp->val == hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE)) { 6662 err = send_settings_rsp(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev); 6663 goto unlock; 6664 } 6665 6666 if (!hdev_is_powered(hdev)) { 6667 hci_dev_change_flag(hdev, HCI_FAST_CONNECTABLE); 6668 err = send_settings_rsp(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev); 6669 new_settings(hdev, sk); 6670 goto unlock; 6671 } 6672 6673 cmd = mgmt_pending_new(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev, data, 6674 len); 6675 if (!cmd) 6676 err = -ENOMEM; 6677 else 6678 err = hci_cmd_sync_queue(hdev, write_fast_connectable_sync, cmd, 6679 fast_connectable_complete); 6680 6681 if (err < 0) { 6682 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_FAST_CONNECTABLE, 6683 MGMT_STATUS_FAILED); 6684 6685 if (cmd) 6686 mgmt_pending_free(cmd); 6687 } 6688 6689 unlock: 6690 hci_dev_unlock(hdev); 6691 6692 return err; 6693 } 6694 6695 static void set_bredr_complete(struct hci_dev *hdev, void *data, int err) 6696 { 6697 struct mgmt_pending_cmd *cmd = data; 6698 6699 bt_dev_dbg(hdev, "err %d", err); 6700 6701 if (err) { 6702 u8 mgmt_err = mgmt_status(err); 6703 6704 /* We need to restore the flag if related HCI commands 6705 * failed. 6706 */ 6707 hci_dev_clear_flag(hdev, HCI_BREDR_ENABLED); 6708 6709 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 6710 } else { 6711 send_settings_rsp(cmd->sk, MGMT_OP_SET_BREDR, hdev); 6712 new_settings(hdev, cmd->sk); 6713 } 6714 6715 mgmt_pending_free(cmd); 6716 } 6717 6718 static int set_bredr_sync(struct hci_dev *hdev, void *data) 6719 { 6720 int status; 6721 6722 status = hci_write_fast_connectable_sync(hdev, false); 6723 6724 if (!status) 6725 status = hci_update_scan_sync(hdev); 6726 6727 /* Since only the advertising data flags will change, there 6728 * is no need to update the scan response data. 6729 */ 6730 if (!status) 6731 status = hci_update_adv_data_sync(hdev, hdev->cur_adv_instance); 6732 6733 return status; 6734 } 6735 6736 static int set_bredr(struct sock *sk, struct hci_dev *hdev, void *data, u16 len) 6737 { 6738 struct mgmt_mode *cp = data; 6739 struct mgmt_pending_cmd *cmd; 6740 int err; 6741 6742 bt_dev_dbg(hdev, "sock %p", sk); 6743 6744 if (!lmp_bredr_capable(hdev) || !lmp_le_capable(hdev)) 6745 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6746 MGMT_STATUS_NOT_SUPPORTED); 6747 6748 if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 6749 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6750 MGMT_STATUS_REJECTED); 6751 6752 if (cp->val != 0x00 && cp->val != 0x01) 6753 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6754 MGMT_STATUS_INVALID_PARAMS); 6755 6756 hci_dev_lock(hdev); 6757 6758 if (cp->val == hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) { 6759 err = send_settings_rsp(sk, MGMT_OP_SET_BREDR, hdev); 6760 goto unlock; 6761 } 6762 6763 if (!hdev_is_powered(hdev)) { 6764 if (!cp->val) { 6765 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE); 6766 hci_dev_clear_flag(hdev, HCI_SSP_ENABLED); 6767 hci_dev_clear_flag(hdev, HCI_LINK_SECURITY); 6768 hci_dev_clear_flag(hdev, HCI_FAST_CONNECTABLE); 6769 hci_dev_clear_flag(hdev, HCI_HS_ENABLED); 6770 } 6771 6772 hci_dev_change_flag(hdev, HCI_BREDR_ENABLED); 6773 6774 err = send_settings_rsp(sk, MGMT_OP_SET_BREDR, hdev); 6775 if (err < 0) 6776 goto unlock; 6777 6778 err = new_settings(hdev, sk); 6779 goto unlock; 6780 } 6781 6782 /* Reject disabling when powered on */ 6783 if (!cp->val) { 6784 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6785 MGMT_STATUS_REJECTED); 6786 goto unlock; 6787 } else { 6788 /* When configuring a dual-mode controller to operate 6789 * with LE only and using a static address, then switching 6790 * BR/EDR back on is not allowed. 6791 * 6792 * Dual-mode controllers shall operate with the public 6793 * address as its identity address for BR/EDR and LE. So 6794 * reject the attempt to create an invalid configuration. 6795 * 6796 * The same restrictions applies when secure connections 6797 * has been enabled. For BR/EDR this is a controller feature 6798 * while for LE it is a host stack feature. This means that 6799 * switching BR/EDR back on when secure connections has been 6800 * enabled is not a supported transaction. 6801 */ 6802 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) && 6803 (bacmp(&hdev->static_addr, BDADDR_ANY) || 6804 hci_dev_test_flag(hdev, HCI_SC_ENABLED))) { 6805 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6806 MGMT_STATUS_REJECTED); 6807 goto unlock; 6808 } 6809 } 6810 6811 cmd = mgmt_pending_new(sk, MGMT_OP_SET_BREDR, hdev, data, len); 6812 if (!cmd) 6813 err = -ENOMEM; 6814 else 6815 err = hci_cmd_sync_queue(hdev, set_bredr_sync, cmd, 6816 set_bredr_complete); 6817 6818 if (err < 0) { 6819 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR, 6820 MGMT_STATUS_FAILED); 6821 if (cmd) 6822 mgmt_pending_free(cmd); 6823 6824 goto unlock; 6825 } 6826 6827 /* We need to flip the bit already here so that 6828 * hci_req_update_adv_data generates the correct flags. 6829 */ 6830 hci_dev_set_flag(hdev, HCI_BREDR_ENABLED); 6831 6832 unlock: 6833 hci_dev_unlock(hdev); 6834 return err; 6835 } 6836 6837 static void set_secure_conn_complete(struct hci_dev *hdev, void *data, int err) 6838 { 6839 struct mgmt_pending_cmd *cmd = data; 6840 struct mgmt_mode *cp; 6841 6842 bt_dev_dbg(hdev, "err %d", err); 6843 6844 if (err) { 6845 u8 mgmt_err = mgmt_status(err); 6846 6847 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err); 6848 goto done; 6849 } 6850 6851 cp = cmd->param; 6852 6853 switch (cp->val) { 6854 case 0x00: 6855 hci_dev_clear_flag(hdev, HCI_SC_ENABLED); 6856 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6857 break; 6858 case 0x01: 6859 hci_dev_set_flag(hdev, HCI_SC_ENABLED); 6860 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6861 break; 6862 case 0x02: 6863 hci_dev_set_flag(hdev, HCI_SC_ENABLED); 6864 hci_dev_set_flag(hdev, HCI_SC_ONLY); 6865 break; 6866 } 6867 6868 send_settings_rsp(cmd->sk, cmd->opcode, hdev); 6869 new_settings(hdev, cmd->sk); 6870 6871 done: 6872 mgmt_pending_free(cmd); 6873 } 6874 6875 static int set_secure_conn_sync(struct hci_dev *hdev, void *data) 6876 { 6877 struct mgmt_pending_cmd *cmd = data; 6878 struct mgmt_mode *cp = cmd->param; 6879 u8 val = !!cp->val; 6880 6881 /* Force write of val */ 6882 hci_dev_set_flag(hdev, HCI_SC_ENABLED); 6883 6884 return hci_write_sc_support_sync(hdev, val); 6885 } 6886 6887 static int set_secure_conn(struct sock *sk, struct hci_dev *hdev, 6888 void *data, u16 len) 6889 { 6890 struct mgmt_mode *cp = data; 6891 struct mgmt_pending_cmd *cmd; 6892 u8 val; 6893 int err; 6894 6895 bt_dev_dbg(hdev, "sock %p", sk); 6896 6897 if (!lmp_sc_capable(hdev) && 6898 !hci_dev_test_flag(hdev, HCI_LE_ENABLED)) 6899 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6900 MGMT_STATUS_NOT_SUPPORTED); 6901 6902 if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) && 6903 lmp_sc_capable(hdev) && 6904 !hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) 6905 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6906 MGMT_STATUS_REJECTED); 6907 6908 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 6909 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6910 MGMT_STATUS_INVALID_PARAMS); 6911 6912 hci_dev_lock(hdev); 6913 6914 if (!hdev_is_powered(hdev) || !lmp_sc_capable(hdev) || 6915 !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) { 6916 bool changed; 6917 6918 if (cp->val) { 6919 changed = !hci_dev_test_and_set_flag(hdev, 6920 HCI_SC_ENABLED); 6921 if (cp->val == 0x02) 6922 hci_dev_set_flag(hdev, HCI_SC_ONLY); 6923 else 6924 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6925 } else { 6926 changed = hci_dev_test_and_clear_flag(hdev, 6927 HCI_SC_ENABLED); 6928 hci_dev_clear_flag(hdev, HCI_SC_ONLY); 6929 } 6930 6931 err = send_settings_rsp(sk, MGMT_OP_SET_SECURE_CONN, hdev); 6932 if (err < 0) 6933 goto failed; 6934 6935 if (changed) 6936 err = new_settings(hdev, sk); 6937 6938 goto failed; 6939 } 6940 6941 val = !!cp->val; 6942 6943 if (val == hci_dev_test_flag(hdev, HCI_SC_ENABLED) && 6944 (cp->val == 0x02) == hci_dev_test_flag(hdev, HCI_SC_ONLY)) { 6945 err = send_settings_rsp(sk, MGMT_OP_SET_SECURE_CONN, hdev); 6946 goto failed; 6947 } 6948 6949 cmd = mgmt_pending_new(sk, MGMT_OP_SET_SECURE_CONN, hdev, data, len); 6950 if (!cmd) 6951 err = -ENOMEM; 6952 else 6953 err = hci_cmd_sync_queue(hdev, set_secure_conn_sync, cmd, 6954 set_secure_conn_complete); 6955 6956 if (err < 0) { 6957 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN, 6958 MGMT_STATUS_FAILED); 6959 if (cmd) 6960 mgmt_pending_free(cmd); 6961 } 6962 6963 failed: 6964 hci_dev_unlock(hdev); 6965 return err; 6966 } 6967 6968 static int set_debug_keys(struct sock *sk, struct hci_dev *hdev, 6969 void *data, u16 len) 6970 { 6971 struct mgmt_mode *cp = data; 6972 bool changed, use_changed; 6973 int err; 6974 6975 bt_dev_dbg(hdev, "sock %p", sk); 6976 6977 if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02) 6978 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEBUG_KEYS, 6979 MGMT_STATUS_INVALID_PARAMS); 6980 6981 hci_dev_lock(hdev); 6982 6983 if (cp->val) 6984 changed = !hci_dev_test_and_set_flag(hdev, HCI_KEEP_DEBUG_KEYS); 6985 else 6986 changed = hci_dev_test_and_clear_flag(hdev, 6987 HCI_KEEP_DEBUG_KEYS); 6988 6989 if (cp->val == 0x02) 6990 use_changed = !hci_dev_test_and_set_flag(hdev, 6991 HCI_USE_DEBUG_KEYS); 6992 else 6993 use_changed = hci_dev_test_and_clear_flag(hdev, 6994 HCI_USE_DEBUG_KEYS); 6995 6996 if (hdev_is_powered(hdev) && use_changed && 6997 hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) { 6998 u8 mode = (cp->val == 0x02) ? 0x01 : 0x00; 6999 hci_send_cmd(hdev, HCI_OP_WRITE_SSP_DEBUG_MODE, 7000 sizeof(mode), &mode); 7001 } 7002 7003 err = send_settings_rsp(sk, MGMT_OP_SET_DEBUG_KEYS, hdev); 7004 if (err < 0) 7005 goto unlock; 7006 7007 if (changed) 7008 err = new_settings(hdev, sk); 7009 7010 unlock: 7011 hci_dev_unlock(hdev); 7012 return err; 7013 } 7014 7015 static int set_privacy(struct sock *sk, struct hci_dev *hdev, void *cp_data, 7016 u16 len) 7017 { 7018 struct mgmt_cp_set_privacy *cp = cp_data; 7019 bool changed; 7020 int err; 7021 7022 bt_dev_dbg(hdev, "sock %p", sk); 7023 7024 if (!lmp_le_capable(hdev)) 7025 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY, 7026 MGMT_STATUS_NOT_SUPPORTED); 7027 7028 if (cp->privacy != 0x00 && cp->privacy != 0x01 && cp->privacy != 0x02) 7029 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY, 7030 MGMT_STATUS_INVALID_PARAMS); 7031 7032 if (hdev_is_powered(hdev)) 7033 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY, 7034 MGMT_STATUS_REJECTED); 7035 7036 hci_dev_lock(hdev); 7037 7038 /* If user space supports this command it is also expected to 7039 * handle IRKs. Therefore, set the HCI_RPA_RESOLVING flag. 7040 */ 7041 hci_dev_set_flag(hdev, HCI_RPA_RESOLVING); 7042 7043 if (cp->privacy) { 7044 changed = !hci_dev_test_and_set_flag(hdev, HCI_PRIVACY); 7045 memcpy(hdev->irk, cp->irk, sizeof(hdev->irk)); 7046 hci_dev_set_flag(hdev, HCI_RPA_EXPIRED); 7047 hci_adv_instances_set_rpa_expired(hdev, true); 7048 if (cp->privacy == 0x02) 7049 hci_dev_set_flag(hdev, HCI_LIMITED_PRIVACY); 7050 else 7051 hci_dev_clear_flag(hdev, HCI_LIMITED_PRIVACY); 7052 } else { 7053 changed = hci_dev_test_and_clear_flag(hdev, HCI_PRIVACY); 7054 memset(hdev->irk, 0, sizeof(hdev->irk)); 7055 hci_dev_clear_flag(hdev, HCI_RPA_EXPIRED); 7056 hci_adv_instances_set_rpa_expired(hdev, false); 7057 hci_dev_clear_flag(hdev, HCI_LIMITED_PRIVACY); 7058 } 7059 7060 err = send_settings_rsp(sk, MGMT_OP_SET_PRIVACY, hdev); 7061 if (err < 0) 7062 goto unlock; 7063 7064 if (changed) 7065 err = new_settings(hdev, sk); 7066 7067 unlock: 7068 hci_dev_unlock(hdev); 7069 return err; 7070 } 7071 7072 static bool irk_is_valid(struct mgmt_irk_info *irk) 7073 { 7074 switch (irk->addr.type) { 7075 case BDADDR_LE_PUBLIC: 7076 return true; 7077 7078 case BDADDR_LE_RANDOM: 7079 /* Two most significant bits shall be set */ 7080 if ((irk->addr.bdaddr.b[5] & 0xc0) != 0xc0) 7081 return false; 7082 return true; 7083 } 7084 7085 return false; 7086 } 7087 7088 static int load_irks(struct sock *sk, struct hci_dev *hdev, void *cp_data, 7089 u16 len) 7090 { 7091 struct mgmt_cp_load_irks *cp = cp_data; 7092 const u16 max_irk_count = ((U16_MAX - sizeof(*cp)) / 7093 sizeof(struct mgmt_irk_info)); 7094 u16 irk_count, expected_len; 7095 int i, err; 7096 7097 bt_dev_dbg(hdev, "sock %p", sk); 7098 7099 if (!lmp_le_capable(hdev)) 7100 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS, 7101 MGMT_STATUS_NOT_SUPPORTED); 7102 7103 irk_count = __le16_to_cpu(cp->irk_count); 7104 if (irk_count > max_irk_count) { 7105 bt_dev_err(hdev, "load_irks: too big irk_count value %u", 7106 irk_count); 7107 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS, 7108 MGMT_STATUS_INVALID_PARAMS); 7109 } 7110 7111 expected_len = struct_size(cp, irks, irk_count); 7112 if (expected_len != len) { 7113 bt_dev_err(hdev, "load_irks: expected %u bytes, got %u bytes", 7114 expected_len, len); 7115 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS, 7116 MGMT_STATUS_INVALID_PARAMS); 7117 } 7118 7119 bt_dev_dbg(hdev, "irk_count %u", irk_count); 7120 7121 for (i = 0; i < irk_count; i++) { 7122 struct mgmt_irk_info *key = &cp->irks[i]; 7123 7124 if (!irk_is_valid(key)) 7125 return mgmt_cmd_status(sk, hdev->id, 7126 MGMT_OP_LOAD_IRKS, 7127 MGMT_STATUS_INVALID_PARAMS); 7128 } 7129 7130 hci_dev_lock(hdev); 7131 7132 hci_smp_irks_clear(hdev); 7133 7134 for (i = 0; i < irk_count; i++) { 7135 struct mgmt_irk_info *irk = &cp->irks[i]; 7136 u8 addr_type = le_addr_type(irk->addr.type); 7137 7138 if (hci_is_blocked_key(hdev, 7139 HCI_BLOCKED_KEY_TYPE_IRK, 7140 irk->val)) { 7141 bt_dev_warn(hdev, "Skipping blocked IRK for %pMR", 7142 &irk->addr.bdaddr); 7143 continue; 7144 } 7145 7146 /* When using SMP over BR/EDR, the addr type should be set to BREDR */ 7147 if (irk->addr.type == BDADDR_BREDR) 7148 addr_type = BDADDR_BREDR; 7149 7150 hci_add_irk(hdev, &irk->addr.bdaddr, 7151 addr_type, irk->val, 7152 BDADDR_ANY); 7153 } 7154 7155 hci_dev_set_flag(hdev, HCI_RPA_RESOLVING); 7156 7157 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_IRKS, 0, NULL, 0); 7158 7159 hci_dev_unlock(hdev); 7160 7161 return err; 7162 } 7163 7164 static bool ltk_is_valid(struct mgmt_ltk_info *key) 7165 { 7166 if (key->initiator != 0x00 && key->initiator != 0x01) 7167 return false; 7168 7169 switch (key->addr.type) { 7170 case BDADDR_LE_PUBLIC: 7171 return true; 7172 7173 case BDADDR_LE_RANDOM: 7174 /* Two most significant bits shall be set */ 7175 if ((key->addr.bdaddr.b[5] & 0xc0) != 0xc0) 7176 return false; 7177 return true; 7178 } 7179 7180 return false; 7181 } 7182 7183 static int load_long_term_keys(struct sock *sk, struct hci_dev *hdev, 7184 void *cp_data, u16 len) 7185 { 7186 struct mgmt_cp_load_long_term_keys *cp = cp_data; 7187 const u16 max_key_count = ((U16_MAX - sizeof(*cp)) / 7188 sizeof(struct mgmt_ltk_info)); 7189 u16 key_count, expected_len; 7190 int i, err; 7191 7192 bt_dev_dbg(hdev, "sock %p", sk); 7193 7194 if (!lmp_le_capable(hdev)) 7195 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 7196 MGMT_STATUS_NOT_SUPPORTED); 7197 7198 key_count = __le16_to_cpu(cp->key_count); 7199 if (key_count > max_key_count) { 7200 bt_dev_err(hdev, "load_ltks: too big key_count value %u", 7201 key_count); 7202 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 7203 MGMT_STATUS_INVALID_PARAMS); 7204 } 7205 7206 expected_len = struct_size(cp, keys, key_count); 7207 if (expected_len != len) { 7208 bt_dev_err(hdev, "load_keys: expected %u bytes, got %u bytes", 7209 expected_len, len); 7210 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 7211 MGMT_STATUS_INVALID_PARAMS); 7212 } 7213 7214 bt_dev_dbg(hdev, "key_count %u", key_count); 7215 7216 for (i = 0; i < key_count; i++) { 7217 struct mgmt_ltk_info *key = &cp->keys[i]; 7218 7219 if (!ltk_is_valid(key)) 7220 return mgmt_cmd_status(sk, hdev->id, 7221 MGMT_OP_LOAD_LONG_TERM_KEYS, 7222 MGMT_STATUS_INVALID_PARAMS); 7223 } 7224 7225 hci_dev_lock(hdev); 7226 7227 hci_smp_ltks_clear(hdev); 7228 7229 for (i = 0; i < key_count; i++) { 7230 struct mgmt_ltk_info *key = &cp->keys[i]; 7231 u8 type, authenticated; 7232 u8 addr_type = le_addr_type(key->addr.type); 7233 7234 if (hci_is_blocked_key(hdev, 7235 HCI_BLOCKED_KEY_TYPE_LTK, 7236 key->val)) { 7237 bt_dev_warn(hdev, "Skipping blocked LTK for %pMR", 7238 &key->addr.bdaddr); 7239 continue; 7240 } 7241 7242 switch (key->type) { 7243 case MGMT_LTK_UNAUTHENTICATED: 7244 authenticated = 0x00; 7245 type = key->initiator ? SMP_LTK : SMP_LTK_RESPONDER; 7246 break; 7247 case MGMT_LTK_AUTHENTICATED: 7248 authenticated = 0x01; 7249 type = key->initiator ? SMP_LTK : SMP_LTK_RESPONDER; 7250 break; 7251 case MGMT_LTK_P256_UNAUTH: 7252 authenticated = 0x00; 7253 type = SMP_LTK_P256; 7254 break; 7255 case MGMT_LTK_P256_AUTH: 7256 authenticated = 0x01; 7257 type = SMP_LTK_P256; 7258 break; 7259 case MGMT_LTK_P256_DEBUG: 7260 authenticated = 0x00; 7261 type = SMP_LTK_P256_DEBUG; 7262 fallthrough; 7263 default: 7264 continue; 7265 } 7266 7267 /* When using SMP over BR/EDR, the addr type should be set to BREDR */ 7268 if (key->addr.type == BDADDR_BREDR) 7269 addr_type = BDADDR_BREDR; 7270 7271 hci_add_ltk(hdev, &key->addr.bdaddr, 7272 addr_type, type, authenticated, 7273 key->val, key->enc_size, key->ediv, key->rand); 7274 } 7275 7276 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 0, 7277 NULL, 0); 7278 7279 hci_dev_unlock(hdev); 7280 7281 return err; 7282 } 7283 7284 static void get_conn_info_complete(struct hci_dev *hdev, void *data, int err) 7285 { 7286 struct mgmt_pending_cmd *cmd = data; 7287 struct hci_conn *conn = cmd->user_data; 7288 struct mgmt_cp_get_conn_info *cp = cmd->param; 7289 struct mgmt_rp_get_conn_info rp; 7290 u8 status; 7291 7292 bt_dev_dbg(hdev, "err %d", err); 7293 7294 memcpy(&rp.addr, &cp->addr, sizeof(rp.addr)); 7295 7296 status = mgmt_status(err); 7297 if (status == MGMT_STATUS_SUCCESS) { 7298 rp.rssi = conn->rssi; 7299 rp.tx_power = conn->tx_power; 7300 rp.max_tx_power = conn->max_tx_power; 7301 } else { 7302 rp.rssi = HCI_RSSI_INVALID; 7303 rp.tx_power = HCI_TX_POWER_INVALID; 7304 rp.max_tx_power = HCI_TX_POWER_INVALID; 7305 } 7306 7307 mgmt_cmd_complete(cmd->sk, cmd->index, MGMT_OP_GET_CONN_INFO, status, 7308 &rp, sizeof(rp)); 7309 7310 mgmt_pending_free(cmd); 7311 } 7312 7313 static int get_conn_info_sync(struct hci_dev *hdev, void *data) 7314 { 7315 struct mgmt_pending_cmd *cmd = data; 7316 struct mgmt_cp_get_conn_info *cp = cmd->param; 7317 struct hci_conn *conn; 7318 int err; 7319 __le16 handle; 7320 7321 /* Make sure we are still connected */ 7322 if (cp->addr.type == BDADDR_BREDR) 7323 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 7324 &cp->addr.bdaddr); 7325 else 7326 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr); 7327 7328 if (!conn || conn->state != BT_CONNECTED) 7329 return MGMT_STATUS_NOT_CONNECTED; 7330 7331 cmd->user_data = conn; 7332 handle = cpu_to_le16(conn->handle); 7333 7334 /* Refresh RSSI each time */ 7335 err = hci_read_rssi_sync(hdev, handle); 7336 7337 /* For LE links TX power does not change thus we don't need to 7338 * query for it once value is known. 7339 */ 7340 if (!err && (!bdaddr_type_is_le(cp->addr.type) || 7341 conn->tx_power == HCI_TX_POWER_INVALID)) 7342 err = hci_read_tx_power_sync(hdev, handle, 0x00); 7343 7344 /* Max TX power needs to be read only once per connection */ 7345 if (!err && conn->max_tx_power == HCI_TX_POWER_INVALID) 7346 err = hci_read_tx_power_sync(hdev, handle, 0x01); 7347 7348 return err; 7349 } 7350 7351 static int get_conn_info(struct sock *sk, struct hci_dev *hdev, void *data, 7352 u16 len) 7353 { 7354 struct mgmt_cp_get_conn_info *cp = data; 7355 struct mgmt_rp_get_conn_info rp; 7356 struct hci_conn *conn; 7357 unsigned long conn_info_age; 7358 int err = 0; 7359 7360 bt_dev_dbg(hdev, "sock %p", sk); 7361 7362 memset(&rp, 0, sizeof(rp)); 7363 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 7364 rp.addr.type = cp->addr.type; 7365 7366 if (!bdaddr_type_is_valid(cp->addr.type)) 7367 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7368 MGMT_STATUS_INVALID_PARAMS, 7369 &rp, sizeof(rp)); 7370 7371 hci_dev_lock(hdev); 7372 7373 if (!hdev_is_powered(hdev)) { 7374 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7375 MGMT_STATUS_NOT_POWERED, &rp, 7376 sizeof(rp)); 7377 goto unlock; 7378 } 7379 7380 if (cp->addr.type == BDADDR_BREDR) 7381 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 7382 &cp->addr.bdaddr); 7383 else 7384 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr); 7385 7386 if (!conn || conn->state != BT_CONNECTED) { 7387 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7388 MGMT_STATUS_NOT_CONNECTED, &rp, 7389 sizeof(rp)); 7390 goto unlock; 7391 } 7392 7393 /* To avoid client trying to guess when to poll again for information we 7394 * calculate conn info age as random value between min/max set in hdev. 7395 */ 7396 conn_info_age = get_random_u32_inclusive(hdev->conn_info_min_age, 7397 hdev->conn_info_max_age - 1); 7398 7399 /* Query controller to refresh cached values if they are too old or were 7400 * never read. 7401 */ 7402 if (time_after(jiffies, conn->conn_info_timestamp + 7403 msecs_to_jiffies(conn_info_age)) || 7404 !conn->conn_info_timestamp) { 7405 struct mgmt_pending_cmd *cmd; 7406 7407 cmd = mgmt_pending_new(sk, MGMT_OP_GET_CONN_INFO, hdev, data, 7408 len); 7409 if (!cmd) { 7410 err = -ENOMEM; 7411 } else { 7412 err = hci_cmd_sync_queue(hdev, get_conn_info_sync, 7413 cmd, get_conn_info_complete); 7414 } 7415 7416 if (err < 0) { 7417 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7418 MGMT_STATUS_FAILED, &rp, sizeof(rp)); 7419 7420 if (cmd) 7421 mgmt_pending_free(cmd); 7422 7423 goto unlock; 7424 } 7425 7426 conn->conn_info_timestamp = jiffies; 7427 } else { 7428 /* Cache is valid, just reply with values cached in hci_conn */ 7429 rp.rssi = conn->rssi; 7430 rp.tx_power = conn->tx_power; 7431 rp.max_tx_power = conn->max_tx_power; 7432 7433 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO, 7434 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 7435 } 7436 7437 unlock: 7438 hci_dev_unlock(hdev); 7439 return err; 7440 } 7441 7442 static void get_clock_info_complete(struct hci_dev *hdev, void *data, int err) 7443 { 7444 struct mgmt_pending_cmd *cmd = data; 7445 struct mgmt_cp_get_clock_info *cp = cmd->param; 7446 struct mgmt_rp_get_clock_info rp; 7447 struct hci_conn *conn = cmd->user_data; 7448 u8 status = mgmt_status(err); 7449 7450 bt_dev_dbg(hdev, "err %d", err); 7451 7452 memset(&rp, 0, sizeof(rp)); 7453 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 7454 rp.addr.type = cp->addr.type; 7455 7456 if (err) 7457 goto complete; 7458 7459 rp.local_clock = cpu_to_le32(hdev->clock); 7460 7461 if (conn) { 7462 rp.piconet_clock = cpu_to_le32(conn->clock); 7463 rp.accuracy = cpu_to_le16(conn->clock_accuracy); 7464 } 7465 7466 complete: 7467 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status, &rp, 7468 sizeof(rp)); 7469 7470 mgmt_pending_free(cmd); 7471 } 7472 7473 static int get_clock_info_sync(struct hci_dev *hdev, void *data) 7474 { 7475 struct mgmt_pending_cmd *cmd = data; 7476 struct mgmt_cp_get_clock_info *cp = cmd->param; 7477 struct hci_cp_read_clock hci_cp; 7478 struct hci_conn *conn; 7479 7480 memset(&hci_cp, 0, sizeof(hci_cp)); 7481 hci_read_clock_sync(hdev, &hci_cp); 7482 7483 /* Make sure connection still exists */ 7484 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr); 7485 if (!conn || conn->state != BT_CONNECTED) 7486 return MGMT_STATUS_NOT_CONNECTED; 7487 7488 cmd->user_data = conn; 7489 hci_cp.handle = cpu_to_le16(conn->handle); 7490 hci_cp.which = 0x01; /* Piconet clock */ 7491 7492 return hci_read_clock_sync(hdev, &hci_cp); 7493 } 7494 7495 static int get_clock_info(struct sock *sk, struct hci_dev *hdev, void *data, 7496 u16 len) 7497 { 7498 struct mgmt_cp_get_clock_info *cp = data; 7499 struct mgmt_rp_get_clock_info rp; 7500 struct mgmt_pending_cmd *cmd; 7501 struct hci_conn *conn; 7502 int err; 7503 7504 bt_dev_dbg(hdev, "sock %p", sk); 7505 7506 memset(&rp, 0, sizeof(rp)); 7507 bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr); 7508 rp.addr.type = cp->addr.type; 7509 7510 if (cp->addr.type != BDADDR_BREDR) 7511 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO, 7512 MGMT_STATUS_INVALID_PARAMS, 7513 &rp, sizeof(rp)); 7514 7515 hci_dev_lock(hdev); 7516 7517 if (!hdev_is_powered(hdev)) { 7518 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO, 7519 MGMT_STATUS_NOT_POWERED, &rp, 7520 sizeof(rp)); 7521 goto unlock; 7522 } 7523 7524 if (bacmp(&cp->addr.bdaddr, BDADDR_ANY)) { 7525 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, 7526 &cp->addr.bdaddr); 7527 if (!conn || conn->state != BT_CONNECTED) { 7528 err = mgmt_cmd_complete(sk, hdev->id, 7529 MGMT_OP_GET_CLOCK_INFO, 7530 MGMT_STATUS_NOT_CONNECTED, 7531 &rp, sizeof(rp)); 7532 goto unlock; 7533 } 7534 } else { 7535 conn = NULL; 7536 } 7537 7538 cmd = mgmt_pending_new(sk, MGMT_OP_GET_CLOCK_INFO, hdev, data, len); 7539 if (!cmd) 7540 err = -ENOMEM; 7541 else 7542 err = hci_cmd_sync_queue(hdev, get_clock_info_sync, cmd, 7543 get_clock_info_complete); 7544 7545 if (err < 0) { 7546 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO, 7547 MGMT_STATUS_FAILED, &rp, sizeof(rp)); 7548 7549 if (cmd) 7550 mgmt_pending_free(cmd); 7551 } 7552 7553 7554 unlock: 7555 hci_dev_unlock(hdev); 7556 return err; 7557 } 7558 7559 static bool is_connected(struct hci_dev *hdev, bdaddr_t *addr, u8 type) 7560 { 7561 struct hci_conn *conn; 7562 7563 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, addr); 7564 if (!conn) 7565 return false; 7566 7567 if (conn->dst_type != type) 7568 return false; 7569 7570 if (conn->state != BT_CONNECTED) 7571 return false; 7572 7573 return true; 7574 } 7575 7576 /* This function requires the caller holds hdev->lock */ 7577 static int hci_conn_params_set(struct hci_dev *hdev, bdaddr_t *addr, 7578 u8 addr_type, u8 auto_connect) 7579 { 7580 struct hci_conn_params *params; 7581 7582 params = hci_conn_params_add(hdev, addr, addr_type); 7583 if (!params) 7584 return -EIO; 7585 7586 if (params->auto_connect == auto_connect) 7587 return 0; 7588 7589 hci_pend_le_list_del_init(params); 7590 7591 switch (auto_connect) { 7592 case HCI_AUTO_CONN_DISABLED: 7593 case HCI_AUTO_CONN_LINK_LOSS: 7594 /* If auto connect is being disabled when we're trying to 7595 * connect to device, keep connecting. 7596 */ 7597 if (params->explicit_connect) 7598 hci_pend_le_list_add(params, &hdev->pend_le_conns); 7599 break; 7600 case HCI_AUTO_CONN_REPORT: 7601 if (params->explicit_connect) 7602 hci_pend_le_list_add(params, &hdev->pend_le_conns); 7603 else 7604 hci_pend_le_list_add(params, &hdev->pend_le_reports); 7605 break; 7606 case HCI_AUTO_CONN_DIRECT: 7607 case HCI_AUTO_CONN_ALWAYS: 7608 if (!is_connected(hdev, addr, addr_type)) 7609 hci_pend_le_list_add(params, &hdev->pend_le_conns); 7610 break; 7611 } 7612 7613 params->auto_connect = auto_connect; 7614 7615 bt_dev_dbg(hdev, "addr %pMR (type %u) auto_connect %u", 7616 addr, addr_type, auto_connect); 7617 7618 return 0; 7619 } 7620 7621 static void device_added(struct sock *sk, struct hci_dev *hdev, 7622 bdaddr_t *bdaddr, u8 type, u8 action) 7623 { 7624 struct mgmt_ev_device_added ev; 7625 7626 bacpy(&ev.addr.bdaddr, bdaddr); 7627 ev.addr.type = type; 7628 ev.action = action; 7629 7630 mgmt_event(MGMT_EV_DEVICE_ADDED, hdev, &ev, sizeof(ev), sk); 7631 } 7632 7633 static int add_device_sync(struct hci_dev *hdev, void *data) 7634 { 7635 return hci_update_passive_scan_sync(hdev); 7636 } 7637 7638 static int add_device(struct sock *sk, struct hci_dev *hdev, 7639 void *data, u16 len) 7640 { 7641 struct mgmt_cp_add_device *cp = data; 7642 u8 auto_conn, addr_type; 7643 struct hci_conn_params *params; 7644 int err; 7645 u32 current_flags = 0; 7646 u32 supported_flags; 7647 7648 bt_dev_dbg(hdev, "sock %p", sk); 7649 7650 if (!bdaddr_type_is_valid(cp->addr.type) || 7651 !bacmp(&cp->addr.bdaddr, BDADDR_ANY)) 7652 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7653 MGMT_STATUS_INVALID_PARAMS, 7654 &cp->addr, sizeof(cp->addr)); 7655 7656 if (cp->action != 0x00 && cp->action != 0x01 && cp->action != 0x02) 7657 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7658 MGMT_STATUS_INVALID_PARAMS, 7659 &cp->addr, sizeof(cp->addr)); 7660 7661 hci_dev_lock(hdev); 7662 7663 if (cp->addr.type == BDADDR_BREDR) { 7664 /* Only incoming connections action is supported for now */ 7665 if (cp->action != 0x01) { 7666 err = mgmt_cmd_complete(sk, hdev->id, 7667 MGMT_OP_ADD_DEVICE, 7668 MGMT_STATUS_INVALID_PARAMS, 7669 &cp->addr, sizeof(cp->addr)); 7670 goto unlock; 7671 } 7672 7673 err = hci_bdaddr_list_add_with_flags(&hdev->accept_list, 7674 &cp->addr.bdaddr, 7675 cp->addr.type, 0); 7676 if (err) 7677 goto unlock; 7678 7679 hci_update_scan(hdev); 7680 7681 goto added; 7682 } 7683 7684 addr_type = le_addr_type(cp->addr.type); 7685 7686 if (cp->action == 0x02) 7687 auto_conn = HCI_AUTO_CONN_ALWAYS; 7688 else if (cp->action == 0x01) 7689 auto_conn = HCI_AUTO_CONN_DIRECT; 7690 else 7691 auto_conn = HCI_AUTO_CONN_REPORT; 7692 7693 /* Kernel internally uses conn_params with resolvable private 7694 * address, but Add Device allows only identity addresses. 7695 * Make sure it is enforced before calling 7696 * hci_conn_params_lookup. 7697 */ 7698 if (!hci_is_identity_address(&cp->addr.bdaddr, addr_type)) { 7699 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7700 MGMT_STATUS_INVALID_PARAMS, 7701 &cp->addr, sizeof(cp->addr)); 7702 goto unlock; 7703 } 7704 7705 /* If the connection parameters don't exist for this device, 7706 * they will be created and configured with defaults. 7707 */ 7708 if (hci_conn_params_set(hdev, &cp->addr.bdaddr, addr_type, 7709 auto_conn) < 0) { 7710 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7711 MGMT_STATUS_FAILED, &cp->addr, 7712 sizeof(cp->addr)); 7713 goto unlock; 7714 } else { 7715 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 7716 addr_type); 7717 if (params) 7718 current_flags = params->flags; 7719 } 7720 7721 err = hci_cmd_sync_queue(hdev, add_device_sync, NULL, NULL); 7722 if (err < 0) 7723 goto unlock; 7724 7725 added: 7726 device_added(sk, hdev, &cp->addr.bdaddr, cp->addr.type, cp->action); 7727 supported_flags = hdev->conn_flags; 7728 device_flags_changed(NULL, hdev, &cp->addr.bdaddr, cp->addr.type, 7729 supported_flags, current_flags); 7730 7731 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE, 7732 MGMT_STATUS_SUCCESS, &cp->addr, 7733 sizeof(cp->addr)); 7734 7735 unlock: 7736 hci_dev_unlock(hdev); 7737 return err; 7738 } 7739 7740 static void device_removed(struct sock *sk, struct hci_dev *hdev, 7741 bdaddr_t *bdaddr, u8 type) 7742 { 7743 struct mgmt_ev_device_removed ev; 7744 7745 bacpy(&ev.addr.bdaddr, bdaddr); 7746 ev.addr.type = type; 7747 7748 mgmt_event(MGMT_EV_DEVICE_REMOVED, hdev, &ev, sizeof(ev), sk); 7749 } 7750 7751 static int remove_device_sync(struct hci_dev *hdev, void *data) 7752 { 7753 return hci_update_passive_scan_sync(hdev); 7754 } 7755 7756 static int remove_device(struct sock *sk, struct hci_dev *hdev, 7757 void *data, u16 len) 7758 { 7759 struct mgmt_cp_remove_device *cp = data; 7760 int err; 7761 7762 bt_dev_dbg(hdev, "sock %p", sk); 7763 7764 hci_dev_lock(hdev); 7765 7766 if (bacmp(&cp->addr.bdaddr, BDADDR_ANY)) { 7767 struct hci_conn_params *params; 7768 u8 addr_type; 7769 7770 if (!bdaddr_type_is_valid(cp->addr.type)) { 7771 err = mgmt_cmd_complete(sk, hdev->id, 7772 MGMT_OP_REMOVE_DEVICE, 7773 MGMT_STATUS_INVALID_PARAMS, 7774 &cp->addr, sizeof(cp->addr)); 7775 goto unlock; 7776 } 7777 7778 if (cp->addr.type == BDADDR_BREDR) { 7779 err = hci_bdaddr_list_del(&hdev->accept_list, 7780 &cp->addr.bdaddr, 7781 cp->addr.type); 7782 if (err) { 7783 err = mgmt_cmd_complete(sk, hdev->id, 7784 MGMT_OP_REMOVE_DEVICE, 7785 MGMT_STATUS_INVALID_PARAMS, 7786 &cp->addr, 7787 sizeof(cp->addr)); 7788 goto unlock; 7789 } 7790 7791 hci_update_scan(hdev); 7792 7793 device_removed(sk, hdev, &cp->addr.bdaddr, 7794 cp->addr.type); 7795 goto complete; 7796 } 7797 7798 addr_type = le_addr_type(cp->addr.type); 7799 7800 /* Kernel internally uses conn_params with resolvable private 7801 * address, but Remove Device allows only identity addresses. 7802 * Make sure it is enforced before calling 7803 * hci_conn_params_lookup. 7804 */ 7805 if (!hci_is_identity_address(&cp->addr.bdaddr, addr_type)) { 7806 err = mgmt_cmd_complete(sk, hdev->id, 7807 MGMT_OP_REMOVE_DEVICE, 7808 MGMT_STATUS_INVALID_PARAMS, 7809 &cp->addr, sizeof(cp->addr)); 7810 goto unlock; 7811 } 7812 7813 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, 7814 addr_type); 7815 if (!params) { 7816 err = mgmt_cmd_complete(sk, hdev->id, 7817 MGMT_OP_REMOVE_DEVICE, 7818 MGMT_STATUS_INVALID_PARAMS, 7819 &cp->addr, sizeof(cp->addr)); 7820 goto unlock; 7821 } 7822 7823 if (params->auto_connect == HCI_AUTO_CONN_DISABLED || 7824 params->auto_connect == HCI_AUTO_CONN_EXPLICIT) { 7825 err = mgmt_cmd_complete(sk, hdev->id, 7826 MGMT_OP_REMOVE_DEVICE, 7827 MGMT_STATUS_INVALID_PARAMS, 7828 &cp->addr, sizeof(cp->addr)); 7829 goto unlock; 7830 } 7831 7832 hci_conn_params_free(params); 7833 7834 device_removed(sk, hdev, &cp->addr.bdaddr, cp->addr.type); 7835 } else { 7836 struct hci_conn_params *p, *tmp; 7837 struct bdaddr_list *b, *btmp; 7838 7839 if (cp->addr.type) { 7840 err = mgmt_cmd_complete(sk, hdev->id, 7841 MGMT_OP_REMOVE_DEVICE, 7842 MGMT_STATUS_INVALID_PARAMS, 7843 &cp->addr, sizeof(cp->addr)); 7844 goto unlock; 7845 } 7846 7847 list_for_each_entry_safe(b, btmp, &hdev->accept_list, list) { 7848 device_removed(sk, hdev, &b->bdaddr, b->bdaddr_type); 7849 list_del(&b->list); 7850 kfree(b); 7851 } 7852 7853 hci_update_scan(hdev); 7854 7855 list_for_each_entry_safe(p, tmp, &hdev->le_conn_params, list) { 7856 if (p->auto_connect == HCI_AUTO_CONN_DISABLED) 7857 continue; 7858 device_removed(sk, hdev, &p->addr, p->addr_type); 7859 if (p->explicit_connect) { 7860 p->auto_connect = HCI_AUTO_CONN_EXPLICIT; 7861 continue; 7862 } 7863 hci_conn_params_free(p); 7864 } 7865 7866 bt_dev_dbg(hdev, "All LE connection parameters were removed"); 7867 } 7868 7869 hci_cmd_sync_queue(hdev, remove_device_sync, NULL, NULL); 7870 7871 complete: 7872 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_REMOVE_DEVICE, 7873 MGMT_STATUS_SUCCESS, &cp->addr, 7874 sizeof(cp->addr)); 7875 unlock: 7876 hci_dev_unlock(hdev); 7877 return err; 7878 } 7879 7880 static int load_conn_param(struct sock *sk, struct hci_dev *hdev, void *data, 7881 u16 len) 7882 { 7883 struct mgmt_cp_load_conn_param *cp = data; 7884 const u16 max_param_count = ((U16_MAX - sizeof(*cp)) / 7885 sizeof(struct mgmt_conn_param)); 7886 u16 param_count, expected_len; 7887 int i; 7888 7889 if (!lmp_le_capable(hdev)) 7890 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 7891 MGMT_STATUS_NOT_SUPPORTED); 7892 7893 param_count = __le16_to_cpu(cp->param_count); 7894 if (param_count > max_param_count) { 7895 bt_dev_err(hdev, "load_conn_param: too big param_count value %u", 7896 param_count); 7897 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 7898 MGMT_STATUS_INVALID_PARAMS); 7899 } 7900 7901 expected_len = struct_size(cp, params, param_count); 7902 if (expected_len != len) { 7903 bt_dev_err(hdev, "load_conn_param: expected %u bytes, got %u bytes", 7904 expected_len, len); 7905 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 7906 MGMT_STATUS_INVALID_PARAMS); 7907 } 7908 7909 bt_dev_dbg(hdev, "param_count %u", param_count); 7910 7911 hci_dev_lock(hdev); 7912 7913 hci_conn_params_clear_disabled(hdev); 7914 7915 for (i = 0; i < param_count; i++) { 7916 struct mgmt_conn_param *param = &cp->params[i]; 7917 struct hci_conn_params *hci_param; 7918 u16 min, max, latency, timeout; 7919 u8 addr_type; 7920 7921 bt_dev_dbg(hdev, "Adding %pMR (type %u)", ¶m->addr.bdaddr, 7922 param->addr.type); 7923 7924 if (param->addr.type == BDADDR_LE_PUBLIC) { 7925 addr_type = ADDR_LE_DEV_PUBLIC; 7926 } else if (param->addr.type == BDADDR_LE_RANDOM) { 7927 addr_type = ADDR_LE_DEV_RANDOM; 7928 } else { 7929 bt_dev_err(hdev, "ignoring invalid connection parameters"); 7930 continue; 7931 } 7932 7933 min = le16_to_cpu(param->min_interval); 7934 max = le16_to_cpu(param->max_interval); 7935 latency = le16_to_cpu(param->latency); 7936 timeout = le16_to_cpu(param->timeout); 7937 7938 bt_dev_dbg(hdev, "min 0x%04x max 0x%04x latency 0x%04x timeout 0x%04x", 7939 min, max, latency, timeout); 7940 7941 if (hci_check_conn_params(min, max, latency, timeout) < 0) { 7942 bt_dev_err(hdev, "ignoring invalid connection parameters"); 7943 continue; 7944 } 7945 7946 hci_param = hci_conn_params_add(hdev, ¶m->addr.bdaddr, 7947 addr_type); 7948 if (!hci_param) { 7949 bt_dev_err(hdev, "failed to add connection parameters"); 7950 continue; 7951 } 7952 7953 hci_param->conn_min_interval = min; 7954 hci_param->conn_max_interval = max; 7955 hci_param->conn_latency = latency; 7956 hci_param->supervision_timeout = timeout; 7957 } 7958 7959 hci_dev_unlock(hdev); 7960 7961 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 0, 7962 NULL, 0); 7963 } 7964 7965 static int set_external_config(struct sock *sk, struct hci_dev *hdev, 7966 void *data, u16 len) 7967 { 7968 struct mgmt_cp_set_external_config *cp = data; 7969 bool changed; 7970 int err; 7971 7972 bt_dev_dbg(hdev, "sock %p", sk); 7973 7974 if (hdev_is_powered(hdev)) 7975 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG, 7976 MGMT_STATUS_REJECTED); 7977 7978 if (cp->config != 0x00 && cp->config != 0x01) 7979 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG, 7980 MGMT_STATUS_INVALID_PARAMS); 7981 7982 if (!test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks)) 7983 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG, 7984 MGMT_STATUS_NOT_SUPPORTED); 7985 7986 hci_dev_lock(hdev); 7987 7988 if (cp->config) 7989 changed = !hci_dev_test_and_set_flag(hdev, HCI_EXT_CONFIGURED); 7990 else 7991 changed = hci_dev_test_and_clear_flag(hdev, HCI_EXT_CONFIGURED); 7992 7993 err = send_options_rsp(sk, MGMT_OP_SET_EXTERNAL_CONFIG, hdev); 7994 if (err < 0) 7995 goto unlock; 7996 7997 if (!changed) 7998 goto unlock; 7999 8000 err = new_options(hdev, sk); 8001 8002 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED) == is_configured(hdev)) { 8003 mgmt_index_removed(hdev); 8004 8005 if (hci_dev_test_and_change_flag(hdev, HCI_UNCONFIGURED)) { 8006 hci_dev_set_flag(hdev, HCI_CONFIG); 8007 hci_dev_set_flag(hdev, HCI_AUTO_OFF); 8008 8009 queue_work(hdev->req_workqueue, &hdev->power_on); 8010 } else { 8011 set_bit(HCI_RAW, &hdev->flags); 8012 mgmt_index_added(hdev); 8013 } 8014 } 8015 8016 unlock: 8017 hci_dev_unlock(hdev); 8018 return err; 8019 } 8020 8021 static int set_public_address(struct sock *sk, struct hci_dev *hdev, 8022 void *data, u16 len) 8023 { 8024 struct mgmt_cp_set_public_address *cp = data; 8025 bool changed; 8026 int err; 8027 8028 bt_dev_dbg(hdev, "sock %p", sk); 8029 8030 if (hdev_is_powered(hdev)) 8031 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS, 8032 MGMT_STATUS_REJECTED); 8033 8034 if (!bacmp(&cp->bdaddr, BDADDR_ANY)) 8035 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS, 8036 MGMT_STATUS_INVALID_PARAMS); 8037 8038 if (!hdev->set_bdaddr) 8039 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS, 8040 MGMT_STATUS_NOT_SUPPORTED); 8041 8042 hci_dev_lock(hdev); 8043 8044 changed = !!bacmp(&hdev->public_addr, &cp->bdaddr); 8045 bacpy(&hdev->public_addr, &cp->bdaddr); 8046 8047 err = send_options_rsp(sk, MGMT_OP_SET_PUBLIC_ADDRESS, hdev); 8048 if (err < 0) 8049 goto unlock; 8050 8051 if (!changed) 8052 goto unlock; 8053 8054 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) 8055 err = new_options(hdev, sk); 8056 8057 if (is_configured(hdev)) { 8058 mgmt_index_removed(hdev); 8059 8060 hci_dev_clear_flag(hdev, HCI_UNCONFIGURED); 8061 8062 hci_dev_set_flag(hdev, HCI_CONFIG); 8063 hci_dev_set_flag(hdev, HCI_AUTO_OFF); 8064 8065 queue_work(hdev->req_workqueue, &hdev->power_on); 8066 } 8067 8068 unlock: 8069 hci_dev_unlock(hdev); 8070 return err; 8071 } 8072 8073 static void read_local_oob_ext_data_complete(struct hci_dev *hdev, void *data, 8074 int err) 8075 { 8076 const struct mgmt_cp_read_local_oob_ext_data *mgmt_cp; 8077 struct mgmt_rp_read_local_oob_ext_data *mgmt_rp; 8078 u8 *h192, *r192, *h256, *r256; 8079 struct mgmt_pending_cmd *cmd = data; 8080 struct sk_buff *skb = cmd->skb; 8081 u8 status = mgmt_status(err); 8082 u16 eir_len; 8083 8084 if (cmd != pending_find(MGMT_OP_READ_LOCAL_OOB_EXT_DATA, hdev)) 8085 return; 8086 8087 if (!status) { 8088 if (!skb) 8089 status = MGMT_STATUS_FAILED; 8090 else if (IS_ERR(skb)) 8091 status = mgmt_status(PTR_ERR(skb)); 8092 else 8093 status = mgmt_status(skb->data[0]); 8094 } 8095 8096 bt_dev_dbg(hdev, "status %u", status); 8097 8098 mgmt_cp = cmd->param; 8099 8100 if (status) { 8101 status = mgmt_status(status); 8102 eir_len = 0; 8103 8104 h192 = NULL; 8105 r192 = NULL; 8106 h256 = NULL; 8107 r256 = NULL; 8108 } else if (!bredr_sc_enabled(hdev)) { 8109 struct hci_rp_read_local_oob_data *rp; 8110 8111 if (skb->len != sizeof(*rp)) { 8112 status = MGMT_STATUS_FAILED; 8113 eir_len = 0; 8114 } else { 8115 status = MGMT_STATUS_SUCCESS; 8116 rp = (void *)skb->data; 8117 8118 eir_len = 5 + 18 + 18; 8119 h192 = rp->hash; 8120 r192 = rp->rand; 8121 h256 = NULL; 8122 r256 = NULL; 8123 } 8124 } else { 8125 struct hci_rp_read_local_oob_ext_data *rp; 8126 8127 if (skb->len != sizeof(*rp)) { 8128 status = MGMT_STATUS_FAILED; 8129 eir_len = 0; 8130 } else { 8131 status = MGMT_STATUS_SUCCESS; 8132 rp = (void *)skb->data; 8133 8134 if (hci_dev_test_flag(hdev, HCI_SC_ONLY)) { 8135 eir_len = 5 + 18 + 18; 8136 h192 = NULL; 8137 r192 = NULL; 8138 } else { 8139 eir_len = 5 + 18 + 18 + 18 + 18; 8140 h192 = rp->hash192; 8141 r192 = rp->rand192; 8142 } 8143 8144 h256 = rp->hash256; 8145 r256 = rp->rand256; 8146 } 8147 } 8148 8149 mgmt_rp = kmalloc(sizeof(*mgmt_rp) + eir_len, GFP_KERNEL); 8150 if (!mgmt_rp) 8151 goto done; 8152 8153 if (eir_len == 0) 8154 goto send_rsp; 8155 8156 eir_len = eir_append_data(mgmt_rp->eir, 0, EIR_CLASS_OF_DEV, 8157 hdev->dev_class, 3); 8158 8159 if (h192 && r192) { 8160 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8161 EIR_SSP_HASH_C192, h192, 16); 8162 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8163 EIR_SSP_RAND_R192, r192, 16); 8164 } 8165 8166 if (h256 && r256) { 8167 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8168 EIR_SSP_HASH_C256, h256, 16); 8169 eir_len = eir_append_data(mgmt_rp->eir, eir_len, 8170 EIR_SSP_RAND_R256, r256, 16); 8171 } 8172 8173 send_rsp: 8174 mgmt_rp->type = mgmt_cp->type; 8175 mgmt_rp->eir_len = cpu_to_le16(eir_len); 8176 8177 err = mgmt_cmd_complete(cmd->sk, hdev->id, 8178 MGMT_OP_READ_LOCAL_OOB_EXT_DATA, status, 8179 mgmt_rp, sizeof(*mgmt_rp) + eir_len); 8180 if (err < 0 || status) 8181 goto done; 8182 8183 hci_sock_set_flag(cmd->sk, HCI_MGMT_OOB_DATA_EVENTS); 8184 8185 err = mgmt_limited_event(MGMT_EV_LOCAL_OOB_DATA_UPDATED, hdev, 8186 mgmt_rp, sizeof(*mgmt_rp) + eir_len, 8187 HCI_MGMT_OOB_DATA_EVENTS, cmd->sk); 8188 done: 8189 if (skb && !IS_ERR(skb)) 8190 kfree_skb(skb); 8191 8192 kfree(mgmt_rp); 8193 mgmt_pending_remove(cmd); 8194 } 8195 8196 static int read_local_ssp_oob_req(struct hci_dev *hdev, struct sock *sk, 8197 struct mgmt_cp_read_local_oob_ext_data *cp) 8198 { 8199 struct mgmt_pending_cmd *cmd; 8200 int err; 8201 8202 cmd = mgmt_pending_add(sk, MGMT_OP_READ_LOCAL_OOB_EXT_DATA, hdev, 8203 cp, sizeof(*cp)); 8204 if (!cmd) 8205 return -ENOMEM; 8206 8207 err = hci_cmd_sync_queue(hdev, read_local_oob_data_sync, cmd, 8208 read_local_oob_ext_data_complete); 8209 8210 if (err < 0) { 8211 mgmt_pending_remove(cmd); 8212 return err; 8213 } 8214 8215 return 0; 8216 } 8217 8218 static int read_local_oob_ext_data(struct sock *sk, struct hci_dev *hdev, 8219 void *data, u16 data_len) 8220 { 8221 struct mgmt_cp_read_local_oob_ext_data *cp = data; 8222 struct mgmt_rp_read_local_oob_ext_data *rp; 8223 size_t rp_len; 8224 u16 eir_len; 8225 u8 status, flags, role, addr[7], hash[16], rand[16]; 8226 int err; 8227 8228 bt_dev_dbg(hdev, "sock %p", sk); 8229 8230 if (hdev_is_powered(hdev)) { 8231 switch (cp->type) { 8232 case BIT(BDADDR_BREDR): 8233 status = mgmt_bredr_support(hdev); 8234 if (status) 8235 eir_len = 0; 8236 else 8237 eir_len = 5; 8238 break; 8239 case (BIT(BDADDR_LE_PUBLIC) | BIT(BDADDR_LE_RANDOM)): 8240 status = mgmt_le_support(hdev); 8241 if (status) 8242 eir_len = 0; 8243 else 8244 eir_len = 9 + 3 + 18 + 18 + 3; 8245 break; 8246 default: 8247 status = MGMT_STATUS_INVALID_PARAMS; 8248 eir_len = 0; 8249 break; 8250 } 8251 } else { 8252 status = MGMT_STATUS_NOT_POWERED; 8253 eir_len = 0; 8254 } 8255 8256 rp_len = sizeof(*rp) + eir_len; 8257 rp = kmalloc(rp_len, GFP_ATOMIC); 8258 if (!rp) 8259 return -ENOMEM; 8260 8261 if (!status && !lmp_ssp_capable(hdev)) { 8262 status = MGMT_STATUS_NOT_SUPPORTED; 8263 eir_len = 0; 8264 } 8265 8266 if (status) 8267 goto complete; 8268 8269 hci_dev_lock(hdev); 8270 8271 eir_len = 0; 8272 switch (cp->type) { 8273 case BIT(BDADDR_BREDR): 8274 if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) { 8275 err = read_local_ssp_oob_req(hdev, sk, cp); 8276 hci_dev_unlock(hdev); 8277 if (!err) 8278 goto done; 8279 8280 status = MGMT_STATUS_FAILED; 8281 goto complete; 8282 } else { 8283 eir_len = eir_append_data(rp->eir, eir_len, 8284 EIR_CLASS_OF_DEV, 8285 hdev->dev_class, 3); 8286 } 8287 break; 8288 case (BIT(BDADDR_LE_PUBLIC) | BIT(BDADDR_LE_RANDOM)): 8289 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED) && 8290 smp_generate_oob(hdev, hash, rand) < 0) { 8291 hci_dev_unlock(hdev); 8292 status = MGMT_STATUS_FAILED; 8293 goto complete; 8294 } 8295 8296 /* This should return the active RPA, but since the RPA 8297 * is only programmed on demand, it is really hard to fill 8298 * this in at the moment. For now disallow retrieving 8299 * local out-of-band data when privacy is in use. 8300 * 8301 * Returning the identity address will not help here since 8302 * pairing happens before the identity resolving key is 8303 * known and thus the connection establishment happens 8304 * based on the RPA and not the identity address. 8305 */ 8306 if (hci_dev_test_flag(hdev, HCI_PRIVACY)) { 8307 hci_dev_unlock(hdev); 8308 status = MGMT_STATUS_REJECTED; 8309 goto complete; 8310 } 8311 8312 if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) || 8313 !bacmp(&hdev->bdaddr, BDADDR_ANY) || 8314 (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) && 8315 bacmp(&hdev->static_addr, BDADDR_ANY))) { 8316 memcpy(addr, &hdev->static_addr, 6); 8317 addr[6] = 0x01; 8318 } else { 8319 memcpy(addr, &hdev->bdaddr, 6); 8320 addr[6] = 0x00; 8321 } 8322 8323 eir_len = eir_append_data(rp->eir, eir_len, EIR_LE_BDADDR, 8324 addr, sizeof(addr)); 8325 8326 if (hci_dev_test_flag(hdev, HCI_ADVERTISING)) 8327 role = 0x02; 8328 else 8329 role = 0x01; 8330 8331 eir_len = eir_append_data(rp->eir, eir_len, EIR_LE_ROLE, 8332 &role, sizeof(role)); 8333 8334 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED)) { 8335 eir_len = eir_append_data(rp->eir, eir_len, 8336 EIR_LE_SC_CONFIRM, 8337 hash, sizeof(hash)); 8338 8339 eir_len = eir_append_data(rp->eir, eir_len, 8340 EIR_LE_SC_RANDOM, 8341 rand, sizeof(rand)); 8342 } 8343 8344 flags = mgmt_get_adv_discov_flags(hdev); 8345 8346 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) 8347 flags |= LE_AD_NO_BREDR; 8348 8349 eir_len = eir_append_data(rp->eir, eir_len, EIR_FLAGS, 8350 &flags, sizeof(flags)); 8351 break; 8352 } 8353 8354 hci_dev_unlock(hdev); 8355 8356 hci_sock_set_flag(sk, HCI_MGMT_OOB_DATA_EVENTS); 8357 8358 status = MGMT_STATUS_SUCCESS; 8359 8360 complete: 8361 rp->type = cp->type; 8362 rp->eir_len = cpu_to_le16(eir_len); 8363 8364 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_EXT_DATA, 8365 status, rp, sizeof(*rp) + eir_len); 8366 if (err < 0 || status) 8367 goto done; 8368 8369 err = mgmt_limited_event(MGMT_EV_LOCAL_OOB_DATA_UPDATED, hdev, 8370 rp, sizeof(*rp) + eir_len, 8371 HCI_MGMT_OOB_DATA_EVENTS, sk); 8372 8373 done: 8374 kfree(rp); 8375 8376 return err; 8377 } 8378 8379 static u32 get_supported_adv_flags(struct hci_dev *hdev) 8380 { 8381 u32 flags = 0; 8382 8383 flags |= MGMT_ADV_FLAG_CONNECTABLE; 8384 flags |= MGMT_ADV_FLAG_DISCOV; 8385 flags |= MGMT_ADV_FLAG_LIMITED_DISCOV; 8386 flags |= MGMT_ADV_FLAG_MANAGED_FLAGS; 8387 flags |= MGMT_ADV_FLAG_APPEARANCE; 8388 flags |= MGMT_ADV_FLAG_LOCAL_NAME; 8389 flags |= MGMT_ADV_PARAM_DURATION; 8390 flags |= MGMT_ADV_PARAM_TIMEOUT; 8391 flags |= MGMT_ADV_PARAM_INTERVALS; 8392 flags |= MGMT_ADV_PARAM_TX_POWER; 8393 flags |= MGMT_ADV_PARAM_SCAN_RSP; 8394 8395 /* In extended adv TX_POWER returned from Set Adv Param 8396 * will be always valid. 8397 */ 8398 if (hdev->adv_tx_power != HCI_TX_POWER_INVALID || ext_adv_capable(hdev)) 8399 flags |= MGMT_ADV_FLAG_TX_POWER; 8400 8401 if (ext_adv_capable(hdev)) { 8402 flags |= MGMT_ADV_FLAG_SEC_1M; 8403 flags |= MGMT_ADV_FLAG_HW_OFFLOAD; 8404 flags |= MGMT_ADV_FLAG_CAN_SET_TX_POWER; 8405 8406 if (le_2m_capable(hdev)) 8407 flags |= MGMT_ADV_FLAG_SEC_2M; 8408 8409 if (le_coded_capable(hdev)) 8410 flags |= MGMT_ADV_FLAG_SEC_CODED; 8411 } 8412 8413 return flags; 8414 } 8415 8416 static int read_adv_features(struct sock *sk, struct hci_dev *hdev, 8417 void *data, u16 data_len) 8418 { 8419 struct mgmt_rp_read_adv_features *rp; 8420 size_t rp_len; 8421 int err; 8422 struct adv_info *adv_instance; 8423 u32 supported_flags; 8424 u8 *instance; 8425 8426 bt_dev_dbg(hdev, "sock %p", sk); 8427 8428 if (!lmp_le_capable(hdev)) 8429 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_ADV_FEATURES, 8430 MGMT_STATUS_REJECTED); 8431 8432 hci_dev_lock(hdev); 8433 8434 rp_len = sizeof(*rp) + hdev->adv_instance_cnt; 8435 rp = kmalloc(rp_len, GFP_ATOMIC); 8436 if (!rp) { 8437 hci_dev_unlock(hdev); 8438 return -ENOMEM; 8439 } 8440 8441 supported_flags = get_supported_adv_flags(hdev); 8442 8443 rp->supported_flags = cpu_to_le32(supported_flags); 8444 rp->max_adv_data_len = max_adv_len(hdev); 8445 rp->max_scan_rsp_len = max_adv_len(hdev); 8446 rp->max_instances = hdev->le_num_of_adv_sets; 8447 rp->num_instances = hdev->adv_instance_cnt; 8448 8449 instance = rp->instance; 8450 list_for_each_entry(adv_instance, &hdev->adv_instances, list) { 8451 /* Only instances 1-le_num_of_adv_sets are externally visible */ 8452 if (adv_instance->instance <= hdev->adv_instance_cnt) { 8453 *instance = adv_instance->instance; 8454 instance++; 8455 } else { 8456 rp->num_instances--; 8457 rp_len--; 8458 } 8459 } 8460 8461 hci_dev_unlock(hdev); 8462 8463 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_ADV_FEATURES, 8464 MGMT_STATUS_SUCCESS, rp, rp_len); 8465 8466 kfree(rp); 8467 8468 return err; 8469 } 8470 8471 static u8 calculate_name_len(struct hci_dev *hdev) 8472 { 8473 u8 buf[HCI_MAX_SHORT_NAME_LENGTH + 3]; 8474 8475 return eir_append_local_name(hdev, buf, 0); 8476 } 8477 8478 static u8 tlv_data_max_len(struct hci_dev *hdev, u32 adv_flags, 8479 bool is_adv_data) 8480 { 8481 u8 max_len = max_adv_len(hdev); 8482 8483 if (is_adv_data) { 8484 if (adv_flags & (MGMT_ADV_FLAG_DISCOV | 8485 MGMT_ADV_FLAG_LIMITED_DISCOV | 8486 MGMT_ADV_FLAG_MANAGED_FLAGS)) 8487 max_len -= 3; 8488 8489 if (adv_flags & MGMT_ADV_FLAG_TX_POWER) 8490 max_len -= 3; 8491 } else { 8492 if (adv_flags & MGMT_ADV_FLAG_LOCAL_NAME) 8493 max_len -= calculate_name_len(hdev); 8494 8495 if (adv_flags & (MGMT_ADV_FLAG_APPEARANCE)) 8496 max_len -= 4; 8497 } 8498 8499 return max_len; 8500 } 8501 8502 static bool flags_managed(u32 adv_flags) 8503 { 8504 return adv_flags & (MGMT_ADV_FLAG_DISCOV | 8505 MGMT_ADV_FLAG_LIMITED_DISCOV | 8506 MGMT_ADV_FLAG_MANAGED_FLAGS); 8507 } 8508 8509 static bool tx_power_managed(u32 adv_flags) 8510 { 8511 return adv_flags & MGMT_ADV_FLAG_TX_POWER; 8512 } 8513 8514 static bool name_managed(u32 adv_flags) 8515 { 8516 return adv_flags & MGMT_ADV_FLAG_LOCAL_NAME; 8517 } 8518 8519 static bool appearance_managed(u32 adv_flags) 8520 { 8521 return adv_flags & MGMT_ADV_FLAG_APPEARANCE; 8522 } 8523 8524 static bool tlv_data_is_valid(struct hci_dev *hdev, u32 adv_flags, u8 *data, 8525 u8 len, bool is_adv_data) 8526 { 8527 int i, cur_len; 8528 u8 max_len; 8529 8530 max_len = tlv_data_max_len(hdev, adv_flags, is_adv_data); 8531 8532 if (len > max_len) 8533 return false; 8534 8535 /* Make sure that the data is correctly formatted. */ 8536 for (i = 0; i < len; i += (cur_len + 1)) { 8537 cur_len = data[i]; 8538 8539 if (!cur_len) 8540 continue; 8541 8542 if (data[i + 1] == EIR_FLAGS && 8543 (!is_adv_data || flags_managed(adv_flags))) 8544 return false; 8545 8546 if (data[i + 1] == EIR_TX_POWER && tx_power_managed(adv_flags)) 8547 return false; 8548 8549 if (data[i + 1] == EIR_NAME_COMPLETE && name_managed(adv_flags)) 8550 return false; 8551 8552 if (data[i + 1] == EIR_NAME_SHORT && name_managed(adv_flags)) 8553 return false; 8554 8555 if (data[i + 1] == EIR_APPEARANCE && 8556 appearance_managed(adv_flags)) 8557 return false; 8558 8559 /* If the current field length would exceed the total data 8560 * length, then it's invalid. 8561 */ 8562 if (i + cur_len >= len) 8563 return false; 8564 } 8565 8566 return true; 8567 } 8568 8569 static bool requested_adv_flags_are_valid(struct hci_dev *hdev, u32 adv_flags) 8570 { 8571 u32 supported_flags, phy_flags; 8572 8573 /* The current implementation only supports a subset of the specified 8574 * flags. Also need to check mutual exclusiveness of sec flags. 8575 */ 8576 supported_flags = get_supported_adv_flags(hdev); 8577 phy_flags = adv_flags & MGMT_ADV_FLAG_SEC_MASK; 8578 if (adv_flags & ~supported_flags || 8579 ((phy_flags && (phy_flags ^ (phy_flags & -phy_flags))))) 8580 return false; 8581 8582 return true; 8583 } 8584 8585 static bool adv_busy(struct hci_dev *hdev) 8586 { 8587 return pending_find(MGMT_OP_SET_LE, hdev); 8588 } 8589 8590 static void add_adv_complete(struct hci_dev *hdev, struct sock *sk, u8 instance, 8591 int err) 8592 { 8593 struct adv_info *adv, *n; 8594 8595 bt_dev_dbg(hdev, "err %d", err); 8596 8597 hci_dev_lock(hdev); 8598 8599 list_for_each_entry_safe(adv, n, &hdev->adv_instances, list) { 8600 u8 instance; 8601 8602 if (!adv->pending) 8603 continue; 8604 8605 if (!err) { 8606 adv->pending = false; 8607 continue; 8608 } 8609 8610 instance = adv->instance; 8611 8612 if (hdev->cur_adv_instance == instance) 8613 cancel_adv_timeout(hdev); 8614 8615 hci_remove_adv_instance(hdev, instance); 8616 mgmt_advertising_removed(sk, hdev, instance); 8617 } 8618 8619 hci_dev_unlock(hdev); 8620 } 8621 8622 static void add_advertising_complete(struct hci_dev *hdev, void *data, int err) 8623 { 8624 struct mgmt_pending_cmd *cmd = data; 8625 struct mgmt_cp_add_advertising *cp = cmd->param; 8626 struct mgmt_rp_add_advertising rp; 8627 8628 memset(&rp, 0, sizeof(rp)); 8629 8630 rp.instance = cp->instance; 8631 8632 if (err) 8633 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 8634 mgmt_status(err)); 8635 else 8636 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 8637 mgmt_status(err), &rp, sizeof(rp)); 8638 8639 add_adv_complete(hdev, cmd->sk, cp->instance, err); 8640 8641 mgmt_pending_free(cmd); 8642 } 8643 8644 static int add_advertising_sync(struct hci_dev *hdev, void *data) 8645 { 8646 struct mgmt_pending_cmd *cmd = data; 8647 struct mgmt_cp_add_advertising *cp = cmd->param; 8648 8649 return hci_schedule_adv_instance_sync(hdev, cp->instance, true); 8650 } 8651 8652 static int add_advertising(struct sock *sk, struct hci_dev *hdev, 8653 void *data, u16 data_len) 8654 { 8655 struct mgmt_cp_add_advertising *cp = data; 8656 struct mgmt_rp_add_advertising rp; 8657 u32 flags; 8658 u8 status; 8659 u16 timeout, duration; 8660 unsigned int prev_instance_cnt; 8661 u8 schedule_instance = 0; 8662 struct adv_info *adv, *next_instance; 8663 int err; 8664 struct mgmt_pending_cmd *cmd; 8665 8666 bt_dev_dbg(hdev, "sock %p", sk); 8667 8668 status = mgmt_le_support(hdev); 8669 if (status) 8670 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8671 status); 8672 8673 if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets) 8674 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8675 MGMT_STATUS_INVALID_PARAMS); 8676 8677 if (data_len != sizeof(*cp) + cp->adv_data_len + cp->scan_rsp_len) 8678 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8679 MGMT_STATUS_INVALID_PARAMS); 8680 8681 flags = __le32_to_cpu(cp->flags); 8682 timeout = __le16_to_cpu(cp->timeout); 8683 duration = __le16_to_cpu(cp->duration); 8684 8685 if (!requested_adv_flags_are_valid(hdev, flags)) 8686 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8687 MGMT_STATUS_INVALID_PARAMS); 8688 8689 hci_dev_lock(hdev); 8690 8691 if (timeout && !hdev_is_powered(hdev)) { 8692 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8693 MGMT_STATUS_REJECTED); 8694 goto unlock; 8695 } 8696 8697 if (adv_busy(hdev)) { 8698 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8699 MGMT_STATUS_BUSY); 8700 goto unlock; 8701 } 8702 8703 if (!tlv_data_is_valid(hdev, flags, cp->data, cp->adv_data_len, true) || 8704 !tlv_data_is_valid(hdev, flags, cp->data + cp->adv_data_len, 8705 cp->scan_rsp_len, false)) { 8706 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8707 MGMT_STATUS_INVALID_PARAMS); 8708 goto unlock; 8709 } 8710 8711 prev_instance_cnt = hdev->adv_instance_cnt; 8712 8713 adv = hci_add_adv_instance(hdev, cp->instance, flags, 8714 cp->adv_data_len, cp->data, 8715 cp->scan_rsp_len, 8716 cp->data + cp->adv_data_len, 8717 timeout, duration, 8718 HCI_ADV_TX_POWER_NO_PREFERENCE, 8719 hdev->le_adv_min_interval, 8720 hdev->le_adv_max_interval, 0); 8721 if (IS_ERR(adv)) { 8722 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8723 MGMT_STATUS_FAILED); 8724 goto unlock; 8725 } 8726 8727 /* Only trigger an advertising added event if a new instance was 8728 * actually added. 8729 */ 8730 if (hdev->adv_instance_cnt > prev_instance_cnt) 8731 mgmt_advertising_added(sk, hdev, cp->instance); 8732 8733 if (hdev->cur_adv_instance == cp->instance) { 8734 /* If the currently advertised instance is being changed then 8735 * cancel the current advertising and schedule the next 8736 * instance. If there is only one instance then the overridden 8737 * advertising data will be visible right away. 8738 */ 8739 cancel_adv_timeout(hdev); 8740 8741 next_instance = hci_get_next_instance(hdev, cp->instance); 8742 if (next_instance) 8743 schedule_instance = next_instance->instance; 8744 } else if (!hdev->adv_instance_timeout) { 8745 /* Immediately advertise the new instance if no other 8746 * instance is currently being advertised. 8747 */ 8748 schedule_instance = cp->instance; 8749 } 8750 8751 /* If the HCI_ADVERTISING flag is set or the device isn't powered or 8752 * there is no instance to be advertised then we have no HCI 8753 * communication to make. Simply return. 8754 */ 8755 if (!hdev_is_powered(hdev) || 8756 hci_dev_test_flag(hdev, HCI_ADVERTISING) || 8757 !schedule_instance) { 8758 rp.instance = cp->instance; 8759 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_ADVERTISING, 8760 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 8761 goto unlock; 8762 } 8763 8764 /* We're good to go, update advertising data, parameters, and start 8765 * advertising. 8766 */ 8767 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_ADVERTISING, hdev, data, 8768 data_len); 8769 if (!cmd) { 8770 err = -ENOMEM; 8771 goto unlock; 8772 } 8773 8774 cp->instance = schedule_instance; 8775 8776 err = hci_cmd_sync_queue(hdev, add_advertising_sync, cmd, 8777 add_advertising_complete); 8778 if (err < 0) 8779 mgmt_pending_free(cmd); 8780 8781 unlock: 8782 hci_dev_unlock(hdev); 8783 8784 return err; 8785 } 8786 8787 static void add_ext_adv_params_complete(struct hci_dev *hdev, void *data, 8788 int err) 8789 { 8790 struct mgmt_pending_cmd *cmd = data; 8791 struct mgmt_cp_add_ext_adv_params *cp = cmd->param; 8792 struct mgmt_rp_add_ext_adv_params rp; 8793 struct adv_info *adv; 8794 u32 flags; 8795 8796 BT_DBG("%s", hdev->name); 8797 8798 hci_dev_lock(hdev); 8799 8800 adv = hci_find_adv_instance(hdev, cp->instance); 8801 if (!adv) 8802 goto unlock; 8803 8804 rp.instance = cp->instance; 8805 rp.tx_power = adv->tx_power; 8806 8807 /* While we're at it, inform userspace of the available space for this 8808 * advertisement, given the flags that will be used. 8809 */ 8810 flags = __le32_to_cpu(cp->flags); 8811 rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true); 8812 rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false); 8813 8814 if (err) { 8815 /* If this advertisement was previously advertising and we 8816 * failed to update it, we signal that it has been removed and 8817 * delete its structure 8818 */ 8819 if (!adv->pending) 8820 mgmt_advertising_removed(cmd->sk, hdev, cp->instance); 8821 8822 hci_remove_adv_instance(hdev, cp->instance); 8823 8824 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 8825 mgmt_status(err)); 8826 } else { 8827 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 8828 mgmt_status(err), &rp, sizeof(rp)); 8829 } 8830 8831 unlock: 8832 if (cmd) 8833 mgmt_pending_free(cmd); 8834 8835 hci_dev_unlock(hdev); 8836 } 8837 8838 static int add_ext_adv_params_sync(struct hci_dev *hdev, void *data) 8839 { 8840 struct mgmt_pending_cmd *cmd = data; 8841 struct mgmt_cp_add_ext_adv_params *cp = cmd->param; 8842 8843 return hci_setup_ext_adv_instance_sync(hdev, cp->instance); 8844 } 8845 8846 static int add_ext_adv_params(struct sock *sk, struct hci_dev *hdev, 8847 void *data, u16 data_len) 8848 { 8849 struct mgmt_cp_add_ext_adv_params *cp = data; 8850 struct mgmt_rp_add_ext_adv_params rp; 8851 struct mgmt_pending_cmd *cmd = NULL; 8852 struct adv_info *adv; 8853 u32 flags, min_interval, max_interval; 8854 u16 timeout, duration; 8855 u8 status; 8856 s8 tx_power; 8857 int err; 8858 8859 BT_DBG("%s", hdev->name); 8860 8861 status = mgmt_le_support(hdev); 8862 if (status) 8863 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8864 status); 8865 8866 if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets) 8867 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8868 MGMT_STATUS_INVALID_PARAMS); 8869 8870 /* The purpose of breaking add_advertising into two separate MGMT calls 8871 * for params and data is to allow more parameters to be added to this 8872 * structure in the future. For this reason, we verify that we have the 8873 * bare minimum structure we know of when the interface was defined. Any 8874 * extra parameters we don't know about will be ignored in this request. 8875 */ 8876 if (data_len < MGMT_ADD_EXT_ADV_PARAMS_MIN_SIZE) 8877 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8878 MGMT_STATUS_INVALID_PARAMS); 8879 8880 flags = __le32_to_cpu(cp->flags); 8881 8882 if (!requested_adv_flags_are_valid(hdev, flags)) 8883 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8884 MGMT_STATUS_INVALID_PARAMS); 8885 8886 hci_dev_lock(hdev); 8887 8888 /* In new interface, we require that we are powered to register */ 8889 if (!hdev_is_powered(hdev)) { 8890 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8891 MGMT_STATUS_REJECTED); 8892 goto unlock; 8893 } 8894 8895 if (adv_busy(hdev)) { 8896 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8897 MGMT_STATUS_BUSY); 8898 goto unlock; 8899 } 8900 8901 /* Parse defined parameters from request, use defaults otherwise */ 8902 timeout = (flags & MGMT_ADV_PARAM_TIMEOUT) ? 8903 __le16_to_cpu(cp->timeout) : 0; 8904 8905 duration = (flags & MGMT_ADV_PARAM_DURATION) ? 8906 __le16_to_cpu(cp->duration) : 8907 hdev->def_multi_adv_rotation_duration; 8908 8909 min_interval = (flags & MGMT_ADV_PARAM_INTERVALS) ? 8910 __le32_to_cpu(cp->min_interval) : 8911 hdev->le_adv_min_interval; 8912 8913 max_interval = (flags & MGMT_ADV_PARAM_INTERVALS) ? 8914 __le32_to_cpu(cp->max_interval) : 8915 hdev->le_adv_max_interval; 8916 8917 tx_power = (flags & MGMT_ADV_PARAM_TX_POWER) ? 8918 cp->tx_power : 8919 HCI_ADV_TX_POWER_NO_PREFERENCE; 8920 8921 /* Create advertising instance with no advertising or response data */ 8922 adv = hci_add_adv_instance(hdev, cp->instance, flags, 0, NULL, 0, NULL, 8923 timeout, duration, tx_power, min_interval, 8924 max_interval, 0); 8925 8926 if (IS_ERR(adv)) { 8927 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS, 8928 MGMT_STATUS_FAILED); 8929 goto unlock; 8930 } 8931 8932 /* Submit request for advertising params if ext adv available */ 8933 if (ext_adv_capable(hdev)) { 8934 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_EXT_ADV_PARAMS, hdev, 8935 data, data_len); 8936 if (!cmd) { 8937 err = -ENOMEM; 8938 hci_remove_adv_instance(hdev, cp->instance); 8939 goto unlock; 8940 } 8941 8942 err = hci_cmd_sync_queue(hdev, add_ext_adv_params_sync, cmd, 8943 add_ext_adv_params_complete); 8944 if (err < 0) 8945 mgmt_pending_free(cmd); 8946 } else { 8947 rp.instance = cp->instance; 8948 rp.tx_power = HCI_ADV_TX_POWER_NO_PREFERENCE; 8949 rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true); 8950 rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false); 8951 err = mgmt_cmd_complete(sk, hdev->id, 8952 MGMT_OP_ADD_EXT_ADV_PARAMS, 8953 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 8954 } 8955 8956 unlock: 8957 hci_dev_unlock(hdev); 8958 8959 return err; 8960 } 8961 8962 static void add_ext_adv_data_complete(struct hci_dev *hdev, void *data, int err) 8963 { 8964 struct mgmt_pending_cmd *cmd = data; 8965 struct mgmt_cp_add_ext_adv_data *cp = cmd->param; 8966 struct mgmt_rp_add_advertising rp; 8967 8968 add_adv_complete(hdev, cmd->sk, cp->instance, err); 8969 8970 memset(&rp, 0, sizeof(rp)); 8971 8972 rp.instance = cp->instance; 8973 8974 if (err) 8975 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 8976 mgmt_status(err)); 8977 else 8978 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 8979 mgmt_status(err), &rp, sizeof(rp)); 8980 8981 mgmt_pending_free(cmd); 8982 } 8983 8984 static int add_ext_adv_data_sync(struct hci_dev *hdev, void *data) 8985 { 8986 struct mgmt_pending_cmd *cmd = data; 8987 struct mgmt_cp_add_ext_adv_data *cp = cmd->param; 8988 int err; 8989 8990 if (ext_adv_capable(hdev)) { 8991 err = hci_update_adv_data_sync(hdev, cp->instance); 8992 if (err) 8993 return err; 8994 8995 err = hci_update_scan_rsp_data_sync(hdev, cp->instance); 8996 if (err) 8997 return err; 8998 8999 return hci_enable_ext_advertising_sync(hdev, cp->instance); 9000 } 9001 9002 return hci_schedule_adv_instance_sync(hdev, cp->instance, true); 9003 } 9004 9005 static int add_ext_adv_data(struct sock *sk, struct hci_dev *hdev, void *data, 9006 u16 data_len) 9007 { 9008 struct mgmt_cp_add_ext_adv_data *cp = data; 9009 struct mgmt_rp_add_ext_adv_data rp; 9010 u8 schedule_instance = 0; 9011 struct adv_info *next_instance; 9012 struct adv_info *adv_instance; 9013 int err = 0; 9014 struct mgmt_pending_cmd *cmd; 9015 9016 BT_DBG("%s", hdev->name); 9017 9018 hci_dev_lock(hdev); 9019 9020 adv_instance = hci_find_adv_instance(hdev, cp->instance); 9021 9022 if (!adv_instance) { 9023 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9024 MGMT_STATUS_INVALID_PARAMS); 9025 goto unlock; 9026 } 9027 9028 /* In new interface, we require that we are powered to register */ 9029 if (!hdev_is_powered(hdev)) { 9030 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9031 MGMT_STATUS_REJECTED); 9032 goto clear_new_instance; 9033 } 9034 9035 if (adv_busy(hdev)) { 9036 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9037 MGMT_STATUS_BUSY); 9038 goto clear_new_instance; 9039 } 9040 9041 /* Validate new data */ 9042 if (!tlv_data_is_valid(hdev, adv_instance->flags, cp->data, 9043 cp->adv_data_len, true) || 9044 !tlv_data_is_valid(hdev, adv_instance->flags, cp->data + 9045 cp->adv_data_len, cp->scan_rsp_len, false)) { 9046 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9047 MGMT_STATUS_INVALID_PARAMS); 9048 goto clear_new_instance; 9049 } 9050 9051 /* Set the data in the advertising instance */ 9052 hci_set_adv_instance_data(hdev, cp->instance, cp->adv_data_len, 9053 cp->data, cp->scan_rsp_len, 9054 cp->data + cp->adv_data_len); 9055 9056 /* If using software rotation, determine next instance to use */ 9057 if (hdev->cur_adv_instance == cp->instance) { 9058 /* If the currently advertised instance is being changed 9059 * then cancel the current advertising and schedule the 9060 * next instance. If there is only one instance then the 9061 * overridden advertising data will be visible right 9062 * away 9063 */ 9064 cancel_adv_timeout(hdev); 9065 9066 next_instance = hci_get_next_instance(hdev, cp->instance); 9067 if (next_instance) 9068 schedule_instance = next_instance->instance; 9069 } else if (!hdev->adv_instance_timeout) { 9070 /* Immediately advertise the new instance if no other 9071 * instance is currently being advertised. 9072 */ 9073 schedule_instance = cp->instance; 9074 } 9075 9076 /* If the HCI_ADVERTISING flag is set or there is no instance to 9077 * be advertised then we have no HCI communication to make. 9078 * Simply return. 9079 */ 9080 if (hci_dev_test_flag(hdev, HCI_ADVERTISING) || !schedule_instance) { 9081 if (adv_instance->pending) { 9082 mgmt_advertising_added(sk, hdev, cp->instance); 9083 adv_instance->pending = false; 9084 } 9085 rp.instance = cp->instance; 9086 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA, 9087 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 9088 goto unlock; 9089 } 9090 9091 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_EXT_ADV_DATA, hdev, data, 9092 data_len); 9093 if (!cmd) { 9094 err = -ENOMEM; 9095 goto clear_new_instance; 9096 } 9097 9098 err = hci_cmd_sync_queue(hdev, add_ext_adv_data_sync, cmd, 9099 add_ext_adv_data_complete); 9100 if (err < 0) { 9101 mgmt_pending_free(cmd); 9102 goto clear_new_instance; 9103 } 9104 9105 /* We were successful in updating data, so trigger advertising_added 9106 * event if this is an instance that wasn't previously advertising. If 9107 * a failure occurs in the requests we initiated, we will remove the 9108 * instance again in add_advertising_complete 9109 */ 9110 if (adv_instance->pending) 9111 mgmt_advertising_added(sk, hdev, cp->instance); 9112 9113 goto unlock; 9114 9115 clear_new_instance: 9116 hci_remove_adv_instance(hdev, cp->instance); 9117 9118 unlock: 9119 hci_dev_unlock(hdev); 9120 9121 return err; 9122 } 9123 9124 static void remove_advertising_complete(struct hci_dev *hdev, void *data, 9125 int err) 9126 { 9127 struct mgmt_pending_cmd *cmd = data; 9128 struct mgmt_cp_remove_advertising *cp = cmd->param; 9129 struct mgmt_rp_remove_advertising rp; 9130 9131 bt_dev_dbg(hdev, "err %d", err); 9132 9133 memset(&rp, 0, sizeof(rp)); 9134 rp.instance = cp->instance; 9135 9136 if (err) 9137 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, 9138 mgmt_status(err)); 9139 else 9140 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, 9141 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 9142 9143 mgmt_pending_free(cmd); 9144 } 9145 9146 static int remove_advertising_sync(struct hci_dev *hdev, void *data) 9147 { 9148 struct mgmt_pending_cmd *cmd = data; 9149 struct mgmt_cp_remove_advertising *cp = cmd->param; 9150 int err; 9151 9152 err = hci_remove_advertising_sync(hdev, cmd->sk, cp->instance, true); 9153 if (err) 9154 return err; 9155 9156 if (list_empty(&hdev->adv_instances)) 9157 err = hci_disable_advertising_sync(hdev); 9158 9159 return err; 9160 } 9161 9162 static int remove_advertising(struct sock *sk, struct hci_dev *hdev, 9163 void *data, u16 data_len) 9164 { 9165 struct mgmt_cp_remove_advertising *cp = data; 9166 struct mgmt_pending_cmd *cmd; 9167 int err; 9168 9169 bt_dev_dbg(hdev, "sock %p", sk); 9170 9171 hci_dev_lock(hdev); 9172 9173 if (cp->instance && !hci_find_adv_instance(hdev, cp->instance)) { 9174 err = mgmt_cmd_status(sk, hdev->id, 9175 MGMT_OP_REMOVE_ADVERTISING, 9176 MGMT_STATUS_INVALID_PARAMS); 9177 goto unlock; 9178 } 9179 9180 if (pending_find(MGMT_OP_SET_LE, hdev)) { 9181 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADVERTISING, 9182 MGMT_STATUS_BUSY); 9183 goto unlock; 9184 } 9185 9186 if (list_empty(&hdev->adv_instances)) { 9187 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADVERTISING, 9188 MGMT_STATUS_INVALID_PARAMS); 9189 goto unlock; 9190 } 9191 9192 cmd = mgmt_pending_new(sk, MGMT_OP_REMOVE_ADVERTISING, hdev, data, 9193 data_len); 9194 if (!cmd) { 9195 err = -ENOMEM; 9196 goto unlock; 9197 } 9198 9199 err = hci_cmd_sync_queue(hdev, remove_advertising_sync, cmd, 9200 remove_advertising_complete); 9201 if (err < 0) 9202 mgmt_pending_free(cmd); 9203 9204 unlock: 9205 hci_dev_unlock(hdev); 9206 9207 return err; 9208 } 9209 9210 static int get_adv_size_info(struct sock *sk, struct hci_dev *hdev, 9211 void *data, u16 data_len) 9212 { 9213 struct mgmt_cp_get_adv_size_info *cp = data; 9214 struct mgmt_rp_get_adv_size_info rp; 9215 u32 flags, supported_flags; 9216 9217 bt_dev_dbg(hdev, "sock %p", sk); 9218 9219 if (!lmp_le_capable(hdev)) 9220 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9221 MGMT_STATUS_REJECTED); 9222 9223 if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets) 9224 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9225 MGMT_STATUS_INVALID_PARAMS); 9226 9227 flags = __le32_to_cpu(cp->flags); 9228 9229 /* The current implementation only supports a subset of the specified 9230 * flags. 9231 */ 9232 supported_flags = get_supported_adv_flags(hdev); 9233 if (flags & ~supported_flags) 9234 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9235 MGMT_STATUS_INVALID_PARAMS); 9236 9237 rp.instance = cp->instance; 9238 rp.flags = cp->flags; 9239 rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true); 9240 rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false); 9241 9242 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO, 9243 MGMT_STATUS_SUCCESS, &rp, sizeof(rp)); 9244 } 9245 9246 static const struct hci_mgmt_handler mgmt_handlers[] = { 9247 { NULL }, /* 0x0000 (no command) */ 9248 { read_version, MGMT_READ_VERSION_SIZE, 9249 HCI_MGMT_NO_HDEV | 9250 HCI_MGMT_UNTRUSTED }, 9251 { read_commands, MGMT_READ_COMMANDS_SIZE, 9252 HCI_MGMT_NO_HDEV | 9253 HCI_MGMT_UNTRUSTED }, 9254 { read_index_list, MGMT_READ_INDEX_LIST_SIZE, 9255 HCI_MGMT_NO_HDEV | 9256 HCI_MGMT_UNTRUSTED }, 9257 { read_controller_info, MGMT_READ_INFO_SIZE, 9258 HCI_MGMT_UNTRUSTED }, 9259 { set_powered, MGMT_SETTING_SIZE }, 9260 { set_discoverable, MGMT_SET_DISCOVERABLE_SIZE }, 9261 { set_connectable, MGMT_SETTING_SIZE }, 9262 { set_fast_connectable, MGMT_SETTING_SIZE }, 9263 { set_bondable, MGMT_SETTING_SIZE }, 9264 { set_link_security, MGMT_SETTING_SIZE }, 9265 { set_ssp, MGMT_SETTING_SIZE }, 9266 { set_hs, MGMT_SETTING_SIZE }, 9267 { set_le, MGMT_SETTING_SIZE }, 9268 { set_dev_class, MGMT_SET_DEV_CLASS_SIZE }, 9269 { set_local_name, MGMT_SET_LOCAL_NAME_SIZE }, 9270 { add_uuid, MGMT_ADD_UUID_SIZE }, 9271 { remove_uuid, MGMT_REMOVE_UUID_SIZE }, 9272 { load_link_keys, MGMT_LOAD_LINK_KEYS_SIZE, 9273 HCI_MGMT_VAR_LEN }, 9274 { load_long_term_keys, MGMT_LOAD_LONG_TERM_KEYS_SIZE, 9275 HCI_MGMT_VAR_LEN }, 9276 { disconnect, MGMT_DISCONNECT_SIZE }, 9277 { get_connections, MGMT_GET_CONNECTIONS_SIZE }, 9278 { pin_code_reply, MGMT_PIN_CODE_REPLY_SIZE }, 9279 { pin_code_neg_reply, MGMT_PIN_CODE_NEG_REPLY_SIZE }, 9280 { set_io_capability, MGMT_SET_IO_CAPABILITY_SIZE }, 9281 { pair_device, MGMT_PAIR_DEVICE_SIZE }, 9282 { cancel_pair_device, MGMT_CANCEL_PAIR_DEVICE_SIZE }, 9283 { unpair_device, MGMT_UNPAIR_DEVICE_SIZE }, 9284 { user_confirm_reply, MGMT_USER_CONFIRM_REPLY_SIZE }, 9285 { user_confirm_neg_reply, MGMT_USER_CONFIRM_NEG_REPLY_SIZE }, 9286 { user_passkey_reply, MGMT_USER_PASSKEY_REPLY_SIZE }, 9287 { user_passkey_neg_reply, MGMT_USER_PASSKEY_NEG_REPLY_SIZE }, 9288 { read_local_oob_data, MGMT_READ_LOCAL_OOB_DATA_SIZE }, 9289 { add_remote_oob_data, MGMT_ADD_REMOTE_OOB_DATA_SIZE, 9290 HCI_MGMT_VAR_LEN }, 9291 { remove_remote_oob_data, MGMT_REMOVE_REMOTE_OOB_DATA_SIZE }, 9292 { start_discovery, MGMT_START_DISCOVERY_SIZE }, 9293 { stop_discovery, MGMT_STOP_DISCOVERY_SIZE }, 9294 { confirm_name, MGMT_CONFIRM_NAME_SIZE }, 9295 { block_device, MGMT_BLOCK_DEVICE_SIZE }, 9296 { unblock_device, MGMT_UNBLOCK_DEVICE_SIZE }, 9297 { set_device_id, MGMT_SET_DEVICE_ID_SIZE }, 9298 { set_advertising, MGMT_SETTING_SIZE }, 9299 { set_bredr, MGMT_SETTING_SIZE }, 9300 { set_static_address, MGMT_SET_STATIC_ADDRESS_SIZE }, 9301 { set_scan_params, MGMT_SET_SCAN_PARAMS_SIZE }, 9302 { set_secure_conn, MGMT_SETTING_SIZE }, 9303 { set_debug_keys, MGMT_SETTING_SIZE }, 9304 { set_privacy, MGMT_SET_PRIVACY_SIZE }, 9305 { load_irks, MGMT_LOAD_IRKS_SIZE, 9306 HCI_MGMT_VAR_LEN }, 9307 { get_conn_info, MGMT_GET_CONN_INFO_SIZE }, 9308 { get_clock_info, MGMT_GET_CLOCK_INFO_SIZE }, 9309 { add_device, MGMT_ADD_DEVICE_SIZE }, 9310 { remove_device, MGMT_REMOVE_DEVICE_SIZE }, 9311 { load_conn_param, MGMT_LOAD_CONN_PARAM_SIZE, 9312 HCI_MGMT_VAR_LEN }, 9313 { read_unconf_index_list, MGMT_READ_UNCONF_INDEX_LIST_SIZE, 9314 HCI_MGMT_NO_HDEV | 9315 HCI_MGMT_UNTRUSTED }, 9316 { read_config_info, MGMT_READ_CONFIG_INFO_SIZE, 9317 HCI_MGMT_UNCONFIGURED | 9318 HCI_MGMT_UNTRUSTED }, 9319 { set_external_config, MGMT_SET_EXTERNAL_CONFIG_SIZE, 9320 HCI_MGMT_UNCONFIGURED }, 9321 { set_public_address, MGMT_SET_PUBLIC_ADDRESS_SIZE, 9322 HCI_MGMT_UNCONFIGURED }, 9323 { start_service_discovery, MGMT_START_SERVICE_DISCOVERY_SIZE, 9324 HCI_MGMT_VAR_LEN }, 9325 { read_local_oob_ext_data, MGMT_READ_LOCAL_OOB_EXT_DATA_SIZE }, 9326 { read_ext_index_list, MGMT_READ_EXT_INDEX_LIST_SIZE, 9327 HCI_MGMT_NO_HDEV | 9328 HCI_MGMT_UNTRUSTED }, 9329 { read_adv_features, MGMT_READ_ADV_FEATURES_SIZE }, 9330 { add_advertising, MGMT_ADD_ADVERTISING_SIZE, 9331 HCI_MGMT_VAR_LEN }, 9332 { remove_advertising, MGMT_REMOVE_ADVERTISING_SIZE }, 9333 { get_adv_size_info, MGMT_GET_ADV_SIZE_INFO_SIZE }, 9334 { start_limited_discovery, MGMT_START_DISCOVERY_SIZE }, 9335 { read_ext_controller_info,MGMT_READ_EXT_INFO_SIZE, 9336 HCI_MGMT_UNTRUSTED }, 9337 { set_appearance, MGMT_SET_APPEARANCE_SIZE }, 9338 { get_phy_configuration, MGMT_GET_PHY_CONFIGURATION_SIZE }, 9339 { set_phy_configuration, MGMT_SET_PHY_CONFIGURATION_SIZE }, 9340 { set_blocked_keys, MGMT_OP_SET_BLOCKED_KEYS_SIZE, 9341 HCI_MGMT_VAR_LEN }, 9342 { set_wideband_speech, MGMT_SETTING_SIZE }, 9343 { read_controller_cap, MGMT_READ_CONTROLLER_CAP_SIZE, 9344 HCI_MGMT_UNTRUSTED }, 9345 { read_exp_features_info, MGMT_READ_EXP_FEATURES_INFO_SIZE, 9346 HCI_MGMT_UNTRUSTED | 9347 HCI_MGMT_HDEV_OPTIONAL }, 9348 { set_exp_feature, MGMT_SET_EXP_FEATURE_SIZE, 9349 HCI_MGMT_VAR_LEN | 9350 HCI_MGMT_HDEV_OPTIONAL }, 9351 { read_def_system_config, MGMT_READ_DEF_SYSTEM_CONFIG_SIZE, 9352 HCI_MGMT_UNTRUSTED }, 9353 { set_def_system_config, MGMT_SET_DEF_SYSTEM_CONFIG_SIZE, 9354 HCI_MGMT_VAR_LEN }, 9355 { read_def_runtime_config, MGMT_READ_DEF_RUNTIME_CONFIG_SIZE, 9356 HCI_MGMT_UNTRUSTED }, 9357 { set_def_runtime_config, MGMT_SET_DEF_RUNTIME_CONFIG_SIZE, 9358 HCI_MGMT_VAR_LEN }, 9359 { get_device_flags, MGMT_GET_DEVICE_FLAGS_SIZE }, 9360 { set_device_flags, MGMT_SET_DEVICE_FLAGS_SIZE }, 9361 { read_adv_mon_features, MGMT_READ_ADV_MONITOR_FEATURES_SIZE }, 9362 { add_adv_patterns_monitor,MGMT_ADD_ADV_PATTERNS_MONITOR_SIZE, 9363 HCI_MGMT_VAR_LEN }, 9364 { remove_adv_monitor, MGMT_REMOVE_ADV_MONITOR_SIZE }, 9365 { add_ext_adv_params, MGMT_ADD_EXT_ADV_PARAMS_MIN_SIZE, 9366 HCI_MGMT_VAR_LEN }, 9367 { add_ext_adv_data, MGMT_ADD_EXT_ADV_DATA_SIZE, 9368 HCI_MGMT_VAR_LEN }, 9369 { add_adv_patterns_monitor_rssi, 9370 MGMT_ADD_ADV_PATTERNS_MONITOR_RSSI_SIZE, 9371 HCI_MGMT_VAR_LEN }, 9372 { set_mesh, MGMT_SET_MESH_RECEIVER_SIZE, 9373 HCI_MGMT_VAR_LEN }, 9374 { mesh_features, MGMT_MESH_READ_FEATURES_SIZE }, 9375 { mesh_send, MGMT_MESH_SEND_SIZE, 9376 HCI_MGMT_VAR_LEN }, 9377 { mesh_send_cancel, MGMT_MESH_SEND_CANCEL_SIZE }, 9378 }; 9379 9380 void mgmt_index_added(struct hci_dev *hdev) 9381 { 9382 struct mgmt_ev_ext_index ev; 9383 9384 if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks)) 9385 return; 9386 9387 switch (hdev->dev_type) { 9388 case HCI_PRIMARY: 9389 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) { 9390 mgmt_index_event(MGMT_EV_UNCONF_INDEX_ADDED, hdev, 9391 NULL, 0, HCI_MGMT_UNCONF_INDEX_EVENTS); 9392 ev.type = 0x01; 9393 } else { 9394 mgmt_index_event(MGMT_EV_INDEX_ADDED, hdev, NULL, 0, 9395 HCI_MGMT_INDEX_EVENTS); 9396 ev.type = 0x00; 9397 } 9398 break; 9399 case HCI_AMP: 9400 ev.type = 0x02; 9401 break; 9402 default: 9403 return; 9404 } 9405 9406 ev.bus = hdev->bus; 9407 9408 mgmt_index_event(MGMT_EV_EXT_INDEX_ADDED, hdev, &ev, sizeof(ev), 9409 HCI_MGMT_EXT_INDEX_EVENTS); 9410 } 9411 9412 void mgmt_index_removed(struct hci_dev *hdev) 9413 { 9414 struct mgmt_ev_ext_index ev; 9415 u8 status = MGMT_STATUS_INVALID_INDEX; 9416 9417 if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks)) 9418 return; 9419 9420 switch (hdev->dev_type) { 9421 case HCI_PRIMARY: 9422 mgmt_pending_foreach(0, hdev, cmd_complete_rsp, &status); 9423 9424 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) { 9425 mgmt_index_event(MGMT_EV_UNCONF_INDEX_REMOVED, hdev, 9426 NULL, 0, HCI_MGMT_UNCONF_INDEX_EVENTS); 9427 ev.type = 0x01; 9428 } else { 9429 mgmt_index_event(MGMT_EV_INDEX_REMOVED, hdev, NULL, 0, 9430 HCI_MGMT_INDEX_EVENTS); 9431 ev.type = 0x00; 9432 } 9433 break; 9434 case HCI_AMP: 9435 ev.type = 0x02; 9436 break; 9437 default: 9438 return; 9439 } 9440 9441 ev.bus = hdev->bus; 9442 9443 mgmt_index_event(MGMT_EV_EXT_INDEX_REMOVED, hdev, &ev, sizeof(ev), 9444 HCI_MGMT_EXT_INDEX_EVENTS); 9445 9446 /* Cancel any remaining timed work */ 9447 if (!hci_dev_test_flag(hdev, HCI_MGMT)) 9448 return; 9449 cancel_delayed_work_sync(&hdev->discov_off); 9450 cancel_delayed_work_sync(&hdev->service_cache); 9451 cancel_delayed_work_sync(&hdev->rpa_expired); 9452 } 9453 9454 void mgmt_power_on(struct hci_dev *hdev, int err) 9455 { 9456 struct cmd_lookup match = { NULL, hdev }; 9457 9458 bt_dev_dbg(hdev, "err %d", err); 9459 9460 hci_dev_lock(hdev); 9461 9462 if (!err) { 9463 restart_le_actions(hdev); 9464 hci_update_passive_scan(hdev); 9465 } 9466 9467 mgmt_pending_foreach(MGMT_OP_SET_POWERED, hdev, settings_rsp, &match); 9468 9469 new_settings(hdev, match.sk); 9470 9471 if (match.sk) 9472 sock_put(match.sk); 9473 9474 hci_dev_unlock(hdev); 9475 } 9476 9477 void __mgmt_power_off(struct hci_dev *hdev) 9478 { 9479 struct cmd_lookup match = { NULL, hdev }; 9480 u8 status, zero_cod[] = { 0, 0, 0 }; 9481 9482 mgmt_pending_foreach(MGMT_OP_SET_POWERED, hdev, settings_rsp, &match); 9483 9484 /* If the power off is because of hdev unregistration let 9485 * use the appropriate INVALID_INDEX status. Otherwise use 9486 * NOT_POWERED. We cover both scenarios here since later in 9487 * mgmt_index_removed() any hci_conn callbacks will have already 9488 * been triggered, potentially causing misleading DISCONNECTED 9489 * status responses. 9490 */ 9491 if (hci_dev_test_flag(hdev, HCI_UNREGISTER)) 9492 status = MGMT_STATUS_INVALID_INDEX; 9493 else 9494 status = MGMT_STATUS_NOT_POWERED; 9495 9496 mgmt_pending_foreach(0, hdev, cmd_complete_rsp, &status); 9497 9498 if (memcmp(hdev->dev_class, zero_cod, sizeof(zero_cod)) != 0) { 9499 mgmt_limited_event(MGMT_EV_CLASS_OF_DEV_CHANGED, hdev, 9500 zero_cod, sizeof(zero_cod), 9501 HCI_MGMT_DEV_CLASS_EVENTS, NULL); 9502 ext_info_changed(hdev, NULL); 9503 } 9504 9505 new_settings(hdev, match.sk); 9506 9507 if (match.sk) 9508 sock_put(match.sk); 9509 } 9510 9511 void mgmt_set_powered_failed(struct hci_dev *hdev, int err) 9512 { 9513 struct mgmt_pending_cmd *cmd; 9514 u8 status; 9515 9516 cmd = pending_find(MGMT_OP_SET_POWERED, hdev); 9517 if (!cmd) 9518 return; 9519 9520 if (err == -ERFKILL) 9521 status = MGMT_STATUS_RFKILLED; 9522 else 9523 status = MGMT_STATUS_FAILED; 9524 9525 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_POWERED, status); 9526 9527 mgmt_pending_remove(cmd); 9528 } 9529 9530 void mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key, 9531 bool persistent) 9532 { 9533 struct mgmt_ev_new_link_key ev; 9534 9535 memset(&ev, 0, sizeof(ev)); 9536 9537 ev.store_hint = persistent; 9538 bacpy(&ev.key.addr.bdaddr, &key->bdaddr); 9539 ev.key.addr.type = link_to_bdaddr(key->link_type, key->bdaddr_type); 9540 ev.key.type = key->type; 9541 memcpy(ev.key.val, key->val, HCI_LINK_KEY_SIZE); 9542 ev.key.pin_len = key->pin_len; 9543 9544 mgmt_event(MGMT_EV_NEW_LINK_KEY, hdev, &ev, sizeof(ev), NULL); 9545 } 9546 9547 static u8 mgmt_ltk_type(struct smp_ltk *ltk) 9548 { 9549 switch (ltk->type) { 9550 case SMP_LTK: 9551 case SMP_LTK_RESPONDER: 9552 if (ltk->authenticated) 9553 return MGMT_LTK_AUTHENTICATED; 9554 return MGMT_LTK_UNAUTHENTICATED; 9555 case SMP_LTK_P256: 9556 if (ltk->authenticated) 9557 return MGMT_LTK_P256_AUTH; 9558 return MGMT_LTK_P256_UNAUTH; 9559 case SMP_LTK_P256_DEBUG: 9560 return MGMT_LTK_P256_DEBUG; 9561 } 9562 9563 return MGMT_LTK_UNAUTHENTICATED; 9564 } 9565 9566 void mgmt_new_ltk(struct hci_dev *hdev, struct smp_ltk *key, bool persistent) 9567 { 9568 struct mgmt_ev_new_long_term_key ev; 9569 9570 memset(&ev, 0, sizeof(ev)); 9571 9572 /* Devices using resolvable or non-resolvable random addresses 9573 * without providing an identity resolving key don't require 9574 * to store long term keys. Their addresses will change the 9575 * next time around. 9576 * 9577 * Only when a remote device provides an identity address 9578 * make sure the long term key is stored. If the remote 9579 * identity is known, the long term keys are internally 9580 * mapped to the identity address. So allow static random 9581 * and public addresses here. 9582 */ 9583 if (key->bdaddr_type == ADDR_LE_DEV_RANDOM && 9584 (key->bdaddr.b[5] & 0xc0) != 0xc0) 9585 ev.store_hint = 0x00; 9586 else 9587 ev.store_hint = persistent; 9588 9589 bacpy(&ev.key.addr.bdaddr, &key->bdaddr); 9590 ev.key.addr.type = link_to_bdaddr(key->link_type, key->bdaddr_type); 9591 ev.key.type = mgmt_ltk_type(key); 9592 ev.key.enc_size = key->enc_size; 9593 ev.key.ediv = key->ediv; 9594 ev.key.rand = key->rand; 9595 9596 if (key->type == SMP_LTK) 9597 ev.key.initiator = 1; 9598 9599 /* Make sure we copy only the significant bytes based on the 9600 * encryption key size, and set the rest of the value to zeroes. 9601 */ 9602 memcpy(ev.key.val, key->val, key->enc_size); 9603 memset(ev.key.val + key->enc_size, 0, 9604 sizeof(ev.key.val) - key->enc_size); 9605 9606 mgmt_event(MGMT_EV_NEW_LONG_TERM_KEY, hdev, &ev, sizeof(ev), NULL); 9607 } 9608 9609 void mgmt_new_irk(struct hci_dev *hdev, struct smp_irk *irk, bool persistent) 9610 { 9611 struct mgmt_ev_new_irk ev; 9612 9613 memset(&ev, 0, sizeof(ev)); 9614 9615 ev.store_hint = persistent; 9616 9617 bacpy(&ev.rpa, &irk->rpa); 9618 bacpy(&ev.irk.addr.bdaddr, &irk->bdaddr); 9619 ev.irk.addr.type = link_to_bdaddr(irk->link_type, irk->addr_type); 9620 memcpy(ev.irk.val, irk->val, sizeof(irk->val)); 9621 9622 mgmt_event(MGMT_EV_NEW_IRK, hdev, &ev, sizeof(ev), NULL); 9623 } 9624 9625 void mgmt_new_csrk(struct hci_dev *hdev, struct smp_csrk *csrk, 9626 bool persistent) 9627 { 9628 struct mgmt_ev_new_csrk ev; 9629 9630 memset(&ev, 0, sizeof(ev)); 9631 9632 /* Devices using resolvable or non-resolvable random addresses 9633 * without providing an identity resolving key don't require 9634 * to store signature resolving keys. Their addresses will change 9635 * the next time around. 9636 * 9637 * Only when a remote device provides an identity address 9638 * make sure the signature resolving key is stored. So allow 9639 * static random and public addresses here. 9640 */ 9641 if (csrk->bdaddr_type == ADDR_LE_DEV_RANDOM && 9642 (csrk->bdaddr.b[5] & 0xc0) != 0xc0) 9643 ev.store_hint = 0x00; 9644 else 9645 ev.store_hint = persistent; 9646 9647 bacpy(&ev.key.addr.bdaddr, &csrk->bdaddr); 9648 ev.key.addr.type = link_to_bdaddr(csrk->link_type, csrk->bdaddr_type); 9649 ev.key.type = csrk->type; 9650 memcpy(ev.key.val, csrk->val, sizeof(csrk->val)); 9651 9652 mgmt_event(MGMT_EV_NEW_CSRK, hdev, &ev, sizeof(ev), NULL); 9653 } 9654 9655 void mgmt_new_conn_param(struct hci_dev *hdev, bdaddr_t *bdaddr, 9656 u8 bdaddr_type, u8 store_hint, u16 min_interval, 9657 u16 max_interval, u16 latency, u16 timeout) 9658 { 9659 struct mgmt_ev_new_conn_param ev; 9660 9661 if (!hci_is_identity_address(bdaddr, bdaddr_type)) 9662 return; 9663 9664 memset(&ev, 0, sizeof(ev)); 9665 bacpy(&ev.addr.bdaddr, bdaddr); 9666 ev.addr.type = link_to_bdaddr(LE_LINK, bdaddr_type); 9667 ev.store_hint = store_hint; 9668 ev.min_interval = cpu_to_le16(min_interval); 9669 ev.max_interval = cpu_to_le16(max_interval); 9670 ev.latency = cpu_to_le16(latency); 9671 ev.timeout = cpu_to_le16(timeout); 9672 9673 mgmt_event(MGMT_EV_NEW_CONN_PARAM, hdev, &ev, sizeof(ev), NULL); 9674 } 9675 9676 void mgmt_device_connected(struct hci_dev *hdev, struct hci_conn *conn, 9677 u8 *name, u8 name_len) 9678 { 9679 struct sk_buff *skb; 9680 struct mgmt_ev_device_connected *ev; 9681 u16 eir_len = 0; 9682 u32 flags = 0; 9683 9684 /* allocate buff for LE or BR/EDR adv */ 9685 if (conn->le_adv_data_len > 0) 9686 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_CONNECTED, 9687 sizeof(*ev) + conn->le_adv_data_len); 9688 else 9689 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_CONNECTED, 9690 sizeof(*ev) + (name ? eir_precalc_len(name_len) : 0) + 9691 eir_precalc_len(sizeof(conn->dev_class))); 9692 9693 ev = skb_put(skb, sizeof(*ev)); 9694 bacpy(&ev->addr.bdaddr, &conn->dst); 9695 ev->addr.type = link_to_bdaddr(conn->type, conn->dst_type); 9696 9697 if (conn->out) 9698 flags |= MGMT_DEV_FOUND_INITIATED_CONN; 9699 9700 ev->flags = __cpu_to_le32(flags); 9701 9702 /* We must ensure that the EIR Data fields are ordered and 9703 * unique. Keep it simple for now and avoid the problem by not 9704 * adding any BR/EDR data to the LE adv. 9705 */ 9706 if (conn->le_adv_data_len > 0) { 9707 skb_put_data(skb, conn->le_adv_data, conn->le_adv_data_len); 9708 eir_len = conn->le_adv_data_len; 9709 } else { 9710 if (name) 9711 eir_len += eir_skb_put_data(skb, EIR_NAME_COMPLETE, name, name_len); 9712 9713 if (memcmp(conn->dev_class, "\0\0\0", sizeof(conn->dev_class))) 9714 eir_len += eir_skb_put_data(skb, EIR_CLASS_OF_DEV, 9715 conn->dev_class, sizeof(conn->dev_class)); 9716 } 9717 9718 ev->eir_len = cpu_to_le16(eir_len); 9719 9720 mgmt_event_skb(skb, NULL); 9721 } 9722 9723 static void disconnect_rsp(struct mgmt_pending_cmd *cmd, void *data) 9724 { 9725 struct sock **sk = data; 9726 9727 cmd->cmd_complete(cmd, 0); 9728 9729 *sk = cmd->sk; 9730 sock_hold(*sk); 9731 9732 mgmt_pending_remove(cmd); 9733 } 9734 9735 static void unpair_device_rsp(struct mgmt_pending_cmd *cmd, void *data) 9736 { 9737 struct hci_dev *hdev = data; 9738 struct mgmt_cp_unpair_device *cp = cmd->param; 9739 9740 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, cmd->sk); 9741 9742 cmd->cmd_complete(cmd, 0); 9743 mgmt_pending_remove(cmd); 9744 } 9745 9746 bool mgmt_powering_down(struct hci_dev *hdev) 9747 { 9748 struct mgmt_pending_cmd *cmd; 9749 struct mgmt_mode *cp; 9750 9751 cmd = pending_find(MGMT_OP_SET_POWERED, hdev); 9752 if (!cmd) 9753 return false; 9754 9755 cp = cmd->param; 9756 if (!cp->val) 9757 return true; 9758 9759 return false; 9760 } 9761 9762 void mgmt_device_disconnected(struct hci_dev *hdev, bdaddr_t *bdaddr, 9763 u8 link_type, u8 addr_type, u8 reason, 9764 bool mgmt_connected) 9765 { 9766 struct mgmt_ev_device_disconnected ev; 9767 struct sock *sk = NULL; 9768 9769 /* The connection is still in hci_conn_hash so test for 1 9770 * instead of 0 to know if this is the last one. 9771 */ 9772 if (mgmt_powering_down(hdev) && hci_conn_count(hdev) == 1) { 9773 cancel_delayed_work(&hdev->power_off); 9774 queue_work(hdev->req_workqueue, &hdev->power_off.work); 9775 } 9776 9777 if (!mgmt_connected) 9778 return; 9779 9780 if (link_type != ACL_LINK && link_type != LE_LINK) 9781 return; 9782 9783 mgmt_pending_foreach(MGMT_OP_DISCONNECT, hdev, disconnect_rsp, &sk); 9784 9785 bacpy(&ev.addr.bdaddr, bdaddr); 9786 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9787 ev.reason = reason; 9788 9789 /* Report disconnects due to suspend */ 9790 if (hdev->suspended) 9791 ev.reason = MGMT_DEV_DISCONN_LOCAL_HOST_SUSPEND; 9792 9793 mgmt_event(MGMT_EV_DEVICE_DISCONNECTED, hdev, &ev, sizeof(ev), sk); 9794 9795 if (sk) 9796 sock_put(sk); 9797 9798 mgmt_pending_foreach(MGMT_OP_UNPAIR_DEVICE, hdev, unpair_device_rsp, 9799 hdev); 9800 } 9801 9802 void mgmt_disconnect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, 9803 u8 link_type, u8 addr_type, u8 status) 9804 { 9805 u8 bdaddr_type = link_to_bdaddr(link_type, addr_type); 9806 struct mgmt_cp_disconnect *cp; 9807 struct mgmt_pending_cmd *cmd; 9808 9809 mgmt_pending_foreach(MGMT_OP_UNPAIR_DEVICE, hdev, unpair_device_rsp, 9810 hdev); 9811 9812 cmd = pending_find(MGMT_OP_DISCONNECT, hdev); 9813 if (!cmd) 9814 return; 9815 9816 cp = cmd->param; 9817 9818 if (bacmp(bdaddr, &cp->addr.bdaddr)) 9819 return; 9820 9821 if (cp->addr.type != bdaddr_type) 9822 return; 9823 9824 cmd->cmd_complete(cmd, mgmt_status(status)); 9825 mgmt_pending_remove(cmd); 9826 } 9827 9828 void mgmt_connect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 9829 u8 addr_type, u8 status) 9830 { 9831 struct mgmt_ev_connect_failed ev; 9832 9833 /* The connection is still in hci_conn_hash so test for 1 9834 * instead of 0 to know if this is the last one. 9835 */ 9836 if (mgmt_powering_down(hdev) && hci_conn_count(hdev) == 1) { 9837 cancel_delayed_work(&hdev->power_off); 9838 queue_work(hdev->req_workqueue, &hdev->power_off.work); 9839 } 9840 9841 bacpy(&ev.addr.bdaddr, bdaddr); 9842 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9843 ev.status = mgmt_status(status); 9844 9845 mgmt_event(MGMT_EV_CONNECT_FAILED, hdev, &ev, sizeof(ev), NULL); 9846 } 9847 9848 void mgmt_pin_code_request(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 secure) 9849 { 9850 struct mgmt_ev_pin_code_request ev; 9851 9852 bacpy(&ev.addr.bdaddr, bdaddr); 9853 ev.addr.type = BDADDR_BREDR; 9854 ev.secure = secure; 9855 9856 mgmt_event(MGMT_EV_PIN_CODE_REQUEST, hdev, &ev, sizeof(ev), NULL); 9857 } 9858 9859 void mgmt_pin_code_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9860 u8 status) 9861 { 9862 struct mgmt_pending_cmd *cmd; 9863 9864 cmd = pending_find(MGMT_OP_PIN_CODE_REPLY, hdev); 9865 if (!cmd) 9866 return; 9867 9868 cmd->cmd_complete(cmd, mgmt_status(status)); 9869 mgmt_pending_remove(cmd); 9870 } 9871 9872 void mgmt_pin_code_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9873 u8 status) 9874 { 9875 struct mgmt_pending_cmd *cmd; 9876 9877 cmd = pending_find(MGMT_OP_PIN_CODE_NEG_REPLY, hdev); 9878 if (!cmd) 9879 return; 9880 9881 cmd->cmd_complete(cmd, mgmt_status(status)); 9882 mgmt_pending_remove(cmd); 9883 } 9884 9885 int mgmt_user_confirm_request(struct hci_dev *hdev, bdaddr_t *bdaddr, 9886 u8 link_type, u8 addr_type, u32 value, 9887 u8 confirm_hint) 9888 { 9889 struct mgmt_ev_user_confirm_request ev; 9890 9891 bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr); 9892 9893 bacpy(&ev.addr.bdaddr, bdaddr); 9894 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9895 ev.confirm_hint = confirm_hint; 9896 ev.value = cpu_to_le32(value); 9897 9898 return mgmt_event(MGMT_EV_USER_CONFIRM_REQUEST, hdev, &ev, sizeof(ev), 9899 NULL); 9900 } 9901 9902 int mgmt_user_passkey_request(struct hci_dev *hdev, bdaddr_t *bdaddr, 9903 u8 link_type, u8 addr_type) 9904 { 9905 struct mgmt_ev_user_passkey_request ev; 9906 9907 bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr); 9908 9909 bacpy(&ev.addr.bdaddr, bdaddr); 9910 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9911 9912 return mgmt_event(MGMT_EV_USER_PASSKEY_REQUEST, hdev, &ev, sizeof(ev), 9913 NULL); 9914 } 9915 9916 static int user_pairing_resp_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9917 u8 link_type, u8 addr_type, u8 status, 9918 u8 opcode) 9919 { 9920 struct mgmt_pending_cmd *cmd; 9921 9922 cmd = pending_find(opcode, hdev); 9923 if (!cmd) 9924 return -ENOENT; 9925 9926 cmd->cmd_complete(cmd, mgmt_status(status)); 9927 mgmt_pending_remove(cmd); 9928 9929 return 0; 9930 } 9931 9932 int mgmt_user_confirm_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9933 u8 link_type, u8 addr_type, u8 status) 9934 { 9935 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9936 status, MGMT_OP_USER_CONFIRM_REPLY); 9937 } 9938 9939 int mgmt_user_confirm_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9940 u8 link_type, u8 addr_type, u8 status) 9941 { 9942 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9943 status, 9944 MGMT_OP_USER_CONFIRM_NEG_REPLY); 9945 } 9946 9947 int mgmt_user_passkey_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9948 u8 link_type, u8 addr_type, u8 status) 9949 { 9950 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9951 status, MGMT_OP_USER_PASSKEY_REPLY); 9952 } 9953 9954 int mgmt_user_passkey_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr, 9955 u8 link_type, u8 addr_type, u8 status) 9956 { 9957 return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type, 9958 status, 9959 MGMT_OP_USER_PASSKEY_NEG_REPLY); 9960 } 9961 9962 int mgmt_user_passkey_notify(struct hci_dev *hdev, bdaddr_t *bdaddr, 9963 u8 link_type, u8 addr_type, u32 passkey, 9964 u8 entered) 9965 { 9966 struct mgmt_ev_passkey_notify ev; 9967 9968 bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr); 9969 9970 bacpy(&ev.addr.bdaddr, bdaddr); 9971 ev.addr.type = link_to_bdaddr(link_type, addr_type); 9972 ev.passkey = __cpu_to_le32(passkey); 9973 ev.entered = entered; 9974 9975 return mgmt_event(MGMT_EV_PASSKEY_NOTIFY, hdev, &ev, sizeof(ev), NULL); 9976 } 9977 9978 void mgmt_auth_failed(struct hci_conn *conn, u8 hci_status) 9979 { 9980 struct mgmt_ev_auth_failed ev; 9981 struct mgmt_pending_cmd *cmd; 9982 u8 status = mgmt_status(hci_status); 9983 9984 bacpy(&ev.addr.bdaddr, &conn->dst); 9985 ev.addr.type = link_to_bdaddr(conn->type, conn->dst_type); 9986 ev.status = status; 9987 9988 cmd = find_pairing(conn); 9989 9990 mgmt_event(MGMT_EV_AUTH_FAILED, conn->hdev, &ev, sizeof(ev), 9991 cmd ? cmd->sk : NULL); 9992 9993 if (cmd) { 9994 cmd->cmd_complete(cmd, status); 9995 mgmt_pending_remove(cmd); 9996 } 9997 } 9998 9999 void mgmt_auth_enable_complete(struct hci_dev *hdev, u8 status) 10000 { 10001 struct cmd_lookup match = { NULL, hdev }; 10002 bool changed; 10003 10004 if (status) { 10005 u8 mgmt_err = mgmt_status(status); 10006 mgmt_pending_foreach(MGMT_OP_SET_LINK_SECURITY, hdev, 10007 cmd_status_rsp, &mgmt_err); 10008 return; 10009 } 10010 10011 if (test_bit(HCI_AUTH, &hdev->flags)) 10012 changed = !hci_dev_test_and_set_flag(hdev, HCI_LINK_SECURITY); 10013 else 10014 changed = hci_dev_test_and_clear_flag(hdev, HCI_LINK_SECURITY); 10015 10016 mgmt_pending_foreach(MGMT_OP_SET_LINK_SECURITY, hdev, settings_rsp, 10017 &match); 10018 10019 if (changed) 10020 new_settings(hdev, match.sk); 10021 10022 if (match.sk) 10023 sock_put(match.sk); 10024 } 10025 10026 static void sk_lookup(struct mgmt_pending_cmd *cmd, void *data) 10027 { 10028 struct cmd_lookup *match = data; 10029 10030 if (match->sk == NULL) { 10031 match->sk = cmd->sk; 10032 sock_hold(match->sk); 10033 } 10034 } 10035 10036 void mgmt_set_class_of_dev_complete(struct hci_dev *hdev, u8 *dev_class, 10037 u8 status) 10038 { 10039 struct cmd_lookup match = { NULL, hdev, mgmt_status(status) }; 10040 10041 mgmt_pending_foreach(MGMT_OP_SET_DEV_CLASS, hdev, sk_lookup, &match); 10042 mgmt_pending_foreach(MGMT_OP_ADD_UUID, hdev, sk_lookup, &match); 10043 mgmt_pending_foreach(MGMT_OP_REMOVE_UUID, hdev, sk_lookup, &match); 10044 10045 if (!status) { 10046 mgmt_limited_event(MGMT_EV_CLASS_OF_DEV_CHANGED, hdev, dev_class, 10047 3, HCI_MGMT_DEV_CLASS_EVENTS, NULL); 10048 ext_info_changed(hdev, NULL); 10049 } 10050 10051 if (match.sk) 10052 sock_put(match.sk); 10053 } 10054 10055 void mgmt_set_local_name_complete(struct hci_dev *hdev, u8 *name, u8 status) 10056 { 10057 struct mgmt_cp_set_local_name ev; 10058 struct mgmt_pending_cmd *cmd; 10059 10060 if (status) 10061 return; 10062 10063 memset(&ev, 0, sizeof(ev)); 10064 memcpy(ev.name, name, HCI_MAX_NAME_LENGTH); 10065 memcpy(ev.short_name, hdev->short_name, HCI_MAX_SHORT_NAME_LENGTH); 10066 10067 cmd = pending_find(MGMT_OP_SET_LOCAL_NAME, hdev); 10068 if (!cmd) { 10069 memcpy(hdev->dev_name, name, sizeof(hdev->dev_name)); 10070 10071 /* If this is a HCI command related to powering on the 10072 * HCI dev don't send any mgmt signals. 10073 */ 10074 if (pending_find(MGMT_OP_SET_POWERED, hdev)) 10075 return; 10076 } 10077 10078 mgmt_limited_event(MGMT_EV_LOCAL_NAME_CHANGED, hdev, &ev, sizeof(ev), 10079 HCI_MGMT_LOCAL_NAME_EVENTS, cmd ? cmd->sk : NULL); 10080 ext_info_changed(hdev, cmd ? cmd->sk : NULL); 10081 } 10082 10083 static inline bool has_uuid(u8 *uuid, u16 uuid_count, u8 (*uuids)[16]) 10084 { 10085 int i; 10086 10087 for (i = 0; i < uuid_count; i++) { 10088 if (!memcmp(uuid, uuids[i], 16)) 10089 return true; 10090 } 10091 10092 return false; 10093 } 10094 10095 static bool eir_has_uuids(u8 *eir, u16 eir_len, u16 uuid_count, u8 (*uuids)[16]) 10096 { 10097 u16 parsed = 0; 10098 10099 while (parsed < eir_len) { 10100 u8 field_len = eir[0]; 10101 u8 uuid[16]; 10102 int i; 10103 10104 if (field_len == 0) 10105 break; 10106 10107 if (eir_len - parsed < field_len + 1) 10108 break; 10109 10110 switch (eir[1]) { 10111 case EIR_UUID16_ALL: 10112 case EIR_UUID16_SOME: 10113 for (i = 0; i + 3 <= field_len; i += 2) { 10114 memcpy(uuid, bluetooth_base_uuid, 16); 10115 uuid[13] = eir[i + 3]; 10116 uuid[12] = eir[i + 2]; 10117 if (has_uuid(uuid, uuid_count, uuids)) 10118 return true; 10119 } 10120 break; 10121 case EIR_UUID32_ALL: 10122 case EIR_UUID32_SOME: 10123 for (i = 0; i + 5 <= field_len; i += 4) { 10124 memcpy(uuid, bluetooth_base_uuid, 16); 10125 uuid[15] = eir[i + 5]; 10126 uuid[14] = eir[i + 4]; 10127 uuid[13] = eir[i + 3]; 10128 uuid[12] = eir[i + 2]; 10129 if (has_uuid(uuid, uuid_count, uuids)) 10130 return true; 10131 } 10132 break; 10133 case EIR_UUID128_ALL: 10134 case EIR_UUID128_SOME: 10135 for (i = 0; i + 17 <= field_len; i += 16) { 10136 memcpy(uuid, eir + i + 2, 16); 10137 if (has_uuid(uuid, uuid_count, uuids)) 10138 return true; 10139 } 10140 break; 10141 } 10142 10143 parsed += field_len + 1; 10144 eir += field_len + 1; 10145 } 10146 10147 return false; 10148 } 10149 10150 static bool is_filter_match(struct hci_dev *hdev, s8 rssi, u8 *eir, 10151 u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len) 10152 { 10153 /* If a RSSI threshold has been specified, and 10154 * HCI_QUIRK_STRICT_DUPLICATE_FILTER is not set, then all results with 10155 * a RSSI smaller than the RSSI threshold will be dropped. If the quirk 10156 * is set, let it through for further processing, as we might need to 10157 * restart the scan. 10158 * 10159 * For BR/EDR devices (pre 1.2) providing no RSSI during inquiry, 10160 * the results are also dropped. 10161 */ 10162 if (hdev->discovery.rssi != HCI_RSSI_INVALID && 10163 (rssi == HCI_RSSI_INVALID || 10164 (rssi < hdev->discovery.rssi && 10165 !test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks)))) 10166 return false; 10167 10168 if (hdev->discovery.uuid_count != 0) { 10169 /* If a list of UUIDs is provided in filter, results with no 10170 * matching UUID should be dropped. 10171 */ 10172 if (!eir_has_uuids(eir, eir_len, hdev->discovery.uuid_count, 10173 hdev->discovery.uuids) && 10174 !eir_has_uuids(scan_rsp, scan_rsp_len, 10175 hdev->discovery.uuid_count, 10176 hdev->discovery.uuids)) 10177 return false; 10178 } 10179 10180 /* If duplicate filtering does not report RSSI changes, then restart 10181 * scanning to ensure updated result with updated RSSI values. 10182 */ 10183 if (test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks)) { 10184 /* Validate RSSI value against the RSSI threshold once more. */ 10185 if (hdev->discovery.rssi != HCI_RSSI_INVALID && 10186 rssi < hdev->discovery.rssi) 10187 return false; 10188 } 10189 10190 return true; 10191 } 10192 10193 void mgmt_adv_monitor_device_lost(struct hci_dev *hdev, u16 handle, 10194 bdaddr_t *bdaddr, u8 addr_type) 10195 { 10196 struct mgmt_ev_adv_monitor_device_lost ev; 10197 10198 ev.monitor_handle = cpu_to_le16(handle); 10199 bacpy(&ev.addr.bdaddr, bdaddr); 10200 ev.addr.type = addr_type; 10201 10202 mgmt_event(MGMT_EV_ADV_MONITOR_DEVICE_LOST, hdev, &ev, sizeof(ev), 10203 NULL); 10204 } 10205 10206 static void mgmt_send_adv_monitor_device_found(struct hci_dev *hdev, 10207 struct sk_buff *skb, 10208 struct sock *skip_sk, 10209 u16 handle) 10210 { 10211 struct sk_buff *advmon_skb; 10212 size_t advmon_skb_len; 10213 __le16 *monitor_handle; 10214 10215 if (!skb) 10216 return; 10217 10218 advmon_skb_len = (sizeof(struct mgmt_ev_adv_monitor_device_found) - 10219 sizeof(struct mgmt_ev_device_found)) + skb->len; 10220 advmon_skb = mgmt_alloc_skb(hdev, MGMT_EV_ADV_MONITOR_DEVICE_FOUND, 10221 advmon_skb_len); 10222 if (!advmon_skb) 10223 return; 10224 10225 /* ADV_MONITOR_DEVICE_FOUND is similar to DEVICE_FOUND event except 10226 * that it also has 'monitor_handle'. Make a copy of DEVICE_FOUND and 10227 * store monitor_handle of the matched monitor. 10228 */ 10229 monitor_handle = skb_put(advmon_skb, sizeof(*monitor_handle)); 10230 *monitor_handle = cpu_to_le16(handle); 10231 skb_put_data(advmon_skb, skb->data, skb->len); 10232 10233 mgmt_event_skb(advmon_skb, skip_sk); 10234 } 10235 10236 static void mgmt_adv_monitor_device_found(struct hci_dev *hdev, 10237 bdaddr_t *bdaddr, bool report_device, 10238 struct sk_buff *skb, 10239 struct sock *skip_sk) 10240 { 10241 struct monitored_device *dev, *tmp; 10242 bool matched = false; 10243 bool notified = false; 10244 10245 /* We have received the Advertisement Report because: 10246 * 1. the kernel has initiated active discovery 10247 * 2. if not, we have pend_le_reports > 0 in which case we are doing 10248 * passive scanning 10249 * 3. if none of the above is true, we have one or more active 10250 * Advertisement Monitor 10251 * 10252 * For case 1 and 2, report all advertisements via MGMT_EV_DEVICE_FOUND 10253 * and report ONLY one advertisement per device for the matched Monitor 10254 * via MGMT_EV_ADV_MONITOR_DEVICE_FOUND event. 10255 * 10256 * For case 3, since we are not active scanning and all advertisements 10257 * received are due to a matched Advertisement Monitor, report all 10258 * advertisements ONLY via MGMT_EV_ADV_MONITOR_DEVICE_FOUND event. 10259 */ 10260 if (report_device && !hdev->advmon_pend_notify) { 10261 mgmt_event_skb(skb, skip_sk); 10262 return; 10263 } 10264 10265 hdev->advmon_pend_notify = false; 10266 10267 list_for_each_entry_safe(dev, tmp, &hdev->monitored_devices, list) { 10268 if (!bacmp(&dev->bdaddr, bdaddr)) { 10269 matched = true; 10270 10271 if (!dev->notified) { 10272 mgmt_send_adv_monitor_device_found(hdev, skb, 10273 skip_sk, 10274 dev->handle); 10275 notified = true; 10276 dev->notified = true; 10277 } 10278 } 10279 10280 if (!dev->notified) 10281 hdev->advmon_pend_notify = true; 10282 } 10283 10284 if (!report_device && 10285 ((matched && !notified) || !msft_monitor_supported(hdev))) { 10286 /* Handle 0 indicates that we are not active scanning and this 10287 * is a subsequent advertisement report for an already matched 10288 * Advertisement Monitor or the controller offloading support 10289 * is not available. 10290 */ 10291 mgmt_send_adv_monitor_device_found(hdev, skb, skip_sk, 0); 10292 } 10293 10294 if (report_device) 10295 mgmt_event_skb(skb, skip_sk); 10296 else 10297 kfree_skb(skb); 10298 } 10299 10300 static void mesh_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, 10301 u8 addr_type, s8 rssi, u32 flags, u8 *eir, 10302 u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len, 10303 u64 instant) 10304 { 10305 struct sk_buff *skb; 10306 struct mgmt_ev_mesh_device_found *ev; 10307 int i, j; 10308 10309 if (!hdev->mesh_ad_types[0]) 10310 goto accepted; 10311 10312 /* Scan for requested AD types */ 10313 if (eir_len > 0) { 10314 for (i = 0; i + 1 < eir_len; i += eir[i] + 1) { 10315 for (j = 0; j < sizeof(hdev->mesh_ad_types); j++) { 10316 if (!hdev->mesh_ad_types[j]) 10317 break; 10318 10319 if (hdev->mesh_ad_types[j] == eir[i + 1]) 10320 goto accepted; 10321 } 10322 } 10323 } 10324 10325 if (scan_rsp_len > 0) { 10326 for (i = 0; i + 1 < scan_rsp_len; i += scan_rsp[i] + 1) { 10327 for (j = 0; j < sizeof(hdev->mesh_ad_types); j++) { 10328 if (!hdev->mesh_ad_types[j]) 10329 break; 10330 10331 if (hdev->mesh_ad_types[j] == scan_rsp[i + 1]) 10332 goto accepted; 10333 } 10334 } 10335 } 10336 10337 return; 10338 10339 accepted: 10340 skb = mgmt_alloc_skb(hdev, MGMT_EV_MESH_DEVICE_FOUND, 10341 sizeof(*ev) + eir_len + scan_rsp_len); 10342 if (!skb) 10343 return; 10344 10345 ev = skb_put(skb, sizeof(*ev)); 10346 10347 bacpy(&ev->addr.bdaddr, bdaddr); 10348 ev->addr.type = link_to_bdaddr(LE_LINK, addr_type); 10349 ev->rssi = rssi; 10350 ev->flags = cpu_to_le32(flags); 10351 ev->instant = cpu_to_le64(instant); 10352 10353 if (eir_len > 0) 10354 /* Copy EIR or advertising data into event */ 10355 skb_put_data(skb, eir, eir_len); 10356 10357 if (scan_rsp_len > 0) 10358 /* Append scan response data to event */ 10359 skb_put_data(skb, scan_rsp, scan_rsp_len); 10360 10361 ev->eir_len = cpu_to_le16(eir_len + scan_rsp_len); 10362 10363 mgmt_event_skb(skb, NULL); 10364 } 10365 10366 void mgmt_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 10367 u8 addr_type, u8 *dev_class, s8 rssi, u32 flags, 10368 u8 *eir, u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len, 10369 u64 instant) 10370 { 10371 struct sk_buff *skb; 10372 struct mgmt_ev_device_found *ev; 10373 bool report_device = hci_discovery_active(hdev); 10374 10375 if (hci_dev_test_flag(hdev, HCI_MESH) && link_type == LE_LINK) 10376 mesh_device_found(hdev, bdaddr, addr_type, rssi, flags, 10377 eir, eir_len, scan_rsp, scan_rsp_len, 10378 instant); 10379 10380 /* Don't send events for a non-kernel initiated discovery. With 10381 * LE one exception is if we have pend_le_reports > 0 in which 10382 * case we're doing passive scanning and want these events. 10383 */ 10384 if (!hci_discovery_active(hdev)) { 10385 if (link_type == ACL_LINK) 10386 return; 10387 if (link_type == LE_LINK && !list_empty(&hdev->pend_le_reports)) 10388 report_device = true; 10389 else if (!hci_is_adv_monitoring(hdev)) 10390 return; 10391 } 10392 10393 if (hdev->discovery.result_filtering) { 10394 /* We are using service discovery */ 10395 if (!is_filter_match(hdev, rssi, eir, eir_len, scan_rsp, 10396 scan_rsp_len)) 10397 return; 10398 } 10399 10400 if (hdev->discovery.limited) { 10401 /* Check for limited discoverable bit */ 10402 if (dev_class) { 10403 if (!(dev_class[1] & 0x20)) 10404 return; 10405 } else { 10406 u8 *flags = eir_get_data(eir, eir_len, EIR_FLAGS, NULL); 10407 if (!flags || !(flags[0] & LE_AD_LIMITED)) 10408 return; 10409 } 10410 } 10411 10412 /* Allocate skb. The 5 extra bytes are for the potential CoD field */ 10413 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_FOUND, 10414 sizeof(*ev) + eir_len + scan_rsp_len + 5); 10415 if (!skb) 10416 return; 10417 10418 ev = skb_put(skb, sizeof(*ev)); 10419 10420 /* In case of device discovery with BR/EDR devices (pre 1.2), the 10421 * RSSI value was reported as 0 when not available. This behavior 10422 * is kept when using device discovery. This is required for full 10423 * backwards compatibility with the API. 10424 * 10425 * However when using service discovery, the value 127 will be 10426 * returned when the RSSI is not available. 10427 */ 10428 if (rssi == HCI_RSSI_INVALID && !hdev->discovery.report_invalid_rssi && 10429 link_type == ACL_LINK) 10430 rssi = 0; 10431 10432 bacpy(&ev->addr.bdaddr, bdaddr); 10433 ev->addr.type = link_to_bdaddr(link_type, addr_type); 10434 ev->rssi = rssi; 10435 ev->flags = cpu_to_le32(flags); 10436 10437 if (eir_len > 0) 10438 /* Copy EIR or advertising data into event */ 10439 skb_put_data(skb, eir, eir_len); 10440 10441 if (dev_class && !eir_get_data(eir, eir_len, EIR_CLASS_OF_DEV, NULL)) { 10442 u8 eir_cod[5]; 10443 10444 eir_len += eir_append_data(eir_cod, 0, EIR_CLASS_OF_DEV, 10445 dev_class, 3); 10446 skb_put_data(skb, eir_cod, sizeof(eir_cod)); 10447 } 10448 10449 if (scan_rsp_len > 0) 10450 /* Append scan response data to event */ 10451 skb_put_data(skb, scan_rsp, scan_rsp_len); 10452 10453 ev->eir_len = cpu_to_le16(eir_len + scan_rsp_len); 10454 10455 mgmt_adv_monitor_device_found(hdev, bdaddr, report_device, skb, NULL); 10456 } 10457 10458 void mgmt_remote_name(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type, 10459 u8 addr_type, s8 rssi, u8 *name, u8 name_len) 10460 { 10461 struct sk_buff *skb; 10462 struct mgmt_ev_device_found *ev; 10463 u16 eir_len = 0; 10464 u32 flags = 0; 10465 10466 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_FOUND, 10467 sizeof(*ev) + (name ? eir_precalc_len(name_len) : 0)); 10468 10469 ev = skb_put(skb, sizeof(*ev)); 10470 bacpy(&ev->addr.bdaddr, bdaddr); 10471 ev->addr.type = link_to_bdaddr(link_type, addr_type); 10472 ev->rssi = rssi; 10473 10474 if (name) 10475 eir_len += eir_skb_put_data(skb, EIR_NAME_COMPLETE, name, name_len); 10476 else 10477 flags = MGMT_DEV_FOUND_NAME_REQUEST_FAILED; 10478 10479 ev->eir_len = cpu_to_le16(eir_len); 10480 ev->flags = cpu_to_le32(flags); 10481 10482 mgmt_event_skb(skb, NULL); 10483 } 10484 10485 void mgmt_discovering(struct hci_dev *hdev, u8 discovering) 10486 { 10487 struct mgmt_ev_discovering ev; 10488 10489 bt_dev_dbg(hdev, "discovering %u", discovering); 10490 10491 memset(&ev, 0, sizeof(ev)); 10492 ev.type = hdev->discovery.type; 10493 ev.discovering = discovering; 10494 10495 mgmt_event(MGMT_EV_DISCOVERING, hdev, &ev, sizeof(ev), NULL); 10496 } 10497 10498 void mgmt_suspending(struct hci_dev *hdev, u8 state) 10499 { 10500 struct mgmt_ev_controller_suspend ev; 10501 10502 ev.suspend_state = state; 10503 mgmt_event(MGMT_EV_CONTROLLER_SUSPEND, hdev, &ev, sizeof(ev), NULL); 10504 } 10505 10506 void mgmt_resuming(struct hci_dev *hdev, u8 reason, bdaddr_t *bdaddr, 10507 u8 addr_type) 10508 { 10509 struct mgmt_ev_controller_resume ev; 10510 10511 ev.wake_reason = reason; 10512 if (bdaddr) { 10513 bacpy(&ev.addr.bdaddr, bdaddr); 10514 ev.addr.type = addr_type; 10515 } else { 10516 memset(&ev.addr, 0, sizeof(ev.addr)); 10517 } 10518 10519 mgmt_event(MGMT_EV_CONTROLLER_RESUME, hdev, &ev, sizeof(ev), NULL); 10520 } 10521 10522 static struct hci_mgmt_chan chan = { 10523 .channel = HCI_CHANNEL_CONTROL, 10524 .handler_count = ARRAY_SIZE(mgmt_handlers), 10525 .handlers = mgmt_handlers, 10526 .hdev_init = mgmt_init_hdev, 10527 }; 10528 10529 int mgmt_init(void) 10530 { 10531 return hci_mgmt_chan_register(&chan); 10532 } 10533 10534 void mgmt_exit(void) 10535 { 10536 hci_mgmt_chan_unregister(&chan); 10537 } 10538 10539 void mgmt_cleanup(struct sock *sk) 10540 { 10541 struct mgmt_mesh_tx *mesh_tx; 10542 struct hci_dev *hdev; 10543 10544 read_lock(&hci_dev_list_lock); 10545 10546 list_for_each_entry(hdev, &hci_dev_list, list) { 10547 do { 10548 mesh_tx = mgmt_mesh_next(hdev, sk); 10549 10550 if (mesh_tx) 10551 mesh_send_complete(hdev, mesh_tx, true); 10552 } while (mesh_tx); 10553 } 10554 10555 read_unlock(&hci_dev_list_lock); 10556 } 10557