Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/*
2 BlueZ - Bluetooth protocol stack for Linux
3 Copyright (c) 2000-2001, 2010, Code Aurora Forum. All rights reserved.
4
5 Written 2000,2001 by Maxim Krasnyansky <maxk@qualcomm.com>
6
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License version 2 as
9 published by the Free Software Foundation;
10
11 THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12 OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13 FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14 IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15 CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16 WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17 ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18 OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20 ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21 COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22 SOFTWARE IS DISCLAIMED.
23*/
24
25/* Bluetooth HCI event handling. */
26
27#include <asm/unaligned.h>
28
29#include <net/bluetooth/bluetooth.h>
30#include <net/bluetooth/hci_core.h>
31#include <net/bluetooth/mgmt.h>
32
33#include "hci_request.h"
34#include "hci_debugfs.h"
35#include "a2mp.h"
36#include "amp.h"
37#include "smp.h"
38
39#define ZERO_KEY "\x00\x00\x00\x00\x00\x00\x00\x00" \
40 "\x00\x00\x00\x00\x00\x00\x00\x00"
41
42/* Handle HCI Event packets */
43
44static void hci_cc_inquiry_cancel(struct hci_dev *hdev, struct sk_buff *skb)
45{
46 __u8 status = *((__u8 *) skb->data);
47
48 BT_DBG("%s status 0x%2.2x", hdev->name, status);
49
50 if (status)
51 return;
52
53 clear_bit(HCI_INQUIRY, &hdev->flags);
54 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
55 wake_up_bit(&hdev->flags, HCI_INQUIRY);
56
57 hci_dev_lock(hdev);
58 /* Set discovery state to stopped if we're not doing LE active
59 * scanning.
60 */
61 if (!hci_dev_test_flag(hdev, HCI_LE_SCAN) ||
62 hdev->le_scan_type != LE_SCAN_ACTIVE)
63 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
64 hci_dev_unlock(hdev);
65
66 hci_conn_check_pending(hdev);
67}
68
69static void hci_cc_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
70{
71 __u8 status = *((__u8 *) skb->data);
72
73 BT_DBG("%s status 0x%2.2x", hdev->name, status);
74
75 if (status)
76 return;
77
78 hci_dev_set_flag(hdev, HCI_PERIODIC_INQ);
79}
80
81static void hci_cc_exit_periodic_inq(struct hci_dev *hdev, struct sk_buff *skb)
82{
83 __u8 status = *((__u8 *) skb->data);
84
85 BT_DBG("%s status 0x%2.2x", hdev->name, status);
86
87 if (status)
88 return;
89
90 hci_dev_clear_flag(hdev, HCI_PERIODIC_INQ);
91
92 hci_conn_check_pending(hdev);
93}
94
95static void hci_cc_remote_name_req_cancel(struct hci_dev *hdev,
96 struct sk_buff *skb)
97{
98 BT_DBG("%s", hdev->name);
99}
100
101static void hci_cc_role_discovery(struct hci_dev *hdev, struct sk_buff *skb)
102{
103 struct hci_rp_role_discovery *rp = (void *) skb->data;
104 struct hci_conn *conn;
105
106 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
107
108 if (rp->status)
109 return;
110
111 hci_dev_lock(hdev);
112
113 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
114 if (conn)
115 conn->role = rp->role;
116
117 hci_dev_unlock(hdev);
118}
119
120static void hci_cc_read_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
121{
122 struct hci_rp_read_link_policy *rp = (void *) skb->data;
123 struct hci_conn *conn;
124
125 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
126
127 if (rp->status)
128 return;
129
130 hci_dev_lock(hdev);
131
132 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
133 if (conn)
134 conn->link_policy = __le16_to_cpu(rp->policy);
135
136 hci_dev_unlock(hdev);
137}
138
139static void hci_cc_write_link_policy(struct hci_dev *hdev, struct sk_buff *skb)
140{
141 struct hci_rp_write_link_policy *rp = (void *) skb->data;
142 struct hci_conn *conn;
143 void *sent;
144
145 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
146
147 if (rp->status)
148 return;
149
150 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LINK_POLICY);
151 if (!sent)
152 return;
153
154 hci_dev_lock(hdev);
155
156 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
157 if (conn)
158 conn->link_policy = get_unaligned_le16(sent + 2);
159
160 hci_dev_unlock(hdev);
161}
162
163static void hci_cc_read_def_link_policy(struct hci_dev *hdev,
164 struct sk_buff *skb)
165{
166 struct hci_rp_read_def_link_policy *rp = (void *) skb->data;
167
168 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
169
170 if (rp->status)
171 return;
172
173 hdev->link_policy = __le16_to_cpu(rp->policy);
174}
175
176static void hci_cc_write_def_link_policy(struct hci_dev *hdev,
177 struct sk_buff *skb)
178{
179 __u8 status = *((__u8 *) skb->data);
180 void *sent;
181
182 BT_DBG("%s status 0x%2.2x", hdev->name, status);
183
184 if (status)
185 return;
186
187 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_DEF_LINK_POLICY);
188 if (!sent)
189 return;
190
191 hdev->link_policy = get_unaligned_le16(sent);
192}
193
194static void hci_cc_reset(struct hci_dev *hdev, struct sk_buff *skb)
195{
196 __u8 status = *((__u8 *) skb->data);
197
198 BT_DBG("%s status 0x%2.2x", hdev->name, status);
199
200 clear_bit(HCI_RESET, &hdev->flags);
201
202 if (status)
203 return;
204
205 /* Reset all non-persistent flags */
206 hci_dev_clear_volatile_flags(hdev);
207
208 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
209
210 hdev->inq_tx_power = HCI_TX_POWER_INVALID;
211 hdev->adv_tx_power = HCI_TX_POWER_INVALID;
212
213 memset(hdev->adv_data, 0, sizeof(hdev->adv_data));
214 hdev->adv_data_len = 0;
215
216 memset(hdev->scan_rsp_data, 0, sizeof(hdev->scan_rsp_data));
217 hdev->scan_rsp_data_len = 0;
218
219 hdev->le_scan_type = LE_SCAN_PASSIVE;
220
221 hdev->ssp_debug_mode = 0;
222
223 hci_bdaddr_list_clear(&hdev->le_white_list);
224 hci_bdaddr_list_clear(&hdev->le_resolv_list);
225}
226
227static void hci_cc_read_stored_link_key(struct hci_dev *hdev,
228 struct sk_buff *skb)
229{
230 struct hci_rp_read_stored_link_key *rp = (void *)skb->data;
231 struct hci_cp_read_stored_link_key *sent;
232
233 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
234
235 sent = hci_sent_cmd_data(hdev, HCI_OP_READ_STORED_LINK_KEY);
236 if (!sent)
237 return;
238
239 if (!rp->status && sent->read_all == 0x01) {
240 hdev->stored_max_keys = rp->max_keys;
241 hdev->stored_num_keys = rp->num_keys;
242 }
243}
244
245static void hci_cc_delete_stored_link_key(struct hci_dev *hdev,
246 struct sk_buff *skb)
247{
248 struct hci_rp_delete_stored_link_key *rp = (void *)skb->data;
249
250 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
251
252 if (rp->status)
253 return;
254
255 if (rp->num_keys <= hdev->stored_num_keys)
256 hdev->stored_num_keys -= rp->num_keys;
257 else
258 hdev->stored_num_keys = 0;
259}
260
261static void hci_cc_write_local_name(struct hci_dev *hdev, struct sk_buff *skb)
262{
263 __u8 status = *((__u8 *) skb->data);
264 void *sent;
265
266 BT_DBG("%s status 0x%2.2x", hdev->name, status);
267
268 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LOCAL_NAME);
269 if (!sent)
270 return;
271
272 hci_dev_lock(hdev);
273
274 if (hci_dev_test_flag(hdev, HCI_MGMT))
275 mgmt_set_local_name_complete(hdev, sent, status);
276 else if (!status)
277 memcpy(hdev->dev_name, sent, HCI_MAX_NAME_LENGTH);
278
279 hci_dev_unlock(hdev);
280}
281
282static void hci_cc_read_local_name(struct hci_dev *hdev, struct sk_buff *skb)
283{
284 struct hci_rp_read_local_name *rp = (void *) skb->data;
285
286 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
287
288 if (rp->status)
289 return;
290
291 if (hci_dev_test_flag(hdev, HCI_SETUP) ||
292 hci_dev_test_flag(hdev, HCI_CONFIG))
293 memcpy(hdev->dev_name, rp->name, HCI_MAX_NAME_LENGTH);
294}
295
296static void hci_cc_write_auth_enable(struct hci_dev *hdev, struct sk_buff *skb)
297{
298 __u8 status = *((__u8 *) skb->data);
299 void *sent;
300
301 BT_DBG("%s status 0x%2.2x", hdev->name, status);
302
303 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_AUTH_ENABLE);
304 if (!sent)
305 return;
306
307 hci_dev_lock(hdev);
308
309 if (!status) {
310 __u8 param = *((__u8 *) sent);
311
312 if (param == AUTH_ENABLED)
313 set_bit(HCI_AUTH, &hdev->flags);
314 else
315 clear_bit(HCI_AUTH, &hdev->flags);
316 }
317
318 if (hci_dev_test_flag(hdev, HCI_MGMT))
319 mgmt_auth_enable_complete(hdev, status);
320
321 hci_dev_unlock(hdev);
322}
323
324static void hci_cc_write_encrypt_mode(struct hci_dev *hdev, struct sk_buff *skb)
325{
326 __u8 status = *((__u8 *) skb->data);
327 __u8 param;
328 void *sent;
329
330 BT_DBG("%s status 0x%2.2x", hdev->name, status);
331
332 if (status)
333 return;
334
335 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_ENCRYPT_MODE);
336 if (!sent)
337 return;
338
339 param = *((__u8 *) sent);
340
341 if (param)
342 set_bit(HCI_ENCRYPT, &hdev->flags);
343 else
344 clear_bit(HCI_ENCRYPT, &hdev->flags);
345}
346
347static void hci_cc_write_scan_enable(struct hci_dev *hdev, struct sk_buff *skb)
348{
349 __u8 status = *((__u8 *) skb->data);
350 __u8 param;
351 void *sent;
352
353 BT_DBG("%s status 0x%2.2x", hdev->name, status);
354
355 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SCAN_ENABLE);
356 if (!sent)
357 return;
358
359 param = *((__u8 *) sent);
360
361 hci_dev_lock(hdev);
362
363 if (status) {
364 hdev->discov_timeout = 0;
365 goto done;
366 }
367
368 if (param & SCAN_INQUIRY)
369 set_bit(HCI_ISCAN, &hdev->flags);
370 else
371 clear_bit(HCI_ISCAN, &hdev->flags);
372
373 if (param & SCAN_PAGE)
374 set_bit(HCI_PSCAN, &hdev->flags);
375 else
376 clear_bit(HCI_PSCAN, &hdev->flags);
377
378done:
379 hci_dev_unlock(hdev);
380}
381
382static void hci_cc_read_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
383{
384 struct hci_rp_read_class_of_dev *rp = (void *) skb->data;
385
386 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
387
388 if (rp->status)
389 return;
390
391 memcpy(hdev->dev_class, rp->dev_class, 3);
392
393 BT_DBG("%s class 0x%.2x%.2x%.2x", hdev->name,
394 hdev->dev_class[2], hdev->dev_class[1], hdev->dev_class[0]);
395}
396
397static void hci_cc_write_class_of_dev(struct hci_dev *hdev, struct sk_buff *skb)
398{
399 __u8 status = *((__u8 *) skb->data);
400 void *sent;
401
402 BT_DBG("%s status 0x%2.2x", hdev->name, status);
403
404 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_CLASS_OF_DEV);
405 if (!sent)
406 return;
407
408 hci_dev_lock(hdev);
409
410 if (status == 0)
411 memcpy(hdev->dev_class, sent, 3);
412
413 if (hci_dev_test_flag(hdev, HCI_MGMT))
414 mgmt_set_class_of_dev_complete(hdev, sent, status);
415
416 hci_dev_unlock(hdev);
417}
418
419static void hci_cc_read_voice_setting(struct hci_dev *hdev, struct sk_buff *skb)
420{
421 struct hci_rp_read_voice_setting *rp = (void *) skb->data;
422 __u16 setting;
423
424 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
425
426 if (rp->status)
427 return;
428
429 setting = __le16_to_cpu(rp->voice_setting);
430
431 if (hdev->voice_setting == setting)
432 return;
433
434 hdev->voice_setting = setting;
435
436 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
437
438 if (hdev->notify)
439 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
440}
441
442static void hci_cc_write_voice_setting(struct hci_dev *hdev,
443 struct sk_buff *skb)
444{
445 __u8 status = *((__u8 *) skb->data);
446 __u16 setting;
447 void *sent;
448
449 BT_DBG("%s status 0x%2.2x", hdev->name, status);
450
451 if (status)
452 return;
453
454 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_VOICE_SETTING);
455 if (!sent)
456 return;
457
458 setting = get_unaligned_le16(sent);
459
460 if (hdev->voice_setting == setting)
461 return;
462
463 hdev->voice_setting = setting;
464
465 BT_DBG("%s voice setting 0x%4.4x", hdev->name, setting);
466
467 if (hdev->notify)
468 hdev->notify(hdev, HCI_NOTIFY_VOICE_SETTING);
469}
470
471static void hci_cc_read_num_supported_iac(struct hci_dev *hdev,
472 struct sk_buff *skb)
473{
474 struct hci_rp_read_num_supported_iac *rp = (void *) skb->data;
475
476 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
477
478 if (rp->status)
479 return;
480
481 hdev->num_iac = rp->num_iac;
482
483 BT_DBG("%s num iac %d", hdev->name, hdev->num_iac);
484}
485
486static void hci_cc_write_ssp_mode(struct hci_dev *hdev, struct sk_buff *skb)
487{
488 __u8 status = *((__u8 *) skb->data);
489 struct hci_cp_write_ssp_mode *sent;
490
491 BT_DBG("%s status 0x%2.2x", hdev->name, status);
492
493 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SSP_MODE);
494 if (!sent)
495 return;
496
497 hci_dev_lock(hdev);
498
499 if (!status) {
500 if (sent->mode)
501 hdev->features[1][0] |= LMP_HOST_SSP;
502 else
503 hdev->features[1][0] &= ~LMP_HOST_SSP;
504 }
505
506 if (hci_dev_test_flag(hdev, HCI_MGMT))
507 mgmt_ssp_enable_complete(hdev, sent->mode, status);
508 else if (!status) {
509 if (sent->mode)
510 hci_dev_set_flag(hdev, HCI_SSP_ENABLED);
511 else
512 hci_dev_clear_flag(hdev, HCI_SSP_ENABLED);
513 }
514
515 hci_dev_unlock(hdev);
516}
517
518static void hci_cc_write_sc_support(struct hci_dev *hdev, struct sk_buff *skb)
519{
520 u8 status = *((u8 *) skb->data);
521 struct hci_cp_write_sc_support *sent;
522
523 BT_DBG("%s status 0x%2.2x", hdev->name, status);
524
525 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SC_SUPPORT);
526 if (!sent)
527 return;
528
529 hci_dev_lock(hdev);
530
531 if (!status) {
532 if (sent->support)
533 hdev->features[1][0] |= LMP_HOST_SC;
534 else
535 hdev->features[1][0] &= ~LMP_HOST_SC;
536 }
537
538 if (!hci_dev_test_flag(hdev, HCI_MGMT) && !status) {
539 if (sent->support)
540 hci_dev_set_flag(hdev, HCI_SC_ENABLED);
541 else
542 hci_dev_clear_flag(hdev, HCI_SC_ENABLED);
543 }
544
545 hci_dev_unlock(hdev);
546}
547
548static void hci_cc_read_local_version(struct hci_dev *hdev, struct sk_buff *skb)
549{
550 struct hci_rp_read_local_version *rp = (void *) skb->data;
551
552 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
553
554 if (rp->status)
555 return;
556
557 if (hci_dev_test_flag(hdev, HCI_SETUP) ||
558 hci_dev_test_flag(hdev, HCI_CONFIG)) {
559 hdev->hci_ver = rp->hci_ver;
560 hdev->hci_rev = __le16_to_cpu(rp->hci_rev);
561 hdev->lmp_ver = rp->lmp_ver;
562 hdev->manufacturer = __le16_to_cpu(rp->manufacturer);
563 hdev->lmp_subver = __le16_to_cpu(rp->lmp_subver);
564 }
565}
566
567static void hci_cc_read_local_commands(struct hci_dev *hdev,
568 struct sk_buff *skb)
569{
570 struct hci_rp_read_local_commands *rp = (void *) skb->data;
571
572 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
573
574 if (rp->status)
575 return;
576
577 if (hci_dev_test_flag(hdev, HCI_SETUP) ||
578 hci_dev_test_flag(hdev, HCI_CONFIG))
579 memcpy(hdev->commands, rp->commands, sizeof(hdev->commands));
580}
581
582static void hci_cc_read_local_features(struct hci_dev *hdev,
583 struct sk_buff *skb)
584{
585 struct hci_rp_read_local_features *rp = (void *) skb->data;
586
587 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
588
589 if (rp->status)
590 return;
591
592 memcpy(hdev->features, rp->features, 8);
593
594 /* Adjust default settings according to features
595 * supported by device. */
596
597 if (hdev->features[0][0] & LMP_3SLOT)
598 hdev->pkt_type |= (HCI_DM3 | HCI_DH3);
599
600 if (hdev->features[0][0] & LMP_5SLOT)
601 hdev->pkt_type |= (HCI_DM5 | HCI_DH5);
602
603 if (hdev->features[0][1] & LMP_HV2) {
604 hdev->pkt_type |= (HCI_HV2);
605 hdev->esco_type |= (ESCO_HV2);
606 }
607
608 if (hdev->features[0][1] & LMP_HV3) {
609 hdev->pkt_type |= (HCI_HV3);
610 hdev->esco_type |= (ESCO_HV3);
611 }
612
613 if (lmp_esco_capable(hdev))
614 hdev->esco_type |= (ESCO_EV3);
615
616 if (hdev->features[0][4] & LMP_EV4)
617 hdev->esco_type |= (ESCO_EV4);
618
619 if (hdev->features[0][4] & LMP_EV5)
620 hdev->esco_type |= (ESCO_EV5);
621
622 if (hdev->features[0][5] & LMP_EDR_ESCO_2M)
623 hdev->esco_type |= (ESCO_2EV3);
624
625 if (hdev->features[0][5] & LMP_EDR_ESCO_3M)
626 hdev->esco_type |= (ESCO_3EV3);
627
628 if (hdev->features[0][5] & LMP_EDR_3S_ESCO)
629 hdev->esco_type |= (ESCO_2EV5 | ESCO_3EV5);
630}
631
632static void hci_cc_read_local_ext_features(struct hci_dev *hdev,
633 struct sk_buff *skb)
634{
635 struct hci_rp_read_local_ext_features *rp = (void *) skb->data;
636
637 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
638
639 if (rp->status)
640 return;
641
642 if (hdev->max_page < rp->max_page)
643 hdev->max_page = rp->max_page;
644
645 if (rp->page < HCI_MAX_PAGES)
646 memcpy(hdev->features[rp->page], rp->features, 8);
647}
648
649static void hci_cc_read_flow_control_mode(struct hci_dev *hdev,
650 struct sk_buff *skb)
651{
652 struct hci_rp_read_flow_control_mode *rp = (void *) skb->data;
653
654 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
655
656 if (rp->status)
657 return;
658
659 hdev->flow_ctl_mode = rp->mode;
660}
661
662static void hci_cc_read_buffer_size(struct hci_dev *hdev, struct sk_buff *skb)
663{
664 struct hci_rp_read_buffer_size *rp = (void *) skb->data;
665
666 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
667
668 if (rp->status)
669 return;
670
671 hdev->acl_mtu = __le16_to_cpu(rp->acl_mtu);
672 hdev->sco_mtu = rp->sco_mtu;
673 hdev->acl_pkts = __le16_to_cpu(rp->acl_max_pkt);
674 hdev->sco_pkts = __le16_to_cpu(rp->sco_max_pkt);
675
676 if (test_bit(HCI_QUIRK_FIXUP_BUFFER_SIZE, &hdev->quirks)) {
677 hdev->sco_mtu = 64;
678 hdev->sco_pkts = 8;
679 }
680
681 hdev->acl_cnt = hdev->acl_pkts;
682 hdev->sco_cnt = hdev->sco_pkts;
683
684 BT_DBG("%s acl mtu %d:%d sco mtu %d:%d", hdev->name, hdev->acl_mtu,
685 hdev->acl_pkts, hdev->sco_mtu, hdev->sco_pkts);
686}
687
688static void hci_cc_read_bd_addr(struct hci_dev *hdev, struct sk_buff *skb)
689{
690 struct hci_rp_read_bd_addr *rp = (void *) skb->data;
691
692 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
693
694 if (rp->status)
695 return;
696
697 if (test_bit(HCI_INIT, &hdev->flags))
698 bacpy(&hdev->bdaddr, &rp->bdaddr);
699
700 if (hci_dev_test_flag(hdev, HCI_SETUP))
701 bacpy(&hdev->setup_addr, &rp->bdaddr);
702}
703
704static void hci_cc_read_page_scan_activity(struct hci_dev *hdev,
705 struct sk_buff *skb)
706{
707 struct hci_rp_read_page_scan_activity *rp = (void *) skb->data;
708
709 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
710
711 if (rp->status)
712 return;
713
714 if (test_bit(HCI_INIT, &hdev->flags)) {
715 hdev->page_scan_interval = __le16_to_cpu(rp->interval);
716 hdev->page_scan_window = __le16_to_cpu(rp->window);
717 }
718}
719
720static void hci_cc_write_page_scan_activity(struct hci_dev *hdev,
721 struct sk_buff *skb)
722{
723 u8 status = *((u8 *) skb->data);
724 struct hci_cp_write_page_scan_activity *sent;
725
726 BT_DBG("%s status 0x%2.2x", hdev->name, status);
727
728 if (status)
729 return;
730
731 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_ACTIVITY);
732 if (!sent)
733 return;
734
735 hdev->page_scan_interval = __le16_to_cpu(sent->interval);
736 hdev->page_scan_window = __le16_to_cpu(sent->window);
737}
738
739static void hci_cc_read_page_scan_type(struct hci_dev *hdev,
740 struct sk_buff *skb)
741{
742 struct hci_rp_read_page_scan_type *rp = (void *) skb->data;
743
744 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
745
746 if (rp->status)
747 return;
748
749 if (test_bit(HCI_INIT, &hdev->flags))
750 hdev->page_scan_type = rp->type;
751}
752
753static void hci_cc_write_page_scan_type(struct hci_dev *hdev,
754 struct sk_buff *skb)
755{
756 u8 status = *((u8 *) skb->data);
757 u8 *type;
758
759 BT_DBG("%s status 0x%2.2x", hdev->name, status);
760
761 if (status)
762 return;
763
764 type = hci_sent_cmd_data(hdev, HCI_OP_WRITE_PAGE_SCAN_TYPE);
765 if (type)
766 hdev->page_scan_type = *type;
767}
768
769static void hci_cc_read_data_block_size(struct hci_dev *hdev,
770 struct sk_buff *skb)
771{
772 struct hci_rp_read_data_block_size *rp = (void *) skb->data;
773
774 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
775
776 if (rp->status)
777 return;
778
779 hdev->block_mtu = __le16_to_cpu(rp->max_acl_len);
780 hdev->block_len = __le16_to_cpu(rp->block_len);
781 hdev->num_blocks = __le16_to_cpu(rp->num_blocks);
782
783 hdev->block_cnt = hdev->num_blocks;
784
785 BT_DBG("%s blk mtu %d cnt %d len %d", hdev->name, hdev->block_mtu,
786 hdev->block_cnt, hdev->block_len);
787}
788
789static void hci_cc_read_clock(struct hci_dev *hdev, struct sk_buff *skb)
790{
791 struct hci_rp_read_clock *rp = (void *) skb->data;
792 struct hci_cp_read_clock *cp;
793 struct hci_conn *conn;
794
795 BT_DBG("%s", hdev->name);
796
797 if (skb->len < sizeof(*rp))
798 return;
799
800 if (rp->status)
801 return;
802
803 hci_dev_lock(hdev);
804
805 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_CLOCK);
806 if (!cp)
807 goto unlock;
808
809 if (cp->which == 0x00) {
810 hdev->clock = le32_to_cpu(rp->clock);
811 goto unlock;
812 }
813
814 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
815 if (conn) {
816 conn->clock = le32_to_cpu(rp->clock);
817 conn->clock_accuracy = le16_to_cpu(rp->accuracy);
818 }
819
820unlock:
821 hci_dev_unlock(hdev);
822}
823
824static void hci_cc_read_local_amp_info(struct hci_dev *hdev,
825 struct sk_buff *skb)
826{
827 struct hci_rp_read_local_amp_info *rp = (void *) skb->data;
828
829 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
830
831 if (rp->status)
832 return;
833
834 hdev->amp_status = rp->amp_status;
835 hdev->amp_total_bw = __le32_to_cpu(rp->total_bw);
836 hdev->amp_max_bw = __le32_to_cpu(rp->max_bw);
837 hdev->amp_min_latency = __le32_to_cpu(rp->min_latency);
838 hdev->amp_max_pdu = __le32_to_cpu(rp->max_pdu);
839 hdev->amp_type = rp->amp_type;
840 hdev->amp_pal_cap = __le16_to_cpu(rp->pal_cap);
841 hdev->amp_assoc_size = __le16_to_cpu(rp->max_assoc_size);
842 hdev->amp_be_flush_to = __le32_to_cpu(rp->be_flush_to);
843 hdev->amp_max_flush_to = __le32_to_cpu(rp->max_flush_to);
844}
845
846static void hci_cc_read_inq_rsp_tx_power(struct hci_dev *hdev,
847 struct sk_buff *skb)
848{
849 struct hci_rp_read_inq_rsp_tx_power *rp = (void *) skb->data;
850
851 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
852
853 if (rp->status)
854 return;
855
856 hdev->inq_tx_power = rp->tx_power;
857}
858
859static void hci_cc_pin_code_reply(struct hci_dev *hdev, struct sk_buff *skb)
860{
861 struct hci_rp_pin_code_reply *rp = (void *) skb->data;
862 struct hci_cp_pin_code_reply *cp;
863 struct hci_conn *conn;
864
865 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
866
867 hci_dev_lock(hdev);
868
869 if (hci_dev_test_flag(hdev, HCI_MGMT))
870 mgmt_pin_code_reply_complete(hdev, &rp->bdaddr, rp->status);
871
872 if (rp->status)
873 goto unlock;
874
875 cp = hci_sent_cmd_data(hdev, HCI_OP_PIN_CODE_REPLY);
876 if (!cp)
877 goto unlock;
878
879 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
880 if (conn)
881 conn->pin_length = cp->pin_len;
882
883unlock:
884 hci_dev_unlock(hdev);
885}
886
887static void hci_cc_pin_code_neg_reply(struct hci_dev *hdev, struct sk_buff *skb)
888{
889 struct hci_rp_pin_code_neg_reply *rp = (void *) skb->data;
890
891 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
892
893 hci_dev_lock(hdev);
894
895 if (hci_dev_test_flag(hdev, HCI_MGMT))
896 mgmt_pin_code_neg_reply_complete(hdev, &rp->bdaddr,
897 rp->status);
898
899 hci_dev_unlock(hdev);
900}
901
902static void hci_cc_le_read_buffer_size(struct hci_dev *hdev,
903 struct sk_buff *skb)
904{
905 struct hci_rp_le_read_buffer_size *rp = (void *) skb->data;
906
907 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
908
909 if (rp->status)
910 return;
911
912 hdev->le_mtu = __le16_to_cpu(rp->le_mtu);
913 hdev->le_pkts = rp->le_max_pkt;
914
915 hdev->le_cnt = hdev->le_pkts;
916
917 BT_DBG("%s le mtu %d:%d", hdev->name, hdev->le_mtu, hdev->le_pkts);
918}
919
920static void hci_cc_le_read_local_features(struct hci_dev *hdev,
921 struct sk_buff *skb)
922{
923 struct hci_rp_le_read_local_features *rp = (void *) skb->data;
924
925 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
926
927 if (rp->status)
928 return;
929
930 memcpy(hdev->le_features, rp->features, 8);
931}
932
933static void hci_cc_le_read_adv_tx_power(struct hci_dev *hdev,
934 struct sk_buff *skb)
935{
936 struct hci_rp_le_read_adv_tx_power *rp = (void *) skb->data;
937
938 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
939
940 if (rp->status)
941 return;
942
943 hdev->adv_tx_power = rp->tx_power;
944}
945
946static void hci_cc_user_confirm_reply(struct hci_dev *hdev, struct sk_buff *skb)
947{
948 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
949
950 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
951
952 hci_dev_lock(hdev);
953
954 if (hci_dev_test_flag(hdev, HCI_MGMT))
955 mgmt_user_confirm_reply_complete(hdev, &rp->bdaddr, ACL_LINK, 0,
956 rp->status);
957
958 hci_dev_unlock(hdev);
959}
960
961static void hci_cc_user_confirm_neg_reply(struct hci_dev *hdev,
962 struct sk_buff *skb)
963{
964 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
965
966 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
967
968 hci_dev_lock(hdev);
969
970 if (hci_dev_test_flag(hdev, HCI_MGMT))
971 mgmt_user_confirm_neg_reply_complete(hdev, &rp->bdaddr,
972 ACL_LINK, 0, rp->status);
973
974 hci_dev_unlock(hdev);
975}
976
977static void hci_cc_user_passkey_reply(struct hci_dev *hdev, struct sk_buff *skb)
978{
979 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
980
981 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
982
983 hci_dev_lock(hdev);
984
985 if (hci_dev_test_flag(hdev, HCI_MGMT))
986 mgmt_user_passkey_reply_complete(hdev, &rp->bdaddr, ACL_LINK,
987 0, rp->status);
988
989 hci_dev_unlock(hdev);
990}
991
992static void hci_cc_user_passkey_neg_reply(struct hci_dev *hdev,
993 struct sk_buff *skb)
994{
995 struct hci_rp_user_confirm_reply *rp = (void *) skb->data;
996
997 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
998
999 hci_dev_lock(hdev);
1000
1001 if (hci_dev_test_flag(hdev, HCI_MGMT))
1002 mgmt_user_passkey_neg_reply_complete(hdev, &rp->bdaddr,
1003 ACL_LINK, 0, rp->status);
1004
1005 hci_dev_unlock(hdev);
1006}
1007
1008static void hci_cc_read_local_oob_data(struct hci_dev *hdev,
1009 struct sk_buff *skb)
1010{
1011 struct hci_rp_read_local_oob_data *rp = (void *) skb->data;
1012
1013 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1014}
1015
1016static void hci_cc_read_local_oob_ext_data(struct hci_dev *hdev,
1017 struct sk_buff *skb)
1018{
1019 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data;
1020
1021 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1022}
1023
1024static void hci_cc_le_set_random_addr(struct hci_dev *hdev, struct sk_buff *skb)
1025{
1026 __u8 status = *((__u8 *) skb->data);
1027 bdaddr_t *sent;
1028
1029 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1030
1031 if (status)
1032 return;
1033
1034 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_RANDOM_ADDR);
1035 if (!sent)
1036 return;
1037
1038 hci_dev_lock(hdev);
1039
1040 bacpy(&hdev->random_addr, sent);
1041
1042 hci_dev_unlock(hdev);
1043}
1044
1045static void hci_cc_le_set_default_phy(struct hci_dev *hdev, struct sk_buff *skb)
1046{
1047 __u8 status = *((__u8 *) skb->data);
1048 struct hci_cp_le_set_default_phy *cp;
1049
1050 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1051
1052 if (status)
1053 return;
1054
1055 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_DEFAULT_PHY);
1056 if (!cp)
1057 return;
1058
1059 hci_dev_lock(hdev);
1060
1061 hdev->le_tx_def_phys = cp->tx_phys;
1062 hdev->le_rx_def_phys = cp->rx_phys;
1063
1064 hci_dev_unlock(hdev);
1065}
1066
1067static void hci_cc_le_set_adv_set_random_addr(struct hci_dev *hdev,
1068 struct sk_buff *skb)
1069{
1070 __u8 status = *((__u8 *) skb->data);
1071 struct hci_cp_le_set_adv_set_rand_addr *cp;
1072 struct adv_info *adv_instance;
1073
1074 if (status)
1075 return;
1076
1077 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_SET_RAND_ADDR);
1078 if (!cp)
1079 return;
1080
1081 hci_dev_lock(hdev);
1082
1083 if (!hdev->cur_adv_instance) {
1084 /* Store in hdev for instance 0 (Set adv and Directed advs) */
1085 bacpy(&hdev->random_addr, &cp->bdaddr);
1086 } else {
1087 adv_instance = hci_find_adv_instance(hdev,
1088 hdev->cur_adv_instance);
1089 if (adv_instance)
1090 bacpy(&adv_instance->random_addr, &cp->bdaddr);
1091 }
1092
1093 hci_dev_unlock(hdev);
1094}
1095
1096static void hci_cc_le_set_adv_enable(struct hci_dev *hdev, struct sk_buff *skb)
1097{
1098 __u8 *sent, status = *((__u8 *) skb->data);
1099
1100 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1101
1102 if (status)
1103 return;
1104
1105 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_ENABLE);
1106 if (!sent)
1107 return;
1108
1109 hci_dev_lock(hdev);
1110
1111 /* If we're doing connection initiation as peripheral. Set a
1112 * timeout in case something goes wrong.
1113 */
1114 if (*sent) {
1115 struct hci_conn *conn;
1116
1117 hci_dev_set_flag(hdev, HCI_LE_ADV);
1118
1119 conn = hci_lookup_le_connect(hdev);
1120 if (conn)
1121 queue_delayed_work(hdev->workqueue,
1122 &conn->le_conn_timeout,
1123 conn->conn_timeout);
1124 } else {
1125 hci_dev_clear_flag(hdev, HCI_LE_ADV);
1126 }
1127
1128 hci_dev_unlock(hdev);
1129}
1130
1131static void hci_cc_le_set_ext_adv_enable(struct hci_dev *hdev,
1132 struct sk_buff *skb)
1133{
1134 struct hci_cp_le_set_ext_adv_enable *cp;
1135 __u8 status = *((__u8 *) skb->data);
1136
1137 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1138
1139 if (status)
1140 return;
1141
1142 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_EXT_ADV_ENABLE);
1143 if (!cp)
1144 return;
1145
1146 hci_dev_lock(hdev);
1147
1148 if (cp->enable) {
1149 struct hci_conn *conn;
1150
1151 hci_dev_set_flag(hdev, HCI_LE_ADV);
1152
1153 conn = hci_lookup_le_connect(hdev);
1154 if (conn)
1155 queue_delayed_work(hdev->workqueue,
1156 &conn->le_conn_timeout,
1157 conn->conn_timeout);
1158 } else {
1159 hci_dev_clear_flag(hdev, HCI_LE_ADV);
1160 }
1161
1162 hci_dev_unlock(hdev);
1163}
1164
1165static void hci_cc_le_set_scan_param(struct hci_dev *hdev, struct sk_buff *skb)
1166{
1167 struct hci_cp_le_set_scan_param *cp;
1168 __u8 status = *((__u8 *) skb->data);
1169
1170 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1171
1172 if (status)
1173 return;
1174
1175 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_PARAM);
1176 if (!cp)
1177 return;
1178
1179 hci_dev_lock(hdev);
1180
1181 hdev->le_scan_type = cp->type;
1182
1183 hci_dev_unlock(hdev);
1184}
1185
1186static void hci_cc_le_set_ext_scan_param(struct hci_dev *hdev,
1187 struct sk_buff *skb)
1188{
1189 struct hci_cp_le_set_ext_scan_params *cp;
1190 __u8 status = *((__u8 *) skb->data);
1191 struct hci_cp_le_scan_phy_params *phy_param;
1192
1193 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1194
1195 if (status)
1196 return;
1197
1198 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_EXT_SCAN_PARAMS);
1199 if (!cp)
1200 return;
1201
1202 phy_param = (void *)cp->data;
1203
1204 hci_dev_lock(hdev);
1205
1206 hdev->le_scan_type = phy_param->type;
1207
1208 hci_dev_unlock(hdev);
1209}
1210
1211static bool has_pending_adv_report(struct hci_dev *hdev)
1212{
1213 struct discovery_state *d = &hdev->discovery;
1214
1215 return bacmp(&d->last_adv_addr, BDADDR_ANY);
1216}
1217
1218static void clear_pending_adv_report(struct hci_dev *hdev)
1219{
1220 struct discovery_state *d = &hdev->discovery;
1221
1222 bacpy(&d->last_adv_addr, BDADDR_ANY);
1223 d->last_adv_data_len = 0;
1224}
1225
1226static void store_pending_adv_report(struct hci_dev *hdev, bdaddr_t *bdaddr,
1227 u8 bdaddr_type, s8 rssi, u32 flags,
1228 u8 *data, u8 len)
1229{
1230 struct discovery_state *d = &hdev->discovery;
1231
1232 bacpy(&d->last_adv_addr, bdaddr);
1233 d->last_adv_addr_type = bdaddr_type;
1234 d->last_adv_rssi = rssi;
1235 d->last_adv_flags = flags;
1236 memcpy(d->last_adv_data, data, len);
1237 d->last_adv_data_len = len;
1238}
1239
1240static void le_set_scan_enable_complete(struct hci_dev *hdev, u8 enable)
1241{
1242 hci_dev_lock(hdev);
1243
1244 switch (enable) {
1245 case LE_SCAN_ENABLE:
1246 hci_dev_set_flag(hdev, HCI_LE_SCAN);
1247 if (hdev->le_scan_type == LE_SCAN_ACTIVE)
1248 clear_pending_adv_report(hdev);
1249 break;
1250
1251 case LE_SCAN_DISABLE:
1252 /* We do this here instead of when setting DISCOVERY_STOPPED
1253 * since the latter would potentially require waiting for
1254 * inquiry to stop too.
1255 */
1256 if (has_pending_adv_report(hdev)) {
1257 struct discovery_state *d = &hdev->discovery;
1258
1259 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
1260 d->last_adv_addr_type, NULL,
1261 d->last_adv_rssi, d->last_adv_flags,
1262 d->last_adv_data,
1263 d->last_adv_data_len, NULL, 0);
1264 }
1265
1266 /* Cancel this timer so that we don't try to disable scanning
1267 * when it's already disabled.
1268 */
1269 cancel_delayed_work(&hdev->le_scan_disable);
1270
1271 hci_dev_clear_flag(hdev, HCI_LE_SCAN);
1272
1273 /* The HCI_LE_SCAN_INTERRUPTED flag indicates that we
1274 * interrupted scanning due to a connect request. Mark
1275 * therefore discovery as stopped. If this was not
1276 * because of a connect request advertising might have
1277 * been disabled because of active scanning, so
1278 * re-enable it again if necessary.
1279 */
1280 if (hci_dev_test_and_clear_flag(hdev, HCI_LE_SCAN_INTERRUPTED))
1281 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1282 else if (!hci_dev_test_flag(hdev, HCI_LE_ADV) &&
1283 hdev->discovery.state == DISCOVERY_FINDING)
1284 hci_req_reenable_advertising(hdev);
1285
1286 break;
1287
1288 default:
1289 bt_dev_err(hdev, "use of reserved LE_Scan_Enable param %d",
1290 enable);
1291 break;
1292 }
1293
1294 hci_dev_unlock(hdev);
1295}
1296
1297static void hci_cc_le_set_scan_enable(struct hci_dev *hdev,
1298 struct sk_buff *skb)
1299{
1300 struct hci_cp_le_set_scan_enable *cp;
1301 __u8 status = *((__u8 *) skb->data);
1302
1303 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1304
1305 if (status)
1306 return;
1307
1308 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_SCAN_ENABLE);
1309 if (!cp)
1310 return;
1311
1312 le_set_scan_enable_complete(hdev, cp->enable);
1313}
1314
1315static void hci_cc_le_set_ext_scan_enable(struct hci_dev *hdev,
1316 struct sk_buff *skb)
1317{
1318 struct hci_cp_le_set_ext_scan_enable *cp;
1319 __u8 status = *((__u8 *) skb->data);
1320
1321 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1322
1323 if (status)
1324 return;
1325
1326 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_EXT_SCAN_ENABLE);
1327 if (!cp)
1328 return;
1329
1330 le_set_scan_enable_complete(hdev, cp->enable);
1331}
1332
1333static void hci_cc_le_read_num_adv_sets(struct hci_dev *hdev,
1334 struct sk_buff *skb)
1335{
1336 struct hci_rp_le_read_num_supported_adv_sets *rp = (void *) skb->data;
1337
1338 BT_DBG("%s status 0x%2.2x No of Adv sets %u", hdev->name, rp->status,
1339 rp->num_of_sets);
1340
1341 if (rp->status)
1342 return;
1343
1344 hdev->le_num_of_adv_sets = rp->num_of_sets;
1345}
1346
1347static void hci_cc_le_read_white_list_size(struct hci_dev *hdev,
1348 struct sk_buff *skb)
1349{
1350 struct hci_rp_le_read_white_list_size *rp = (void *) skb->data;
1351
1352 BT_DBG("%s status 0x%2.2x size %u", hdev->name, rp->status, rp->size);
1353
1354 if (rp->status)
1355 return;
1356
1357 hdev->le_white_list_size = rp->size;
1358}
1359
1360static void hci_cc_le_clear_white_list(struct hci_dev *hdev,
1361 struct sk_buff *skb)
1362{
1363 __u8 status = *((__u8 *) skb->data);
1364
1365 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1366
1367 if (status)
1368 return;
1369
1370 hci_bdaddr_list_clear(&hdev->le_white_list);
1371}
1372
1373static void hci_cc_le_add_to_white_list(struct hci_dev *hdev,
1374 struct sk_buff *skb)
1375{
1376 struct hci_cp_le_add_to_white_list *sent;
1377 __u8 status = *((__u8 *) skb->data);
1378
1379 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1380
1381 if (status)
1382 return;
1383
1384 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_ADD_TO_WHITE_LIST);
1385 if (!sent)
1386 return;
1387
1388 hci_bdaddr_list_add(&hdev->le_white_list, &sent->bdaddr,
1389 sent->bdaddr_type);
1390}
1391
1392static void hci_cc_le_del_from_white_list(struct hci_dev *hdev,
1393 struct sk_buff *skb)
1394{
1395 struct hci_cp_le_del_from_white_list *sent;
1396 __u8 status = *((__u8 *) skb->data);
1397
1398 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1399
1400 if (status)
1401 return;
1402
1403 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_DEL_FROM_WHITE_LIST);
1404 if (!sent)
1405 return;
1406
1407 hci_bdaddr_list_del(&hdev->le_white_list, &sent->bdaddr,
1408 sent->bdaddr_type);
1409}
1410
1411static void hci_cc_le_read_supported_states(struct hci_dev *hdev,
1412 struct sk_buff *skb)
1413{
1414 struct hci_rp_le_read_supported_states *rp = (void *) skb->data;
1415
1416 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1417
1418 if (rp->status)
1419 return;
1420
1421 memcpy(hdev->le_states, rp->le_states, 8);
1422}
1423
1424static void hci_cc_le_read_def_data_len(struct hci_dev *hdev,
1425 struct sk_buff *skb)
1426{
1427 struct hci_rp_le_read_def_data_len *rp = (void *) skb->data;
1428
1429 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1430
1431 if (rp->status)
1432 return;
1433
1434 hdev->le_def_tx_len = le16_to_cpu(rp->tx_len);
1435 hdev->le_def_tx_time = le16_to_cpu(rp->tx_time);
1436}
1437
1438static void hci_cc_le_write_def_data_len(struct hci_dev *hdev,
1439 struct sk_buff *skb)
1440{
1441 struct hci_cp_le_write_def_data_len *sent;
1442 __u8 status = *((__u8 *) skb->data);
1443
1444 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1445
1446 if (status)
1447 return;
1448
1449 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_WRITE_DEF_DATA_LEN);
1450 if (!sent)
1451 return;
1452
1453 hdev->le_def_tx_len = le16_to_cpu(sent->tx_len);
1454 hdev->le_def_tx_time = le16_to_cpu(sent->tx_time);
1455}
1456
1457static void hci_cc_le_add_to_resolv_list(struct hci_dev *hdev,
1458 struct sk_buff *skb)
1459{
1460 struct hci_cp_le_add_to_resolv_list *sent;
1461 __u8 status = *((__u8 *) skb->data);
1462
1463 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1464
1465 if (status)
1466 return;
1467
1468 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_ADD_TO_RESOLV_LIST);
1469 if (!sent)
1470 return;
1471
1472 hci_bdaddr_list_add_with_irk(&hdev->le_resolv_list, &sent->bdaddr,
1473 sent->bdaddr_type, sent->peer_irk,
1474 sent->local_irk);
1475}
1476
1477static void hci_cc_le_del_from_resolv_list(struct hci_dev *hdev,
1478 struct sk_buff *skb)
1479{
1480 struct hci_cp_le_del_from_resolv_list *sent;
1481 __u8 status = *((__u8 *) skb->data);
1482
1483 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1484
1485 if (status)
1486 return;
1487
1488 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_DEL_FROM_RESOLV_LIST);
1489 if (!sent)
1490 return;
1491
1492 hci_bdaddr_list_del_with_irk(&hdev->le_resolv_list, &sent->bdaddr,
1493 sent->bdaddr_type);
1494}
1495
1496static void hci_cc_le_clear_resolv_list(struct hci_dev *hdev,
1497 struct sk_buff *skb)
1498{
1499 __u8 status = *((__u8 *) skb->data);
1500
1501 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1502
1503 if (status)
1504 return;
1505
1506 hci_bdaddr_list_clear(&hdev->le_resolv_list);
1507}
1508
1509static void hci_cc_le_read_resolv_list_size(struct hci_dev *hdev,
1510 struct sk_buff *skb)
1511{
1512 struct hci_rp_le_read_resolv_list_size *rp = (void *) skb->data;
1513
1514 BT_DBG("%s status 0x%2.2x size %u", hdev->name, rp->status, rp->size);
1515
1516 if (rp->status)
1517 return;
1518
1519 hdev->le_resolv_list_size = rp->size;
1520}
1521
1522static void hci_cc_le_set_addr_resolution_enable(struct hci_dev *hdev,
1523 struct sk_buff *skb)
1524{
1525 __u8 *sent, status = *((__u8 *) skb->data);
1526
1527 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1528
1529 if (status)
1530 return;
1531
1532 sent = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADDR_RESOLV_ENABLE);
1533 if (!sent)
1534 return;
1535
1536 hci_dev_lock(hdev);
1537
1538 if (*sent)
1539 hci_dev_set_flag(hdev, HCI_LL_RPA_RESOLUTION);
1540 else
1541 hci_dev_clear_flag(hdev, HCI_LL_RPA_RESOLUTION);
1542
1543 hci_dev_unlock(hdev);
1544}
1545
1546static void hci_cc_le_read_max_data_len(struct hci_dev *hdev,
1547 struct sk_buff *skb)
1548{
1549 struct hci_rp_le_read_max_data_len *rp = (void *) skb->data;
1550
1551 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1552
1553 if (rp->status)
1554 return;
1555
1556 hdev->le_max_tx_len = le16_to_cpu(rp->tx_len);
1557 hdev->le_max_tx_time = le16_to_cpu(rp->tx_time);
1558 hdev->le_max_rx_len = le16_to_cpu(rp->rx_len);
1559 hdev->le_max_rx_time = le16_to_cpu(rp->rx_time);
1560}
1561
1562static void hci_cc_write_le_host_supported(struct hci_dev *hdev,
1563 struct sk_buff *skb)
1564{
1565 struct hci_cp_write_le_host_supported *sent;
1566 __u8 status = *((__u8 *) skb->data);
1567
1568 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1569
1570 if (status)
1571 return;
1572
1573 sent = hci_sent_cmd_data(hdev, HCI_OP_WRITE_LE_HOST_SUPPORTED);
1574 if (!sent)
1575 return;
1576
1577 hci_dev_lock(hdev);
1578
1579 if (sent->le) {
1580 hdev->features[1][0] |= LMP_HOST_LE;
1581 hci_dev_set_flag(hdev, HCI_LE_ENABLED);
1582 } else {
1583 hdev->features[1][0] &= ~LMP_HOST_LE;
1584 hci_dev_clear_flag(hdev, HCI_LE_ENABLED);
1585 hci_dev_clear_flag(hdev, HCI_ADVERTISING);
1586 }
1587
1588 if (sent->simul)
1589 hdev->features[1][0] |= LMP_HOST_LE_BREDR;
1590 else
1591 hdev->features[1][0] &= ~LMP_HOST_LE_BREDR;
1592
1593 hci_dev_unlock(hdev);
1594}
1595
1596static void hci_cc_set_adv_param(struct hci_dev *hdev, struct sk_buff *skb)
1597{
1598 struct hci_cp_le_set_adv_param *cp;
1599 u8 status = *((u8 *) skb->data);
1600
1601 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1602
1603 if (status)
1604 return;
1605
1606 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_ADV_PARAM);
1607 if (!cp)
1608 return;
1609
1610 hci_dev_lock(hdev);
1611 hdev->adv_addr_type = cp->own_address_type;
1612 hci_dev_unlock(hdev);
1613}
1614
1615static void hci_cc_set_ext_adv_param(struct hci_dev *hdev, struct sk_buff *skb)
1616{
1617 struct hci_rp_le_set_ext_adv_params *rp = (void *) skb->data;
1618 struct hci_cp_le_set_ext_adv_params *cp;
1619 struct adv_info *adv_instance;
1620
1621 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1622
1623 if (rp->status)
1624 return;
1625
1626 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_SET_EXT_ADV_PARAMS);
1627 if (!cp)
1628 return;
1629
1630 hci_dev_lock(hdev);
1631 hdev->adv_addr_type = cp->own_addr_type;
1632 if (!hdev->cur_adv_instance) {
1633 /* Store in hdev for instance 0 */
1634 hdev->adv_tx_power = rp->tx_power;
1635 } else {
1636 adv_instance = hci_find_adv_instance(hdev,
1637 hdev->cur_adv_instance);
1638 if (adv_instance)
1639 adv_instance->tx_power = rp->tx_power;
1640 }
1641 /* Update adv data as tx power is known now */
1642 hci_req_update_adv_data(hdev, hdev->cur_adv_instance);
1643 hci_dev_unlock(hdev);
1644}
1645
1646static void hci_cc_read_rssi(struct hci_dev *hdev, struct sk_buff *skb)
1647{
1648 struct hci_rp_read_rssi *rp = (void *) skb->data;
1649 struct hci_conn *conn;
1650
1651 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1652
1653 if (rp->status)
1654 return;
1655
1656 hci_dev_lock(hdev);
1657
1658 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1659 if (conn)
1660 conn->rssi = rp->rssi;
1661
1662 hci_dev_unlock(hdev);
1663}
1664
1665static void hci_cc_read_tx_power(struct hci_dev *hdev, struct sk_buff *skb)
1666{
1667 struct hci_cp_read_tx_power *sent;
1668 struct hci_rp_read_tx_power *rp = (void *) skb->data;
1669 struct hci_conn *conn;
1670
1671 BT_DBG("%s status 0x%2.2x", hdev->name, rp->status);
1672
1673 if (rp->status)
1674 return;
1675
1676 sent = hci_sent_cmd_data(hdev, HCI_OP_READ_TX_POWER);
1677 if (!sent)
1678 return;
1679
1680 hci_dev_lock(hdev);
1681
1682 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(rp->handle));
1683 if (!conn)
1684 goto unlock;
1685
1686 switch (sent->type) {
1687 case 0x00:
1688 conn->tx_power = rp->tx_power;
1689 break;
1690 case 0x01:
1691 conn->max_tx_power = rp->tx_power;
1692 break;
1693 }
1694
1695unlock:
1696 hci_dev_unlock(hdev);
1697}
1698
1699static void hci_cc_write_ssp_debug_mode(struct hci_dev *hdev, struct sk_buff *skb)
1700{
1701 u8 status = *((u8 *) skb->data);
1702 u8 *mode;
1703
1704 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1705
1706 if (status)
1707 return;
1708
1709 mode = hci_sent_cmd_data(hdev, HCI_OP_WRITE_SSP_DEBUG_MODE);
1710 if (mode)
1711 hdev->ssp_debug_mode = *mode;
1712}
1713
1714static void hci_cs_inquiry(struct hci_dev *hdev, __u8 status)
1715{
1716 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1717
1718 if (status) {
1719 hci_conn_check_pending(hdev);
1720 return;
1721 }
1722
1723 set_bit(HCI_INQUIRY, &hdev->flags);
1724}
1725
1726static void hci_cs_create_conn(struct hci_dev *hdev, __u8 status)
1727{
1728 struct hci_cp_create_conn *cp;
1729 struct hci_conn *conn;
1730
1731 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1732
1733 cp = hci_sent_cmd_data(hdev, HCI_OP_CREATE_CONN);
1734 if (!cp)
1735 return;
1736
1737 hci_dev_lock(hdev);
1738
1739 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1740
1741 BT_DBG("%s bdaddr %pMR hcon %p", hdev->name, &cp->bdaddr, conn);
1742
1743 if (status) {
1744 if (conn && conn->state == BT_CONNECT) {
1745 if (status != 0x0c || conn->attempt > 2) {
1746 conn->state = BT_CLOSED;
1747 hci_connect_cfm(conn, status);
1748 hci_conn_del(conn);
1749 } else
1750 conn->state = BT_CONNECT2;
1751 }
1752 } else {
1753 if (!conn) {
1754 conn = hci_conn_add(hdev, ACL_LINK, &cp->bdaddr,
1755 HCI_ROLE_MASTER);
1756 if (!conn)
1757 bt_dev_err(hdev, "no memory for new connection");
1758 }
1759 }
1760
1761 hci_dev_unlock(hdev);
1762}
1763
1764static void hci_cs_add_sco(struct hci_dev *hdev, __u8 status)
1765{
1766 struct hci_cp_add_sco *cp;
1767 struct hci_conn *acl, *sco;
1768 __u16 handle;
1769
1770 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1771
1772 if (!status)
1773 return;
1774
1775 cp = hci_sent_cmd_data(hdev, HCI_OP_ADD_SCO);
1776 if (!cp)
1777 return;
1778
1779 handle = __le16_to_cpu(cp->handle);
1780
1781 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
1782
1783 hci_dev_lock(hdev);
1784
1785 acl = hci_conn_hash_lookup_handle(hdev, handle);
1786 if (acl) {
1787 sco = acl->link;
1788 if (sco) {
1789 sco->state = BT_CLOSED;
1790
1791 hci_connect_cfm(sco, status);
1792 hci_conn_del(sco);
1793 }
1794 }
1795
1796 hci_dev_unlock(hdev);
1797}
1798
1799static void hci_cs_auth_requested(struct hci_dev *hdev, __u8 status)
1800{
1801 struct hci_cp_auth_requested *cp;
1802 struct hci_conn *conn;
1803
1804 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1805
1806 if (!status)
1807 return;
1808
1809 cp = hci_sent_cmd_data(hdev, HCI_OP_AUTH_REQUESTED);
1810 if (!cp)
1811 return;
1812
1813 hci_dev_lock(hdev);
1814
1815 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1816 if (conn) {
1817 if (conn->state == BT_CONFIG) {
1818 hci_connect_cfm(conn, status);
1819 hci_conn_drop(conn);
1820 }
1821 }
1822
1823 hci_dev_unlock(hdev);
1824}
1825
1826static void hci_cs_set_conn_encrypt(struct hci_dev *hdev, __u8 status)
1827{
1828 struct hci_cp_set_conn_encrypt *cp;
1829 struct hci_conn *conn;
1830
1831 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1832
1833 if (!status)
1834 return;
1835
1836 cp = hci_sent_cmd_data(hdev, HCI_OP_SET_CONN_ENCRYPT);
1837 if (!cp)
1838 return;
1839
1840 hci_dev_lock(hdev);
1841
1842 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
1843 if (conn) {
1844 if (conn->state == BT_CONFIG) {
1845 hci_connect_cfm(conn, status);
1846 hci_conn_drop(conn);
1847 }
1848 }
1849
1850 hci_dev_unlock(hdev);
1851}
1852
1853static int hci_outgoing_auth_needed(struct hci_dev *hdev,
1854 struct hci_conn *conn)
1855{
1856 if (conn->state != BT_CONFIG || !conn->out)
1857 return 0;
1858
1859 if (conn->pending_sec_level == BT_SECURITY_SDP)
1860 return 0;
1861
1862 /* Only request authentication for SSP connections or non-SSP
1863 * devices with sec_level MEDIUM or HIGH or if MITM protection
1864 * is requested.
1865 */
1866 if (!hci_conn_ssp_enabled(conn) && !(conn->auth_type & 0x01) &&
1867 conn->pending_sec_level != BT_SECURITY_FIPS &&
1868 conn->pending_sec_level != BT_SECURITY_HIGH &&
1869 conn->pending_sec_level != BT_SECURITY_MEDIUM)
1870 return 0;
1871
1872 return 1;
1873}
1874
1875static int hci_resolve_name(struct hci_dev *hdev,
1876 struct inquiry_entry *e)
1877{
1878 struct hci_cp_remote_name_req cp;
1879
1880 memset(&cp, 0, sizeof(cp));
1881
1882 bacpy(&cp.bdaddr, &e->data.bdaddr);
1883 cp.pscan_rep_mode = e->data.pscan_rep_mode;
1884 cp.pscan_mode = e->data.pscan_mode;
1885 cp.clock_offset = e->data.clock_offset;
1886
1887 return hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
1888}
1889
1890static bool hci_resolve_next_name(struct hci_dev *hdev)
1891{
1892 struct discovery_state *discov = &hdev->discovery;
1893 struct inquiry_entry *e;
1894
1895 if (list_empty(&discov->resolve))
1896 return false;
1897
1898 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
1899 if (!e)
1900 return false;
1901
1902 if (hci_resolve_name(hdev, e) == 0) {
1903 e->name_state = NAME_PENDING;
1904 return true;
1905 }
1906
1907 return false;
1908}
1909
1910static void hci_check_pending_name(struct hci_dev *hdev, struct hci_conn *conn,
1911 bdaddr_t *bdaddr, u8 *name, u8 name_len)
1912{
1913 struct discovery_state *discov = &hdev->discovery;
1914 struct inquiry_entry *e;
1915
1916 /* Update the mgmt connected state if necessary. Be careful with
1917 * conn objects that exist but are not (yet) connected however.
1918 * Only those in BT_CONFIG or BT_CONNECTED states can be
1919 * considered connected.
1920 */
1921 if (conn &&
1922 (conn->state == BT_CONFIG || conn->state == BT_CONNECTED) &&
1923 !test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
1924 mgmt_device_connected(hdev, conn, 0, name, name_len);
1925
1926 if (discov->state == DISCOVERY_STOPPED)
1927 return;
1928
1929 if (discov->state == DISCOVERY_STOPPING)
1930 goto discov_complete;
1931
1932 if (discov->state != DISCOVERY_RESOLVING)
1933 return;
1934
1935 e = hci_inquiry_cache_lookup_resolve(hdev, bdaddr, NAME_PENDING);
1936 /* If the device was not found in a list of found devices names of which
1937 * are pending. there is no need to continue resolving a next name as it
1938 * will be done upon receiving another Remote Name Request Complete
1939 * Event */
1940 if (!e)
1941 return;
1942
1943 list_del(&e->list);
1944 if (name) {
1945 e->name_state = NAME_KNOWN;
1946 mgmt_remote_name(hdev, bdaddr, ACL_LINK, 0x00,
1947 e->data.rssi, name, name_len);
1948 } else {
1949 e->name_state = NAME_NOT_KNOWN;
1950 }
1951
1952 if (hci_resolve_next_name(hdev))
1953 return;
1954
1955discov_complete:
1956 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
1957}
1958
1959static void hci_cs_remote_name_req(struct hci_dev *hdev, __u8 status)
1960{
1961 struct hci_cp_remote_name_req *cp;
1962 struct hci_conn *conn;
1963
1964 BT_DBG("%s status 0x%2.2x", hdev->name, status);
1965
1966 /* If successful wait for the name req complete event before
1967 * checking for the need to do authentication */
1968 if (!status)
1969 return;
1970
1971 cp = hci_sent_cmd_data(hdev, HCI_OP_REMOTE_NAME_REQ);
1972 if (!cp)
1973 return;
1974
1975 hci_dev_lock(hdev);
1976
1977 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
1978
1979 if (hci_dev_test_flag(hdev, HCI_MGMT))
1980 hci_check_pending_name(hdev, conn, &cp->bdaddr, NULL, 0);
1981
1982 if (!conn)
1983 goto unlock;
1984
1985 if (!hci_outgoing_auth_needed(hdev, conn))
1986 goto unlock;
1987
1988 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
1989 struct hci_cp_auth_requested auth_cp;
1990
1991 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
1992
1993 auth_cp.handle = __cpu_to_le16(conn->handle);
1994 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED,
1995 sizeof(auth_cp), &auth_cp);
1996 }
1997
1998unlock:
1999 hci_dev_unlock(hdev);
2000}
2001
2002static void hci_cs_read_remote_features(struct hci_dev *hdev, __u8 status)
2003{
2004 struct hci_cp_read_remote_features *cp;
2005 struct hci_conn *conn;
2006
2007 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2008
2009 if (!status)
2010 return;
2011
2012 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_FEATURES);
2013 if (!cp)
2014 return;
2015
2016 hci_dev_lock(hdev);
2017
2018 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2019 if (conn) {
2020 if (conn->state == BT_CONFIG) {
2021 hci_connect_cfm(conn, status);
2022 hci_conn_drop(conn);
2023 }
2024 }
2025
2026 hci_dev_unlock(hdev);
2027}
2028
2029static void hci_cs_read_remote_ext_features(struct hci_dev *hdev, __u8 status)
2030{
2031 struct hci_cp_read_remote_ext_features *cp;
2032 struct hci_conn *conn;
2033
2034 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2035
2036 if (!status)
2037 return;
2038
2039 cp = hci_sent_cmd_data(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES);
2040 if (!cp)
2041 return;
2042
2043 hci_dev_lock(hdev);
2044
2045 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2046 if (conn) {
2047 if (conn->state == BT_CONFIG) {
2048 hci_connect_cfm(conn, status);
2049 hci_conn_drop(conn);
2050 }
2051 }
2052
2053 hci_dev_unlock(hdev);
2054}
2055
2056static void hci_cs_setup_sync_conn(struct hci_dev *hdev, __u8 status)
2057{
2058 struct hci_cp_setup_sync_conn *cp;
2059 struct hci_conn *acl, *sco;
2060 __u16 handle;
2061
2062 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2063
2064 if (!status)
2065 return;
2066
2067 cp = hci_sent_cmd_data(hdev, HCI_OP_SETUP_SYNC_CONN);
2068 if (!cp)
2069 return;
2070
2071 handle = __le16_to_cpu(cp->handle);
2072
2073 BT_DBG("%s handle 0x%4.4x", hdev->name, handle);
2074
2075 hci_dev_lock(hdev);
2076
2077 acl = hci_conn_hash_lookup_handle(hdev, handle);
2078 if (acl) {
2079 sco = acl->link;
2080 if (sco) {
2081 sco->state = BT_CLOSED;
2082
2083 hci_connect_cfm(sco, status);
2084 hci_conn_del(sco);
2085 }
2086 }
2087
2088 hci_dev_unlock(hdev);
2089}
2090
2091static void hci_cs_sniff_mode(struct hci_dev *hdev, __u8 status)
2092{
2093 struct hci_cp_sniff_mode *cp;
2094 struct hci_conn *conn;
2095
2096 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2097
2098 if (!status)
2099 return;
2100
2101 cp = hci_sent_cmd_data(hdev, HCI_OP_SNIFF_MODE);
2102 if (!cp)
2103 return;
2104
2105 hci_dev_lock(hdev);
2106
2107 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2108 if (conn) {
2109 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
2110
2111 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
2112 hci_sco_setup(conn, status);
2113 }
2114
2115 hci_dev_unlock(hdev);
2116}
2117
2118static void hci_cs_exit_sniff_mode(struct hci_dev *hdev, __u8 status)
2119{
2120 struct hci_cp_exit_sniff_mode *cp;
2121 struct hci_conn *conn;
2122
2123 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2124
2125 if (!status)
2126 return;
2127
2128 cp = hci_sent_cmd_data(hdev, HCI_OP_EXIT_SNIFF_MODE);
2129 if (!cp)
2130 return;
2131
2132 hci_dev_lock(hdev);
2133
2134 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2135 if (conn) {
2136 clear_bit(HCI_CONN_MODE_CHANGE_PEND, &conn->flags);
2137
2138 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
2139 hci_sco_setup(conn, status);
2140 }
2141
2142 hci_dev_unlock(hdev);
2143}
2144
2145static void hci_cs_disconnect(struct hci_dev *hdev, u8 status)
2146{
2147 struct hci_cp_disconnect *cp;
2148 struct hci_conn *conn;
2149
2150 if (!status)
2151 return;
2152
2153 cp = hci_sent_cmd_data(hdev, HCI_OP_DISCONNECT);
2154 if (!cp)
2155 return;
2156
2157 hci_dev_lock(hdev);
2158
2159 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2160 if (conn)
2161 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
2162 conn->dst_type, status);
2163
2164 hci_dev_unlock(hdev);
2165}
2166
2167static void cs_le_create_conn(struct hci_dev *hdev, bdaddr_t *peer_addr,
2168 u8 peer_addr_type, u8 own_address_type,
2169 u8 filter_policy)
2170{
2171 struct hci_conn *conn;
2172
2173 conn = hci_conn_hash_lookup_le(hdev, peer_addr,
2174 peer_addr_type);
2175 if (!conn)
2176 return;
2177
2178 /* Store the initiator and responder address information which
2179 * is needed for SMP. These values will not change during the
2180 * lifetime of the connection.
2181 */
2182 conn->init_addr_type = own_address_type;
2183 if (own_address_type == ADDR_LE_DEV_RANDOM)
2184 bacpy(&conn->init_addr, &hdev->random_addr);
2185 else
2186 bacpy(&conn->init_addr, &hdev->bdaddr);
2187
2188 conn->resp_addr_type = peer_addr_type;
2189 bacpy(&conn->resp_addr, peer_addr);
2190
2191 /* We don't want the connection attempt to stick around
2192 * indefinitely since LE doesn't have a page timeout concept
2193 * like BR/EDR. Set a timer for any connection that doesn't use
2194 * the white list for connecting.
2195 */
2196 if (filter_policy == HCI_LE_USE_PEER_ADDR)
2197 queue_delayed_work(conn->hdev->workqueue,
2198 &conn->le_conn_timeout,
2199 conn->conn_timeout);
2200}
2201
2202static void hci_cs_le_create_conn(struct hci_dev *hdev, u8 status)
2203{
2204 struct hci_cp_le_create_conn *cp;
2205
2206 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2207
2208 /* All connection failure handling is taken care of by the
2209 * hci_le_conn_failed function which is triggered by the HCI
2210 * request completion callbacks used for connecting.
2211 */
2212 if (status)
2213 return;
2214
2215 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_CREATE_CONN);
2216 if (!cp)
2217 return;
2218
2219 hci_dev_lock(hdev);
2220
2221 cs_le_create_conn(hdev, &cp->peer_addr, cp->peer_addr_type,
2222 cp->own_address_type, cp->filter_policy);
2223
2224 hci_dev_unlock(hdev);
2225}
2226
2227static void hci_cs_le_ext_create_conn(struct hci_dev *hdev, u8 status)
2228{
2229 struct hci_cp_le_ext_create_conn *cp;
2230
2231 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2232
2233 /* All connection failure handling is taken care of by the
2234 * hci_le_conn_failed function which is triggered by the HCI
2235 * request completion callbacks used for connecting.
2236 */
2237 if (status)
2238 return;
2239
2240 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_EXT_CREATE_CONN);
2241 if (!cp)
2242 return;
2243
2244 hci_dev_lock(hdev);
2245
2246 cs_le_create_conn(hdev, &cp->peer_addr, cp->peer_addr_type,
2247 cp->own_addr_type, cp->filter_policy);
2248
2249 hci_dev_unlock(hdev);
2250}
2251
2252static void hci_cs_le_read_remote_features(struct hci_dev *hdev, u8 status)
2253{
2254 struct hci_cp_le_read_remote_features *cp;
2255 struct hci_conn *conn;
2256
2257 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2258
2259 if (!status)
2260 return;
2261
2262 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_READ_REMOTE_FEATURES);
2263 if (!cp)
2264 return;
2265
2266 hci_dev_lock(hdev);
2267
2268 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2269 if (conn) {
2270 if (conn->state == BT_CONFIG) {
2271 hci_connect_cfm(conn, status);
2272 hci_conn_drop(conn);
2273 }
2274 }
2275
2276 hci_dev_unlock(hdev);
2277}
2278
2279static void hci_cs_le_start_enc(struct hci_dev *hdev, u8 status)
2280{
2281 struct hci_cp_le_start_enc *cp;
2282 struct hci_conn *conn;
2283
2284 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2285
2286 if (!status)
2287 return;
2288
2289 hci_dev_lock(hdev);
2290
2291 cp = hci_sent_cmd_data(hdev, HCI_OP_LE_START_ENC);
2292 if (!cp)
2293 goto unlock;
2294
2295 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(cp->handle));
2296 if (!conn)
2297 goto unlock;
2298
2299 if (conn->state != BT_CONNECTED)
2300 goto unlock;
2301
2302 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
2303 hci_conn_drop(conn);
2304
2305unlock:
2306 hci_dev_unlock(hdev);
2307}
2308
2309static void hci_cs_switch_role(struct hci_dev *hdev, u8 status)
2310{
2311 struct hci_cp_switch_role *cp;
2312 struct hci_conn *conn;
2313
2314 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2315
2316 if (!status)
2317 return;
2318
2319 cp = hci_sent_cmd_data(hdev, HCI_OP_SWITCH_ROLE);
2320 if (!cp)
2321 return;
2322
2323 hci_dev_lock(hdev);
2324
2325 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->bdaddr);
2326 if (conn)
2327 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
2328
2329 hci_dev_unlock(hdev);
2330}
2331
2332static void hci_inquiry_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2333{
2334 __u8 status = *((__u8 *) skb->data);
2335 struct discovery_state *discov = &hdev->discovery;
2336 struct inquiry_entry *e;
2337
2338 BT_DBG("%s status 0x%2.2x", hdev->name, status);
2339
2340 hci_conn_check_pending(hdev);
2341
2342 if (!test_and_clear_bit(HCI_INQUIRY, &hdev->flags))
2343 return;
2344
2345 smp_mb__after_atomic(); /* wake_up_bit advises about this barrier */
2346 wake_up_bit(&hdev->flags, HCI_INQUIRY);
2347
2348 if (!hci_dev_test_flag(hdev, HCI_MGMT))
2349 return;
2350
2351 hci_dev_lock(hdev);
2352
2353 if (discov->state != DISCOVERY_FINDING)
2354 goto unlock;
2355
2356 if (list_empty(&discov->resolve)) {
2357 /* When BR/EDR inquiry is active and no LE scanning is in
2358 * progress, then change discovery state to indicate completion.
2359 *
2360 * When running LE scanning and BR/EDR inquiry simultaneously
2361 * and the LE scan already finished, then change the discovery
2362 * state to indicate completion.
2363 */
2364 if (!hci_dev_test_flag(hdev, HCI_LE_SCAN) ||
2365 !test_bit(HCI_QUIRK_SIMULTANEOUS_DISCOVERY, &hdev->quirks))
2366 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2367 goto unlock;
2368 }
2369
2370 e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
2371 if (e && hci_resolve_name(hdev, e) == 0) {
2372 e->name_state = NAME_PENDING;
2373 hci_discovery_set_state(hdev, DISCOVERY_RESOLVING);
2374 } else {
2375 /* When BR/EDR inquiry is active and no LE scanning is in
2376 * progress, then change discovery state to indicate completion.
2377 *
2378 * When running LE scanning and BR/EDR inquiry simultaneously
2379 * and the LE scan already finished, then change the discovery
2380 * state to indicate completion.
2381 */
2382 if (!hci_dev_test_flag(hdev, HCI_LE_SCAN) ||
2383 !test_bit(HCI_QUIRK_SIMULTANEOUS_DISCOVERY, &hdev->quirks))
2384 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
2385 }
2386
2387unlock:
2388 hci_dev_unlock(hdev);
2389}
2390
2391static void hci_inquiry_result_evt(struct hci_dev *hdev, struct sk_buff *skb)
2392{
2393 struct inquiry_data data;
2394 struct inquiry_info *info = (void *) (skb->data + 1);
2395 int num_rsp = *((__u8 *) skb->data);
2396
2397 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
2398
2399 if (!num_rsp)
2400 return;
2401
2402 if (hci_dev_test_flag(hdev, HCI_PERIODIC_INQ))
2403 return;
2404
2405 hci_dev_lock(hdev);
2406
2407 for (; num_rsp; num_rsp--, info++) {
2408 u32 flags;
2409
2410 bacpy(&data.bdaddr, &info->bdaddr);
2411 data.pscan_rep_mode = info->pscan_rep_mode;
2412 data.pscan_period_mode = info->pscan_period_mode;
2413 data.pscan_mode = info->pscan_mode;
2414 memcpy(data.dev_class, info->dev_class, 3);
2415 data.clock_offset = info->clock_offset;
2416 data.rssi = HCI_RSSI_INVALID;
2417 data.ssp_mode = 0x00;
2418
2419 flags = hci_inquiry_cache_update(hdev, &data, false);
2420
2421 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
2422 info->dev_class, HCI_RSSI_INVALID,
2423 flags, NULL, 0, NULL, 0);
2424 }
2425
2426 hci_dev_unlock(hdev);
2427}
2428
2429static void hci_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2430{
2431 struct hci_ev_conn_complete *ev = (void *) skb->data;
2432 struct hci_conn *conn;
2433
2434 BT_DBG("%s", hdev->name);
2435
2436 hci_dev_lock(hdev);
2437
2438 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
2439 if (!conn) {
2440 if (ev->link_type != SCO_LINK)
2441 goto unlock;
2442
2443 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
2444 if (!conn)
2445 goto unlock;
2446
2447 conn->type = SCO_LINK;
2448 }
2449
2450 if (!ev->status) {
2451 conn->handle = __le16_to_cpu(ev->handle);
2452
2453 if (conn->type == ACL_LINK) {
2454 conn->state = BT_CONFIG;
2455 hci_conn_hold(conn);
2456
2457 if (!conn->out && !hci_conn_ssp_enabled(conn) &&
2458 !hci_find_link_key(hdev, &ev->bdaddr))
2459 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
2460 else
2461 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2462 } else
2463 conn->state = BT_CONNECTED;
2464
2465 hci_debugfs_create_conn(conn);
2466 hci_conn_add_sysfs(conn);
2467
2468 if (test_bit(HCI_AUTH, &hdev->flags))
2469 set_bit(HCI_CONN_AUTH, &conn->flags);
2470
2471 if (test_bit(HCI_ENCRYPT, &hdev->flags))
2472 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2473
2474 /* Get remote features */
2475 if (conn->type == ACL_LINK) {
2476 struct hci_cp_read_remote_features cp;
2477 cp.handle = ev->handle;
2478 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_FEATURES,
2479 sizeof(cp), &cp);
2480
2481 hci_req_update_scan(hdev);
2482 }
2483
2484 /* Set packet type for incoming connection */
2485 if (!conn->out && hdev->hci_ver < BLUETOOTH_VER_2_0) {
2486 struct hci_cp_change_conn_ptype cp;
2487 cp.handle = ev->handle;
2488 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2489 hci_send_cmd(hdev, HCI_OP_CHANGE_CONN_PTYPE, sizeof(cp),
2490 &cp);
2491 }
2492 } else {
2493 conn->state = BT_CLOSED;
2494 if (conn->type == ACL_LINK)
2495 mgmt_connect_failed(hdev, &conn->dst, conn->type,
2496 conn->dst_type, ev->status);
2497 }
2498
2499 if (conn->type == ACL_LINK)
2500 hci_sco_setup(conn, ev->status);
2501
2502 if (ev->status) {
2503 hci_connect_cfm(conn, ev->status);
2504 hci_conn_del(conn);
2505 } else if (ev->link_type != ACL_LINK)
2506 hci_connect_cfm(conn, ev->status);
2507
2508unlock:
2509 hci_dev_unlock(hdev);
2510
2511 hci_conn_check_pending(hdev);
2512}
2513
2514static void hci_reject_conn(struct hci_dev *hdev, bdaddr_t *bdaddr)
2515{
2516 struct hci_cp_reject_conn_req cp;
2517
2518 bacpy(&cp.bdaddr, bdaddr);
2519 cp.reason = HCI_ERROR_REJ_BAD_ADDR;
2520 hci_send_cmd(hdev, HCI_OP_REJECT_CONN_REQ, sizeof(cp), &cp);
2521}
2522
2523static void hci_conn_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
2524{
2525 struct hci_ev_conn_request *ev = (void *) skb->data;
2526 int mask = hdev->link_mode;
2527 struct inquiry_entry *ie;
2528 struct hci_conn *conn;
2529 __u8 flags = 0;
2530
2531 BT_DBG("%s bdaddr %pMR type 0x%x", hdev->name, &ev->bdaddr,
2532 ev->link_type);
2533
2534 mask |= hci_proto_connect_ind(hdev, &ev->bdaddr, ev->link_type,
2535 &flags);
2536
2537 if (!(mask & HCI_LM_ACCEPT)) {
2538 hci_reject_conn(hdev, &ev->bdaddr);
2539 return;
2540 }
2541
2542 if (hci_bdaddr_list_lookup(&hdev->blacklist, &ev->bdaddr,
2543 BDADDR_BREDR)) {
2544 hci_reject_conn(hdev, &ev->bdaddr);
2545 return;
2546 }
2547
2548 /* Require HCI_CONNECTABLE or a whitelist entry to accept the
2549 * connection. These features are only touched through mgmt so
2550 * only do the checks if HCI_MGMT is set.
2551 */
2552 if (hci_dev_test_flag(hdev, HCI_MGMT) &&
2553 !hci_dev_test_flag(hdev, HCI_CONNECTABLE) &&
2554 !hci_bdaddr_list_lookup(&hdev->whitelist, &ev->bdaddr,
2555 BDADDR_BREDR)) {
2556 hci_reject_conn(hdev, &ev->bdaddr);
2557 return;
2558 }
2559
2560 /* Connection accepted */
2561
2562 hci_dev_lock(hdev);
2563
2564 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
2565 if (ie)
2566 memcpy(ie->data.dev_class, ev->dev_class, 3);
2567
2568 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type,
2569 &ev->bdaddr);
2570 if (!conn) {
2571 conn = hci_conn_add(hdev, ev->link_type, &ev->bdaddr,
2572 HCI_ROLE_SLAVE);
2573 if (!conn) {
2574 bt_dev_err(hdev, "no memory for new connection");
2575 hci_dev_unlock(hdev);
2576 return;
2577 }
2578 }
2579
2580 memcpy(conn->dev_class, ev->dev_class, 3);
2581
2582 hci_dev_unlock(hdev);
2583
2584 if (ev->link_type == ACL_LINK ||
2585 (!(flags & HCI_PROTO_DEFER) && !lmp_esco_capable(hdev))) {
2586 struct hci_cp_accept_conn_req cp;
2587 conn->state = BT_CONNECT;
2588
2589 bacpy(&cp.bdaddr, &ev->bdaddr);
2590
2591 if (lmp_rswitch_capable(hdev) && (mask & HCI_LM_MASTER))
2592 cp.role = 0x00; /* Become master */
2593 else
2594 cp.role = 0x01; /* Remain slave */
2595
2596 hci_send_cmd(hdev, HCI_OP_ACCEPT_CONN_REQ, sizeof(cp), &cp);
2597 } else if (!(flags & HCI_PROTO_DEFER)) {
2598 struct hci_cp_accept_sync_conn_req cp;
2599 conn->state = BT_CONNECT;
2600
2601 bacpy(&cp.bdaddr, &ev->bdaddr);
2602 cp.pkt_type = cpu_to_le16(conn->pkt_type);
2603
2604 cp.tx_bandwidth = cpu_to_le32(0x00001f40);
2605 cp.rx_bandwidth = cpu_to_le32(0x00001f40);
2606 cp.max_latency = cpu_to_le16(0xffff);
2607 cp.content_format = cpu_to_le16(hdev->voice_setting);
2608 cp.retrans_effort = 0xff;
2609
2610 hci_send_cmd(hdev, HCI_OP_ACCEPT_SYNC_CONN_REQ, sizeof(cp),
2611 &cp);
2612 } else {
2613 conn->state = BT_CONNECT2;
2614 hci_connect_cfm(conn, 0);
2615 }
2616}
2617
2618static u8 hci_to_mgmt_reason(u8 err)
2619{
2620 switch (err) {
2621 case HCI_ERROR_CONNECTION_TIMEOUT:
2622 return MGMT_DEV_DISCONN_TIMEOUT;
2623 case HCI_ERROR_REMOTE_USER_TERM:
2624 case HCI_ERROR_REMOTE_LOW_RESOURCES:
2625 case HCI_ERROR_REMOTE_POWER_OFF:
2626 return MGMT_DEV_DISCONN_REMOTE;
2627 case HCI_ERROR_LOCAL_HOST_TERM:
2628 return MGMT_DEV_DISCONN_LOCAL_HOST;
2629 default:
2630 return MGMT_DEV_DISCONN_UNKNOWN;
2631 }
2632}
2633
2634static void hci_disconn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2635{
2636 struct hci_ev_disconn_complete *ev = (void *) skb->data;
2637 u8 reason;
2638 struct hci_conn_params *params;
2639 struct hci_conn *conn;
2640 bool mgmt_connected;
2641 u8 type;
2642
2643 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2644
2645 hci_dev_lock(hdev);
2646
2647 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2648 if (!conn)
2649 goto unlock;
2650
2651 if (ev->status) {
2652 mgmt_disconnect_failed(hdev, &conn->dst, conn->type,
2653 conn->dst_type, ev->status);
2654 goto unlock;
2655 }
2656
2657 conn->state = BT_CLOSED;
2658
2659 mgmt_connected = test_and_clear_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags);
2660
2661 if (test_bit(HCI_CONN_AUTH_FAILURE, &conn->flags))
2662 reason = MGMT_DEV_DISCONN_AUTH_FAILURE;
2663 else
2664 reason = hci_to_mgmt_reason(ev->reason);
2665
2666 mgmt_device_disconnected(hdev, &conn->dst, conn->type, conn->dst_type,
2667 reason, mgmt_connected);
2668
2669 if (conn->type == ACL_LINK) {
2670 if (test_bit(HCI_CONN_FLUSH_KEY, &conn->flags))
2671 hci_remove_link_key(hdev, &conn->dst);
2672
2673 hci_req_update_scan(hdev);
2674 }
2675
2676 params = hci_conn_params_lookup(hdev, &conn->dst, conn->dst_type);
2677 if (params) {
2678 switch (params->auto_connect) {
2679 case HCI_AUTO_CONN_LINK_LOSS:
2680 if (ev->reason != HCI_ERROR_CONNECTION_TIMEOUT)
2681 break;
2682 /* Fall through */
2683
2684 case HCI_AUTO_CONN_DIRECT:
2685 case HCI_AUTO_CONN_ALWAYS:
2686 list_del_init(¶ms->action);
2687 list_add(¶ms->action, &hdev->pend_le_conns);
2688 hci_update_background_scan(hdev);
2689 break;
2690
2691 default:
2692 break;
2693 }
2694 }
2695
2696 type = conn->type;
2697
2698 hci_disconn_cfm(conn, ev->reason);
2699 hci_conn_del(conn);
2700
2701 /* Re-enable advertising if necessary, since it might
2702 * have been disabled by the connection. From the
2703 * HCI_LE_Set_Advertise_Enable command description in
2704 * the core specification (v4.0):
2705 * "The Controller shall continue advertising until the Host
2706 * issues an LE_Set_Advertise_Enable command with
2707 * Advertising_Enable set to 0x00 (Advertising is disabled)
2708 * or until a connection is created or until the Advertising
2709 * is timed out due to Directed Advertising."
2710 */
2711 if (type == LE_LINK)
2712 hci_req_reenable_advertising(hdev);
2713
2714unlock:
2715 hci_dev_unlock(hdev);
2716}
2717
2718static void hci_auth_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
2719{
2720 struct hci_ev_auth_complete *ev = (void *) skb->data;
2721 struct hci_conn *conn;
2722
2723 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2724
2725 hci_dev_lock(hdev);
2726
2727 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2728 if (!conn)
2729 goto unlock;
2730
2731 if (!ev->status) {
2732 clear_bit(HCI_CONN_AUTH_FAILURE, &conn->flags);
2733
2734 if (!hci_conn_ssp_enabled(conn) &&
2735 test_bit(HCI_CONN_REAUTH_PEND, &conn->flags)) {
2736 bt_dev_info(hdev, "re-auth of legacy device is not possible.");
2737 } else {
2738 set_bit(HCI_CONN_AUTH, &conn->flags);
2739 conn->sec_level = conn->pending_sec_level;
2740 }
2741 } else {
2742 if (ev->status == HCI_ERROR_PIN_OR_KEY_MISSING)
2743 set_bit(HCI_CONN_AUTH_FAILURE, &conn->flags);
2744
2745 mgmt_auth_failed(conn, ev->status);
2746 }
2747
2748 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
2749 clear_bit(HCI_CONN_REAUTH_PEND, &conn->flags);
2750
2751 if (conn->state == BT_CONFIG) {
2752 if (!ev->status && hci_conn_ssp_enabled(conn)) {
2753 struct hci_cp_set_conn_encrypt cp;
2754 cp.handle = ev->handle;
2755 cp.encrypt = 0x01;
2756 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2757 &cp);
2758 } else {
2759 conn->state = BT_CONNECTED;
2760 hci_connect_cfm(conn, ev->status);
2761 hci_conn_drop(conn);
2762 }
2763 } else {
2764 hci_auth_cfm(conn, ev->status);
2765
2766 hci_conn_hold(conn);
2767 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
2768 hci_conn_drop(conn);
2769 }
2770
2771 if (test_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags)) {
2772 if (!ev->status) {
2773 struct hci_cp_set_conn_encrypt cp;
2774 cp.handle = ev->handle;
2775 cp.encrypt = 0x01;
2776 hci_send_cmd(hdev, HCI_OP_SET_CONN_ENCRYPT, sizeof(cp),
2777 &cp);
2778 } else {
2779 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2780 hci_encrypt_cfm(conn, ev->status, 0x00);
2781 }
2782 }
2783
2784unlock:
2785 hci_dev_unlock(hdev);
2786}
2787
2788static void hci_remote_name_evt(struct hci_dev *hdev, struct sk_buff *skb)
2789{
2790 struct hci_ev_remote_name *ev = (void *) skb->data;
2791 struct hci_conn *conn;
2792
2793 BT_DBG("%s", hdev->name);
2794
2795 hci_conn_check_pending(hdev);
2796
2797 hci_dev_lock(hdev);
2798
2799 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
2800
2801 if (!hci_dev_test_flag(hdev, HCI_MGMT))
2802 goto check_auth;
2803
2804 if (ev->status == 0)
2805 hci_check_pending_name(hdev, conn, &ev->bdaddr, ev->name,
2806 strnlen(ev->name, HCI_MAX_NAME_LENGTH));
2807 else
2808 hci_check_pending_name(hdev, conn, &ev->bdaddr, NULL, 0);
2809
2810check_auth:
2811 if (!conn)
2812 goto unlock;
2813
2814 if (!hci_outgoing_auth_needed(hdev, conn))
2815 goto unlock;
2816
2817 if (!test_and_set_bit(HCI_CONN_AUTH_PEND, &conn->flags)) {
2818 struct hci_cp_auth_requested cp;
2819
2820 set_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags);
2821
2822 cp.handle = __cpu_to_le16(conn->handle);
2823 hci_send_cmd(hdev, HCI_OP_AUTH_REQUESTED, sizeof(cp), &cp);
2824 }
2825
2826unlock:
2827 hci_dev_unlock(hdev);
2828}
2829
2830static void read_enc_key_size_complete(struct hci_dev *hdev, u8 status,
2831 u16 opcode, struct sk_buff *skb)
2832{
2833 const struct hci_rp_read_enc_key_size *rp;
2834 struct hci_conn *conn;
2835 u16 handle;
2836
2837 BT_DBG("%s status 0x%02x", hdev->name, status);
2838
2839 if (!skb || skb->len < sizeof(*rp)) {
2840 bt_dev_err(hdev, "invalid read key size response");
2841 return;
2842 }
2843
2844 rp = (void *)skb->data;
2845 handle = le16_to_cpu(rp->handle);
2846
2847 hci_dev_lock(hdev);
2848
2849 conn = hci_conn_hash_lookup_handle(hdev, handle);
2850 if (!conn)
2851 goto unlock;
2852
2853 /* If we fail to read the encryption key size, assume maximum
2854 * (which is the same we do also when this HCI command isn't
2855 * supported.
2856 */
2857 if (rp->status) {
2858 bt_dev_err(hdev, "failed to read key size for handle %u",
2859 handle);
2860 conn->enc_key_size = HCI_LINK_KEY_SIZE;
2861 } else {
2862 conn->enc_key_size = rp->key_size;
2863 }
2864
2865 if (conn->state == BT_CONFIG) {
2866 conn->state = BT_CONNECTED;
2867 hci_connect_cfm(conn, 0);
2868 hci_conn_drop(conn);
2869 } else {
2870 u8 encrypt;
2871
2872 if (!test_bit(HCI_CONN_ENCRYPT, &conn->flags))
2873 encrypt = 0x00;
2874 else if (test_bit(HCI_CONN_AES_CCM, &conn->flags))
2875 encrypt = 0x02;
2876 else
2877 encrypt = 0x01;
2878
2879 hci_encrypt_cfm(conn, 0, encrypt);
2880 }
2881
2882unlock:
2883 hci_dev_unlock(hdev);
2884}
2885
2886static void hci_encrypt_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
2887{
2888 struct hci_ev_encrypt_change *ev = (void *) skb->data;
2889 struct hci_conn *conn;
2890
2891 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2892
2893 hci_dev_lock(hdev);
2894
2895 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
2896 if (!conn)
2897 goto unlock;
2898
2899 if (!ev->status) {
2900 if (ev->encrypt) {
2901 /* Encryption implies authentication */
2902 set_bit(HCI_CONN_AUTH, &conn->flags);
2903 set_bit(HCI_CONN_ENCRYPT, &conn->flags);
2904 conn->sec_level = conn->pending_sec_level;
2905
2906 /* P-256 authentication key implies FIPS */
2907 if (conn->key_type == HCI_LK_AUTH_COMBINATION_P256)
2908 set_bit(HCI_CONN_FIPS, &conn->flags);
2909
2910 if ((conn->type == ACL_LINK && ev->encrypt == 0x02) ||
2911 conn->type == LE_LINK)
2912 set_bit(HCI_CONN_AES_CCM, &conn->flags);
2913 } else {
2914 clear_bit(HCI_CONN_ENCRYPT, &conn->flags);
2915 clear_bit(HCI_CONN_AES_CCM, &conn->flags);
2916 }
2917 }
2918
2919 /* We should disregard the current RPA and generate a new one
2920 * whenever the encryption procedure fails.
2921 */
2922 if (ev->status && conn->type == LE_LINK) {
2923 hci_dev_set_flag(hdev, HCI_RPA_EXPIRED);
2924 hci_adv_instances_set_rpa_expired(hdev, true);
2925 }
2926
2927 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
2928
2929 if (ev->status && conn->state == BT_CONNECTED) {
2930 if (ev->status == HCI_ERROR_PIN_OR_KEY_MISSING)
2931 set_bit(HCI_CONN_AUTH_FAILURE, &conn->flags);
2932
2933 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
2934 hci_conn_drop(conn);
2935 goto unlock;
2936 }
2937
2938 /* In Secure Connections Only mode, do not allow any connections
2939 * that are not encrypted with AES-CCM using a P-256 authenticated
2940 * combination key.
2941 */
2942 if (hci_dev_test_flag(hdev, HCI_SC_ONLY) &&
2943 (!test_bit(HCI_CONN_AES_CCM, &conn->flags) ||
2944 conn->key_type != HCI_LK_AUTH_COMBINATION_P256)) {
2945 hci_connect_cfm(conn, HCI_ERROR_AUTH_FAILURE);
2946 hci_conn_drop(conn);
2947 goto unlock;
2948 }
2949
2950 /* Try reading the encryption key size for encrypted ACL links */
2951 if (!ev->status && ev->encrypt && conn->type == ACL_LINK) {
2952 struct hci_cp_read_enc_key_size cp;
2953 struct hci_request req;
2954
2955 /* Only send HCI_Read_Encryption_Key_Size if the
2956 * controller really supports it. If it doesn't, assume
2957 * the default size (16).
2958 */
2959 if (!(hdev->commands[20] & 0x10)) {
2960 conn->enc_key_size = HCI_LINK_KEY_SIZE;
2961 goto notify;
2962 }
2963
2964 hci_req_init(&req, hdev);
2965
2966 cp.handle = cpu_to_le16(conn->handle);
2967 hci_req_add(&req, HCI_OP_READ_ENC_KEY_SIZE, sizeof(cp), &cp);
2968
2969 if (hci_req_run_skb(&req, read_enc_key_size_complete)) {
2970 bt_dev_err(hdev, "sending read key size failed");
2971 conn->enc_key_size = HCI_LINK_KEY_SIZE;
2972 goto notify;
2973 }
2974
2975 goto unlock;
2976 }
2977
2978notify:
2979 if (conn->state == BT_CONFIG) {
2980 if (!ev->status)
2981 conn->state = BT_CONNECTED;
2982
2983 hci_connect_cfm(conn, ev->status);
2984 hci_conn_drop(conn);
2985 } else
2986 hci_encrypt_cfm(conn, ev->status, ev->encrypt);
2987
2988unlock:
2989 hci_dev_unlock(hdev);
2990}
2991
2992static void hci_change_link_key_complete_evt(struct hci_dev *hdev,
2993 struct sk_buff *skb)
2994{
2995 struct hci_ev_change_link_key_complete *ev = (void *) skb->data;
2996 struct hci_conn *conn;
2997
2998 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
2999
3000 hci_dev_lock(hdev);
3001
3002 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3003 if (conn) {
3004 if (!ev->status)
3005 set_bit(HCI_CONN_SECURE, &conn->flags);
3006
3007 clear_bit(HCI_CONN_AUTH_PEND, &conn->flags);
3008
3009 hci_key_change_cfm(conn, ev->status);
3010 }
3011
3012 hci_dev_unlock(hdev);
3013}
3014
3015static void hci_remote_features_evt(struct hci_dev *hdev,
3016 struct sk_buff *skb)
3017{
3018 struct hci_ev_remote_features *ev = (void *) skb->data;
3019 struct hci_conn *conn;
3020
3021 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3022
3023 hci_dev_lock(hdev);
3024
3025 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3026 if (!conn)
3027 goto unlock;
3028
3029 if (!ev->status)
3030 memcpy(conn->features[0], ev->features, 8);
3031
3032 if (conn->state != BT_CONFIG)
3033 goto unlock;
3034
3035 if (!ev->status && lmp_ext_feat_capable(hdev) &&
3036 lmp_ext_feat_capable(conn)) {
3037 struct hci_cp_read_remote_ext_features cp;
3038 cp.handle = ev->handle;
3039 cp.page = 0x01;
3040 hci_send_cmd(hdev, HCI_OP_READ_REMOTE_EXT_FEATURES,
3041 sizeof(cp), &cp);
3042 goto unlock;
3043 }
3044
3045 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
3046 struct hci_cp_remote_name_req cp;
3047 memset(&cp, 0, sizeof(cp));
3048 bacpy(&cp.bdaddr, &conn->dst);
3049 cp.pscan_rep_mode = 0x02;
3050 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
3051 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
3052 mgmt_device_connected(hdev, conn, 0, NULL, 0);
3053
3054 if (!hci_outgoing_auth_needed(hdev, conn)) {
3055 conn->state = BT_CONNECTED;
3056 hci_connect_cfm(conn, ev->status);
3057 hci_conn_drop(conn);
3058 }
3059
3060unlock:
3061 hci_dev_unlock(hdev);
3062}
3063
3064static void hci_cmd_complete_evt(struct hci_dev *hdev, struct sk_buff *skb,
3065 u16 *opcode, u8 *status,
3066 hci_req_complete_t *req_complete,
3067 hci_req_complete_skb_t *req_complete_skb)
3068{
3069 struct hci_ev_cmd_complete *ev = (void *) skb->data;
3070
3071 *opcode = __le16_to_cpu(ev->opcode);
3072 *status = skb->data[sizeof(*ev)];
3073
3074 skb_pull(skb, sizeof(*ev));
3075
3076 switch (*opcode) {
3077 case HCI_OP_INQUIRY_CANCEL:
3078 hci_cc_inquiry_cancel(hdev, skb);
3079 break;
3080
3081 case HCI_OP_PERIODIC_INQ:
3082 hci_cc_periodic_inq(hdev, skb);
3083 break;
3084
3085 case HCI_OP_EXIT_PERIODIC_INQ:
3086 hci_cc_exit_periodic_inq(hdev, skb);
3087 break;
3088
3089 case HCI_OP_REMOTE_NAME_REQ_CANCEL:
3090 hci_cc_remote_name_req_cancel(hdev, skb);
3091 break;
3092
3093 case HCI_OP_ROLE_DISCOVERY:
3094 hci_cc_role_discovery(hdev, skb);
3095 break;
3096
3097 case HCI_OP_READ_LINK_POLICY:
3098 hci_cc_read_link_policy(hdev, skb);
3099 break;
3100
3101 case HCI_OP_WRITE_LINK_POLICY:
3102 hci_cc_write_link_policy(hdev, skb);
3103 break;
3104
3105 case HCI_OP_READ_DEF_LINK_POLICY:
3106 hci_cc_read_def_link_policy(hdev, skb);
3107 break;
3108
3109 case HCI_OP_WRITE_DEF_LINK_POLICY:
3110 hci_cc_write_def_link_policy(hdev, skb);
3111 break;
3112
3113 case HCI_OP_RESET:
3114 hci_cc_reset(hdev, skb);
3115 break;
3116
3117 case HCI_OP_READ_STORED_LINK_KEY:
3118 hci_cc_read_stored_link_key(hdev, skb);
3119 break;
3120
3121 case HCI_OP_DELETE_STORED_LINK_KEY:
3122 hci_cc_delete_stored_link_key(hdev, skb);
3123 break;
3124
3125 case HCI_OP_WRITE_LOCAL_NAME:
3126 hci_cc_write_local_name(hdev, skb);
3127 break;
3128
3129 case HCI_OP_READ_LOCAL_NAME:
3130 hci_cc_read_local_name(hdev, skb);
3131 break;
3132
3133 case HCI_OP_WRITE_AUTH_ENABLE:
3134 hci_cc_write_auth_enable(hdev, skb);
3135 break;
3136
3137 case HCI_OP_WRITE_ENCRYPT_MODE:
3138 hci_cc_write_encrypt_mode(hdev, skb);
3139 break;
3140
3141 case HCI_OP_WRITE_SCAN_ENABLE:
3142 hci_cc_write_scan_enable(hdev, skb);
3143 break;
3144
3145 case HCI_OP_READ_CLASS_OF_DEV:
3146 hci_cc_read_class_of_dev(hdev, skb);
3147 break;
3148
3149 case HCI_OP_WRITE_CLASS_OF_DEV:
3150 hci_cc_write_class_of_dev(hdev, skb);
3151 break;
3152
3153 case HCI_OP_READ_VOICE_SETTING:
3154 hci_cc_read_voice_setting(hdev, skb);
3155 break;
3156
3157 case HCI_OP_WRITE_VOICE_SETTING:
3158 hci_cc_write_voice_setting(hdev, skb);
3159 break;
3160
3161 case HCI_OP_READ_NUM_SUPPORTED_IAC:
3162 hci_cc_read_num_supported_iac(hdev, skb);
3163 break;
3164
3165 case HCI_OP_WRITE_SSP_MODE:
3166 hci_cc_write_ssp_mode(hdev, skb);
3167 break;
3168
3169 case HCI_OP_WRITE_SC_SUPPORT:
3170 hci_cc_write_sc_support(hdev, skb);
3171 break;
3172
3173 case HCI_OP_READ_LOCAL_VERSION:
3174 hci_cc_read_local_version(hdev, skb);
3175 break;
3176
3177 case HCI_OP_READ_LOCAL_COMMANDS:
3178 hci_cc_read_local_commands(hdev, skb);
3179 break;
3180
3181 case HCI_OP_READ_LOCAL_FEATURES:
3182 hci_cc_read_local_features(hdev, skb);
3183 break;
3184
3185 case HCI_OP_READ_LOCAL_EXT_FEATURES:
3186 hci_cc_read_local_ext_features(hdev, skb);
3187 break;
3188
3189 case HCI_OP_READ_BUFFER_SIZE:
3190 hci_cc_read_buffer_size(hdev, skb);
3191 break;
3192
3193 case HCI_OP_READ_BD_ADDR:
3194 hci_cc_read_bd_addr(hdev, skb);
3195 break;
3196
3197 case HCI_OP_READ_PAGE_SCAN_ACTIVITY:
3198 hci_cc_read_page_scan_activity(hdev, skb);
3199 break;
3200
3201 case HCI_OP_WRITE_PAGE_SCAN_ACTIVITY:
3202 hci_cc_write_page_scan_activity(hdev, skb);
3203 break;
3204
3205 case HCI_OP_READ_PAGE_SCAN_TYPE:
3206 hci_cc_read_page_scan_type(hdev, skb);
3207 break;
3208
3209 case HCI_OP_WRITE_PAGE_SCAN_TYPE:
3210 hci_cc_write_page_scan_type(hdev, skb);
3211 break;
3212
3213 case HCI_OP_READ_DATA_BLOCK_SIZE:
3214 hci_cc_read_data_block_size(hdev, skb);
3215 break;
3216
3217 case HCI_OP_READ_FLOW_CONTROL_MODE:
3218 hci_cc_read_flow_control_mode(hdev, skb);
3219 break;
3220
3221 case HCI_OP_READ_LOCAL_AMP_INFO:
3222 hci_cc_read_local_amp_info(hdev, skb);
3223 break;
3224
3225 case HCI_OP_READ_CLOCK:
3226 hci_cc_read_clock(hdev, skb);
3227 break;
3228
3229 case HCI_OP_READ_INQ_RSP_TX_POWER:
3230 hci_cc_read_inq_rsp_tx_power(hdev, skb);
3231 break;
3232
3233 case HCI_OP_PIN_CODE_REPLY:
3234 hci_cc_pin_code_reply(hdev, skb);
3235 break;
3236
3237 case HCI_OP_PIN_CODE_NEG_REPLY:
3238 hci_cc_pin_code_neg_reply(hdev, skb);
3239 break;
3240
3241 case HCI_OP_READ_LOCAL_OOB_DATA:
3242 hci_cc_read_local_oob_data(hdev, skb);
3243 break;
3244
3245 case HCI_OP_READ_LOCAL_OOB_EXT_DATA:
3246 hci_cc_read_local_oob_ext_data(hdev, skb);
3247 break;
3248
3249 case HCI_OP_LE_READ_BUFFER_SIZE:
3250 hci_cc_le_read_buffer_size(hdev, skb);
3251 break;
3252
3253 case HCI_OP_LE_READ_LOCAL_FEATURES:
3254 hci_cc_le_read_local_features(hdev, skb);
3255 break;
3256
3257 case HCI_OP_LE_READ_ADV_TX_POWER:
3258 hci_cc_le_read_adv_tx_power(hdev, skb);
3259 break;
3260
3261 case HCI_OP_USER_CONFIRM_REPLY:
3262 hci_cc_user_confirm_reply(hdev, skb);
3263 break;
3264
3265 case HCI_OP_USER_CONFIRM_NEG_REPLY:
3266 hci_cc_user_confirm_neg_reply(hdev, skb);
3267 break;
3268
3269 case HCI_OP_USER_PASSKEY_REPLY:
3270 hci_cc_user_passkey_reply(hdev, skb);
3271 break;
3272
3273 case HCI_OP_USER_PASSKEY_NEG_REPLY:
3274 hci_cc_user_passkey_neg_reply(hdev, skb);
3275 break;
3276
3277 case HCI_OP_LE_SET_RANDOM_ADDR:
3278 hci_cc_le_set_random_addr(hdev, skb);
3279 break;
3280
3281 case HCI_OP_LE_SET_ADV_ENABLE:
3282 hci_cc_le_set_adv_enable(hdev, skb);
3283 break;
3284
3285 case HCI_OP_LE_SET_SCAN_PARAM:
3286 hci_cc_le_set_scan_param(hdev, skb);
3287 break;
3288
3289 case HCI_OP_LE_SET_SCAN_ENABLE:
3290 hci_cc_le_set_scan_enable(hdev, skb);
3291 break;
3292
3293 case HCI_OP_LE_READ_WHITE_LIST_SIZE:
3294 hci_cc_le_read_white_list_size(hdev, skb);
3295 break;
3296
3297 case HCI_OP_LE_CLEAR_WHITE_LIST:
3298 hci_cc_le_clear_white_list(hdev, skb);
3299 break;
3300
3301 case HCI_OP_LE_ADD_TO_WHITE_LIST:
3302 hci_cc_le_add_to_white_list(hdev, skb);
3303 break;
3304
3305 case HCI_OP_LE_DEL_FROM_WHITE_LIST:
3306 hci_cc_le_del_from_white_list(hdev, skb);
3307 break;
3308
3309 case HCI_OP_LE_READ_SUPPORTED_STATES:
3310 hci_cc_le_read_supported_states(hdev, skb);
3311 break;
3312
3313 case HCI_OP_LE_READ_DEF_DATA_LEN:
3314 hci_cc_le_read_def_data_len(hdev, skb);
3315 break;
3316
3317 case HCI_OP_LE_WRITE_DEF_DATA_LEN:
3318 hci_cc_le_write_def_data_len(hdev, skb);
3319 break;
3320
3321 case HCI_OP_LE_ADD_TO_RESOLV_LIST:
3322 hci_cc_le_add_to_resolv_list(hdev, skb);
3323 break;
3324
3325 case HCI_OP_LE_DEL_FROM_RESOLV_LIST:
3326 hci_cc_le_del_from_resolv_list(hdev, skb);
3327 break;
3328
3329 case HCI_OP_LE_CLEAR_RESOLV_LIST:
3330 hci_cc_le_clear_resolv_list(hdev, skb);
3331 break;
3332
3333 case HCI_OP_LE_READ_RESOLV_LIST_SIZE:
3334 hci_cc_le_read_resolv_list_size(hdev, skb);
3335 break;
3336
3337 case HCI_OP_LE_SET_ADDR_RESOLV_ENABLE:
3338 hci_cc_le_set_addr_resolution_enable(hdev, skb);
3339 break;
3340
3341 case HCI_OP_LE_READ_MAX_DATA_LEN:
3342 hci_cc_le_read_max_data_len(hdev, skb);
3343 break;
3344
3345 case HCI_OP_WRITE_LE_HOST_SUPPORTED:
3346 hci_cc_write_le_host_supported(hdev, skb);
3347 break;
3348
3349 case HCI_OP_LE_SET_ADV_PARAM:
3350 hci_cc_set_adv_param(hdev, skb);
3351 break;
3352
3353 case HCI_OP_READ_RSSI:
3354 hci_cc_read_rssi(hdev, skb);
3355 break;
3356
3357 case HCI_OP_READ_TX_POWER:
3358 hci_cc_read_tx_power(hdev, skb);
3359 break;
3360
3361 case HCI_OP_WRITE_SSP_DEBUG_MODE:
3362 hci_cc_write_ssp_debug_mode(hdev, skb);
3363 break;
3364
3365 case HCI_OP_LE_SET_EXT_SCAN_PARAMS:
3366 hci_cc_le_set_ext_scan_param(hdev, skb);
3367 break;
3368
3369 case HCI_OP_LE_SET_EXT_SCAN_ENABLE:
3370 hci_cc_le_set_ext_scan_enable(hdev, skb);
3371 break;
3372
3373 case HCI_OP_LE_SET_DEFAULT_PHY:
3374 hci_cc_le_set_default_phy(hdev, skb);
3375 break;
3376
3377 case HCI_OP_LE_READ_NUM_SUPPORTED_ADV_SETS:
3378 hci_cc_le_read_num_adv_sets(hdev, skb);
3379 break;
3380
3381 case HCI_OP_LE_SET_EXT_ADV_PARAMS:
3382 hci_cc_set_ext_adv_param(hdev, skb);
3383 break;
3384
3385 case HCI_OP_LE_SET_EXT_ADV_ENABLE:
3386 hci_cc_le_set_ext_adv_enable(hdev, skb);
3387 break;
3388
3389 case HCI_OP_LE_SET_ADV_SET_RAND_ADDR:
3390 hci_cc_le_set_adv_set_random_addr(hdev, skb);
3391 break;
3392
3393 default:
3394 BT_DBG("%s opcode 0x%4.4x", hdev->name, *opcode);
3395 break;
3396 }
3397
3398 if (*opcode != HCI_OP_NOP)
3399 cancel_delayed_work(&hdev->cmd_timer);
3400
3401 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags))
3402 atomic_set(&hdev->cmd_cnt, 1);
3403
3404 hci_req_cmd_complete(hdev, *opcode, *status, req_complete,
3405 req_complete_skb);
3406
3407 if (hci_dev_test_flag(hdev, HCI_CMD_PENDING)) {
3408 bt_dev_err(hdev,
3409 "unexpected event for opcode 0x%4.4x", *opcode);
3410 return;
3411 }
3412
3413 if (atomic_read(&hdev->cmd_cnt) && !skb_queue_empty(&hdev->cmd_q))
3414 queue_work(hdev->workqueue, &hdev->cmd_work);
3415}
3416
3417static void hci_cmd_status_evt(struct hci_dev *hdev, struct sk_buff *skb,
3418 u16 *opcode, u8 *status,
3419 hci_req_complete_t *req_complete,
3420 hci_req_complete_skb_t *req_complete_skb)
3421{
3422 struct hci_ev_cmd_status *ev = (void *) skb->data;
3423
3424 skb_pull(skb, sizeof(*ev));
3425
3426 *opcode = __le16_to_cpu(ev->opcode);
3427 *status = ev->status;
3428
3429 switch (*opcode) {
3430 case HCI_OP_INQUIRY:
3431 hci_cs_inquiry(hdev, ev->status);
3432 break;
3433
3434 case HCI_OP_CREATE_CONN:
3435 hci_cs_create_conn(hdev, ev->status);
3436 break;
3437
3438 case HCI_OP_DISCONNECT:
3439 hci_cs_disconnect(hdev, ev->status);
3440 break;
3441
3442 case HCI_OP_ADD_SCO:
3443 hci_cs_add_sco(hdev, ev->status);
3444 break;
3445
3446 case HCI_OP_AUTH_REQUESTED:
3447 hci_cs_auth_requested(hdev, ev->status);
3448 break;
3449
3450 case HCI_OP_SET_CONN_ENCRYPT:
3451 hci_cs_set_conn_encrypt(hdev, ev->status);
3452 break;
3453
3454 case HCI_OP_REMOTE_NAME_REQ:
3455 hci_cs_remote_name_req(hdev, ev->status);
3456 break;
3457
3458 case HCI_OP_READ_REMOTE_FEATURES:
3459 hci_cs_read_remote_features(hdev, ev->status);
3460 break;
3461
3462 case HCI_OP_READ_REMOTE_EXT_FEATURES:
3463 hci_cs_read_remote_ext_features(hdev, ev->status);
3464 break;
3465
3466 case HCI_OP_SETUP_SYNC_CONN:
3467 hci_cs_setup_sync_conn(hdev, ev->status);
3468 break;
3469
3470 case HCI_OP_SNIFF_MODE:
3471 hci_cs_sniff_mode(hdev, ev->status);
3472 break;
3473
3474 case HCI_OP_EXIT_SNIFF_MODE:
3475 hci_cs_exit_sniff_mode(hdev, ev->status);
3476 break;
3477
3478 case HCI_OP_SWITCH_ROLE:
3479 hci_cs_switch_role(hdev, ev->status);
3480 break;
3481
3482 case HCI_OP_LE_CREATE_CONN:
3483 hci_cs_le_create_conn(hdev, ev->status);
3484 break;
3485
3486 case HCI_OP_LE_READ_REMOTE_FEATURES:
3487 hci_cs_le_read_remote_features(hdev, ev->status);
3488 break;
3489
3490 case HCI_OP_LE_START_ENC:
3491 hci_cs_le_start_enc(hdev, ev->status);
3492 break;
3493
3494 case HCI_OP_LE_EXT_CREATE_CONN:
3495 hci_cs_le_ext_create_conn(hdev, ev->status);
3496 break;
3497
3498 default:
3499 BT_DBG("%s opcode 0x%4.4x", hdev->name, *opcode);
3500 break;
3501 }
3502
3503 if (*opcode != HCI_OP_NOP)
3504 cancel_delayed_work(&hdev->cmd_timer);
3505
3506 if (ev->ncmd && !test_bit(HCI_RESET, &hdev->flags))
3507 atomic_set(&hdev->cmd_cnt, 1);
3508
3509 /* Indicate request completion if the command failed. Also, if
3510 * we're not waiting for a special event and we get a success
3511 * command status we should try to flag the request as completed
3512 * (since for this kind of commands there will not be a command
3513 * complete event).
3514 */
3515 if (ev->status ||
3516 (hdev->sent_cmd && !bt_cb(hdev->sent_cmd)->hci.req_event))
3517 hci_req_cmd_complete(hdev, *opcode, ev->status, req_complete,
3518 req_complete_skb);
3519
3520 if (hci_dev_test_flag(hdev, HCI_CMD_PENDING)) {
3521 bt_dev_err(hdev,
3522 "unexpected event for opcode 0x%4.4x", *opcode);
3523 return;
3524 }
3525
3526 if (atomic_read(&hdev->cmd_cnt) && !skb_queue_empty(&hdev->cmd_q))
3527 queue_work(hdev->workqueue, &hdev->cmd_work);
3528}
3529
3530static void hci_hardware_error_evt(struct hci_dev *hdev, struct sk_buff *skb)
3531{
3532 struct hci_ev_hardware_error *ev = (void *) skb->data;
3533
3534 hdev->hw_error_code = ev->code;
3535
3536 queue_work(hdev->req_workqueue, &hdev->error_reset);
3537}
3538
3539static void hci_role_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3540{
3541 struct hci_ev_role_change *ev = (void *) skb->data;
3542 struct hci_conn *conn;
3543
3544 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3545
3546 hci_dev_lock(hdev);
3547
3548 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3549 if (conn) {
3550 if (!ev->status)
3551 conn->role = ev->role;
3552
3553 clear_bit(HCI_CONN_RSWITCH_PEND, &conn->flags);
3554
3555 hci_role_switch_cfm(conn, ev->status, ev->role);
3556 }
3557
3558 hci_dev_unlock(hdev);
3559}
3560
3561static void hci_num_comp_pkts_evt(struct hci_dev *hdev, struct sk_buff *skb)
3562{
3563 struct hci_ev_num_comp_pkts *ev = (void *) skb->data;
3564 int i;
3565
3566 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_PACKET_BASED) {
3567 bt_dev_err(hdev, "wrong event for mode %d", hdev->flow_ctl_mode);
3568 return;
3569 }
3570
3571 if (skb->len < sizeof(*ev) ||
3572 skb->len < struct_size(ev, handles, ev->num_hndl)) {
3573 BT_DBG("%s bad parameters", hdev->name);
3574 return;
3575 }
3576
3577 BT_DBG("%s num_hndl %d", hdev->name, ev->num_hndl);
3578
3579 for (i = 0; i < ev->num_hndl; i++) {
3580 struct hci_comp_pkts_info *info = &ev->handles[i];
3581 struct hci_conn *conn;
3582 __u16 handle, count;
3583
3584 handle = __le16_to_cpu(info->handle);
3585 count = __le16_to_cpu(info->count);
3586
3587 conn = hci_conn_hash_lookup_handle(hdev, handle);
3588 if (!conn)
3589 continue;
3590
3591 conn->sent -= count;
3592
3593 switch (conn->type) {
3594 case ACL_LINK:
3595 hdev->acl_cnt += count;
3596 if (hdev->acl_cnt > hdev->acl_pkts)
3597 hdev->acl_cnt = hdev->acl_pkts;
3598 break;
3599
3600 case LE_LINK:
3601 if (hdev->le_pkts) {
3602 hdev->le_cnt += count;
3603 if (hdev->le_cnt > hdev->le_pkts)
3604 hdev->le_cnt = hdev->le_pkts;
3605 } else {
3606 hdev->acl_cnt += count;
3607 if (hdev->acl_cnt > hdev->acl_pkts)
3608 hdev->acl_cnt = hdev->acl_pkts;
3609 }
3610 break;
3611
3612 case SCO_LINK:
3613 hdev->sco_cnt += count;
3614 if (hdev->sco_cnt > hdev->sco_pkts)
3615 hdev->sco_cnt = hdev->sco_pkts;
3616 break;
3617
3618 default:
3619 bt_dev_err(hdev, "unknown type %d conn %p",
3620 conn->type, conn);
3621 break;
3622 }
3623 }
3624
3625 queue_work(hdev->workqueue, &hdev->tx_work);
3626}
3627
3628static struct hci_conn *__hci_conn_lookup_handle(struct hci_dev *hdev,
3629 __u16 handle)
3630{
3631 struct hci_chan *chan;
3632
3633 switch (hdev->dev_type) {
3634 case HCI_PRIMARY:
3635 return hci_conn_hash_lookup_handle(hdev, handle);
3636 case HCI_AMP:
3637 chan = hci_chan_lookup_handle(hdev, handle);
3638 if (chan)
3639 return chan->conn;
3640 break;
3641 default:
3642 bt_dev_err(hdev, "unknown dev_type %d", hdev->dev_type);
3643 break;
3644 }
3645
3646 return NULL;
3647}
3648
3649static void hci_num_comp_blocks_evt(struct hci_dev *hdev, struct sk_buff *skb)
3650{
3651 struct hci_ev_num_comp_blocks *ev = (void *) skb->data;
3652 int i;
3653
3654 if (hdev->flow_ctl_mode != HCI_FLOW_CTL_MODE_BLOCK_BASED) {
3655 bt_dev_err(hdev, "wrong event for mode %d", hdev->flow_ctl_mode);
3656 return;
3657 }
3658
3659 if (skb->len < sizeof(*ev) ||
3660 skb->len < struct_size(ev, handles, ev->num_hndl)) {
3661 BT_DBG("%s bad parameters", hdev->name);
3662 return;
3663 }
3664
3665 BT_DBG("%s num_blocks %d num_hndl %d", hdev->name, ev->num_blocks,
3666 ev->num_hndl);
3667
3668 for (i = 0; i < ev->num_hndl; i++) {
3669 struct hci_comp_blocks_info *info = &ev->handles[i];
3670 struct hci_conn *conn = NULL;
3671 __u16 handle, block_count;
3672
3673 handle = __le16_to_cpu(info->handle);
3674 block_count = __le16_to_cpu(info->blocks);
3675
3676 conn = __hci_conn_lookup_handle(hdev, handle);
3677 if (!conn)
3678 continue;
3679
3680 conn->sent -= block_count;
3681
3682 switch (conn->type) {
3683 case ACL_LINK:
3684 case AMP_LINK:
3685 hdev->block_cnt += block_count;
3686 if (hdev->block_cnt > hdev->num_blocks)
3687 hdev->block_cnt = hdev->num_blocks;
3688 break;
3689
3690 default:
3691 bt_dev_err(hdev, "unknown type %d conn %p",
3692 conn->type, conn);
3693 break;
3694 }
3695 }
3696
3697 queue_work(hdev->workqueue, &hdev->tx_work);
3698}
3699
3700static void hci_mode_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3701{
3702 struct hci_ev_mode_change *ev = (void *) skb->data;
3703 struct hci_conn *conn;
3704
3705 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3706
3707 hci_dev_lock(hdev);
3708
3709 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3710 if (conn) {
3711 conn->mode = ev->mode;
3712
3713 if (!test_and_clear_bit(HCI_CONN_MODE_CHANGE_PEND,
3714 &conn->flags)) {
3715 if (conn->mode == HCI_CM_ACTIVE)
3716 set_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3717 else
3718 clear_bit(HCI_CONN_POWER_SAVE, &conn->flags);
3719 }
3720
3721 if (test_and_clear_bit(HCI_CONN_SCO_SETUP_PEND, &conn->flags))
3722 hci_sco_setup(conn, ev->status);
3723 }
3724
3725 hci_dev_unlock(hdev);
3726}
3727
3728static void hci_pin_code_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3729{
3730 struct hci_ev_pin_code_req *ev = (void *) skb->data;
3731 struct hci_conn *conn;
3732
3733 BT_DBG("%s", hdev->name);
3734
3735 hci_dev_lock(hdev);
3736
3737 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3738 if (!conn)
3739 goto unlock;
3740
3741 if (conn->state == BT_CONNECTED) {
3742 hci_conn_hold(conn);
3743 conn->disc_timeout = HCI_PAIRING_TIMEOUT;
3744 hci_conn_drop(conn);
3745 }
3746
3747 if (!hci_dev_test_flag(hdev, HCI_BONDABLE) &&
3748 !test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags)) {
3749 hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY,
3750 sizeof(ev->bdaddr), &ev->bdaddr);
3751 } else if (hci_dev_test_flag(hdev, HCI_MGMT)) {
3752 u8 secure;
3753
3754 if (conn->pending_sec_level == BT_SECURITY_HIGH)
3755 secure = 1;
3756 else
3757 secure = 0;
3758
3759 mgmt_pin_code_request(hdev, &ev->bdaddr, secure);
3760 }
3761
3762unlock:
3763 hci_dev_unlock(hdev);
3764}
3765
3766static void conn_set_key(struct hci_conn *conn, u8 key_type, u8 pin_len)
3767{
3768 if (key_type == HCI_LK_CHANGED_COMBINATION)
3769 return;
3770
3771 conn->pin_length = pin_len;
3772 conn->key_type = key_type;
3773
3774 switch (key_type) {
3775 case HCI_LK_LOCAL_UNIT:
3776 case HCI_LK_REMOTE_UNIT:
3777 case HCI_LK_DEBUG_COMBINATION:
3778 return;
3779 case HCI_LK_COMBINATION:
3780 if (pin_len == 16)
3781 conn->pending_sec_level = BT_SECURITY_HIGH;
3782 else
3783 conn->pending_sec_level = BT_SECURITY_MEDIUM;
3784 break;
3785 case HCI_LK_UNAUTH_COMBINATION_P192:
3786 case HCI_LK_UNAUTH_COMBINATION_P256:
3787 conn->pending_sec_level = BT_SECURITY_MEDIUM;
3788 break;
3789 case HCI_LK_AUTH_COMBINATION_P192:
3790 conn->pending_sec_level = BT_SECURITY_HIGH;
3791 break;
3792 case HCI_LK_AUTH_COMBINATION_P256:
3793 conn->pending_sec_level = BT_SECURITY_FIPS;
3794 break;
3795 }
3796}
3797
3798static void hci_link_key_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
3799{
3800 struct hci_ev_link_key_req *ev = (void *) skb->data;
3801 struct hci_cp_link_key_reply cp;
3802 struct hci_conn *conn;
3803 struct link_key *key;
3804
3805 BT_DBG("%s", hdev->name);
3806
3807 if (!hci_dev_test_flag(hdev, HCI_MGMT))
3808 return;
3809
3810 hci_dev_lock(hdev);
3811
3812 key = hci_find_link_key(hdev, &ev->bdaddr);
3813 if (!key) {
3814 BT_DBG("%s link key not found for %pMR", hdev->name,
3815 &ev->bdaddr);
3816 goto not_found;
3817 }
3818
3819 BT_DBG("%s found key type %u for %pMR", hdev->name, key->type,
3820 &ev->bdaddr);
3821
3822 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3823 if (conn) {
3824 clear_bit(HCI_CONN_NEW_LINK_KEY, &conn->flags);
3825
3826 if ((key->type == HCI_LK_UNAUTH_COMBINATION_P192 ||
3827 key->type == HCI_LK_UNAUTH_COMBINATION_P256) &&
3828 conn->auth_type != 0xff && (conn->auth_type & 0x01)) {
3829 BT_DBG("%s ignoring unauthenticated key", hdev->name);
3830 goto not_found;
3831 }
3832
3833 if (key->type == HCI_LK_COMBINATION && key->pin_len < 16 &&
3834 (conn->pending_sec_level == BT_SECURITY_HIGH ||
3835 conn->pending_sec_level == BT_SECURITY_FIPS)) {
3836 BT_DBG("%s ignoring key unauthenticated for high security",
3837 hdev->name);
3838 goto not_found;
3839 }
3840
3841 conn_set_key(conn, key->type, key->pin_len);
3842 }
3843
3844 bacpy(&cp.bdaddr, &ev->bdaddr);
3845 memcpy(cp.link_key, key->val, HCI_LINK_KEY_SIZE);
3846
3847 hci_send_cmd(hdev, HCI_OP_LINK_KEY_REPLY, sizeof(cp), &cp);
3848
3849 hci_dev_unlock(hdev);
3850
3851 return;
3852
3853not_found:
3854 hci_send_cmd(hdev, HCI_OP_LINK_KEY_NEG_REPLY, 6, &ev->bdaddr);
3855 hci_dev_unlock(hdev);
3856}
3857
3858static void hci_link_key_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
3859{
3860 struct hci_ev_link_key_notify *ev = (void *) skb->data;
3861 struct hci_conn *conn;
3862 struct link_key *key;
3863 bool persistent;
3864 u8 pin_len = 0;
3865
3866 BT_DBG("%s", hdev->name);
3867
3868 hci_dev_lock(hdev);
3869
3870 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
3871 if (!conn)
3872 goto unlock;
3873
3874 hci_conn_hold(conn);
3875 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
3876 hci_conn_drop(conn);
3877
3878 set_bit(HCI_CONN_NEW_LINK_KEY, &conn->flags);
3879 conn_set_key(conn, ev->key_type, conn->pin_length);
3880
3881 if (!hci_dev_test_flag(hdev, HCI_MGMT))
3882 goto unlock;
3883
3884 key = hci_add_link_key(hdev, conn, &ev->bdaddr, ev->link_key,
3885 ev->key_type, pin_len, &persistent);
3886 if (!key)
3887 goto unlock;
3888
3889 /* Update connection information since adding the key will have
3890 * fixed up the type in the case of changed combination keys.
3891 */
3892 if (ev->key_type == HCI_LK_CHANGED_COMBINATION)
3893 conn_set_key(conn, key->type, key->pin_len);
3894
3895 mgmt_new_link_key(hdev, key, persistent);
3896
3897 /* Keep debug keys around only if the HCI_KEEP_DEBUG_KEYS flag
3898 * is set. If it's not set simply remove the key from the kernel
3899 * list (we've still notified user space about it but with
3900 * store_hint being 0).
3901 */
3902 if (key->type == HCI_LK_DEBUG_COMBINATION &&
3903 !hci_dev_test_flag(hdev, HCI_KEEP_DEBUG_KEYS)) {
3904 list_del_rcu(&key->list);
3905 kfree_rcu(key, rcu);
3906 goto unlock;
3907 }
3908
3909 if (persistent)
3910 clear_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3911 else
3912 set_bit(HCI_CONN_FLUSH_KEY, &conn->flags);
3913
3914unlock:
3915 hci_dev_unlock(hdev);
3916}
3917
3918static void hci_clock_offset_evt(struct hci_dev *hdev, struct sk_buff *skb)
3919{
3920 struct hci_ev_clock_offset *ev = (void *) skb->data;
3921 struct hci_conn *conn;
3922
3923 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3924
3925 hci_dev_lock(hdev);
3926
3927 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3928 if (conn && !ev->status) {
3929 struct inquiry_entry *ie;
3930
3931 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
3932 if (ie) {
3933 ie->data.clock_offset = ev->clock_offset;
3934 ie->timestamp = jiffies;
3935 }
3936 }
3937
3938 hci_dev_unlock(hdev);
3939}
3940
3941static void hci_pkt_type_change_evt(struct hci_dev *hdev, struct sk_buff *skb)
3942{
3943 struct hci_ev_pkt_type_change *ev = (void *) skb->data;
3944 struct hci_conn *conn;
3945
3946 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
3947
3948 hci_dev_lock(hdev);
3949
3950 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
3951 if (conn && !ev->status)
3952 conn->pkt_type = __le16_to_cpu(ev->pkt_type);
3953
3954 hci_dev_unlock(hdev);
3955}
3956
3957static void hci_pscan_rep_mode_evt(struct hci_dev *hdev, struct sk_buff *skb)
3958{
3959 struct hci_ev_pscan_rep_mode *ev = (void *) skb->data;
3960 struct inquiry_entry *ie;
3961
3962 BT_DBG("%s", hdev->name);
3963
3964 hci_dev_lock(hdev);
3965
3966 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
3967 if (ie) {
3968 ie->data.pscan_rep_mode = ev->pscan_rep_mode;
3969 ie->timestamp = jiffies;
3970 }
3971
3972 hci_dev_unlock(hdev);
3973}
3974
3975static void hci_inquiry_result_with_rssi_evt(struct hci_dev *hdev,
3976 struct sk_buff *skb)
3977{
3978 struct inquiry_data data;
3979 int num_rsp = *((__u8 *) skb->data);
3980
3981 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
3982
3983 if (!num_rsp)
3984 return;
3985
3986 if (hci_dev_test_flag(hdev, HCI_PERIODIC_INQ))
3987 return;
3988
3989 hci_dev_lock(hdev);
3990
3991 if ((skb->len - 1) / num_rsp != sizeof(struct inquiry_info_with_rssi)) {
3992 struct inquiry_info_with_rssi_and_pscan_mode *info;
3993 info = (void *) (skb->data + 1);
3994
3995 for (; num_rsp; num_rsp--, info++) {
3996 u32 flags;
3997
3998 bacpy(&data.bdaddr, &info->bdaddr);
3999 data.pscan_rep_mode = info->pscan_rep_mode;
4000 data.pscan_period_mode = info->pscan_period_mode;
4001 data.pscan_mode = info->pscan_mode;
4002 memcpy(data.dev_class, info->dev_class, 3);
4003 data.clock_offset = info->clock_offset;
4004 data.rssi = info->rssi;
4005 data.ssp_mode = 0x00;
4006
4007 flags = hci_inquiry_cache_update(hdev, &data, false);
4008
4009 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
4010 info->dev_class, info->rssi,
4011 flags, NULL, 0, NULL, 0);
4012 }
4013 } else {
4014 struct inquiry_info_with_rssi *info = (void *) (skb->data + 1);
4015
4016 for (; num_rsp; num_rsp--, info++) {
4017 u32 flags;
4018
4019 bacpy(&data.bdaddr, &info->bdaddr);
4020 data.pscan_rep_mode = info->pscan_rep_mode;
4021 data.pscan_period_mode = info->pscan_period_mode;
4022 data.pscan_mode = 0x00;
4023 memcpy(data.dev_class, info->dev_class, 3);
4024 data.clock_offset = info->clock_offset;
4025 data.rssi = info->rssi;
4026 data.ssp_mode = 0x00;
4027
4028 flags = hci_inquiry_cache_update(hdev, &data, false);
4029
4030 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
4031 info->dev_class, info->rssi,
4032 flags, NULL, 0, NULL, 0);
4033 }
4034 }
4035
4036 hci_dev_unlock(hdev);
4037}
4038
4039static void hci_remote_ext_features_evt(struct hci_dev *hdev,
4040 struct sk_buff *skb)
4041{
4042 struct hci_ev_remote_ext_features *ev = (void *) skb->data;
4043 struct hci_conn *conn;
4044
4045 BT_DBG("%s", hdev->name);
4046
4047 hci_dev_lock(hdev);
4048
4049 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4050 if (!conn)
4051 goto unlock;
4052
4053 if (ev->page < HCI_MAX_PAGES)
4054 memcpy(conn->features[ev->page], ev->features, 8);
4055
4056 if (!ev->status && ev->page == 0x01) {
4057 struct inquiry_entry *ie;
4058
4059 ie = hci_inquiry_cache_lookup(hdev, &conn->dst);
4060 if (ie)
4061 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
4062
4063 if (ev->features[0] & LMP_HOST_SSP) {
4064 set_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
4065 } else {
4066 /* It is mandatory by the Bluetooth specification that
4067 * Extended Inquiry Results are only used when Secure
4068 * Simple Pairing is enabled, but some devices violate
4069 * this.
4070 *
4071 * To make these devices work, the internal SSP
4072 * enabled flag needs to be cleared if the remote host
4073 * features do not indicate SSP support */
4074 clear_bit(HCI_CONN_SSP_ENABLED, &conn->flags);
4075 }
4076
4077 if (ev->features[0] & LMP_HOST_SC)
4078 set_bit(HCI_CONN_SC_ENABLED, &conn->flags);
4079 }
4080
4081 if (conn->state != BT_CONFIG)
4082 goto unlock;
4083
4084 if (!ev->status && !test_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags)) {
4085 struct hci_cp_remote_name_req cp;
4086 memset(&cp, 0, sizeof(cp));
4087 bacpy(&cp.bdaddr, &conn->dst);
4088 cp.pscan_rep_mode = 0x02;
4089 hci_send_cmd(hdev, HCI_OP_REMOTE_NAME_REQ, sizeof(cp), &cp);
4090 } else if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
4091 mgmt_device_connected(hdev, conn, 0, NULL, 0);
4092
4093 if (!hci_outgoing_auth_needed(hdev, conn)) {
4094 conn->state = BT_CONNECTED;
4095 hci_connect_cfm(conn, ev->status);
4096 hci_conn_drop(conn);
4097 }
4098
4099unlock:
4100 hci_dev_unlock(hdev);
4101}
4102
4103static void hci_sync_conn_complete_evt(struct hci_dev *hdev,
4104 struct sk_buff *skb)
4105{
4106 struct hci_ev_sync_conn_complete *ev = (void *) skb->data;
4107 struct hci_conn *conn;
4108
4109 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4110
4111 hci_dev_lock(hdev);
4112
4113 conn = hci_conn_hash_lookup_ba(hdev, ev->link_type, &ev->bdaddr);
4114 if (!conn) {
4115 if (ev->link_type == ESCO_LINK)
4116 goto unlock;
4117
4118 /* When the link type in the event indicates SCO connection
4119 * and lookup of the connection object fails, then check
4120 * if an eSCO connection object exists.
4121 *
4122 * The core limits the synchronous connections to either
4123 * SCO or eSCO. The eSCO connection is preferred and tried
4124 * to be setup first and until successfully established,
4125 * the link type will be hinted as eSCO.
4126 */
4127 conn = hci_conn_hash_lookup_ba(hdev, ESCO_LINK, &ev->bdaddr);
4128 if (!conn)
4129 goto unlock;
4130 }
4131
4132 switch (ev->status) {
4133 case 0x00:
4134 conn->handle = __le16_to_cpu(ev->handle);
4135 conn->state = BT_CONNECTED;
4136 conn->type = ev->link_type;
4137
4138 hci_debugfs_create_conn(conn);
4139 hci_conn_add_sysfs(conn);
4140 break;
4141
4142 case 0x10: /* Connection Accept Timeout */
4143 case 0x0d: /* Connection Rejected due to Limited Resources */
4144 case 0x11: /* Unsupported Feature or Parameter Value */
4145 case 0x1c: /* SCO interval rejected */
4146 case 0x1a: /* Unsupported Remote Feature */
4147 case 0x1f: /* Unspecified error */
4148 case 0x20: /* Unsupported LMP Parameter value */
4149 if (conn->out) {
4150 conn->pkt_type = (hdev->esco_type & SCO_ESCO_MASK) |
4151 (hdev->esco_type & EDR_ESCO_MASK);
4152 if (hci_setup_sync(conn, conn->link->handle))
4153 goto unlock;
4154 }
4155 /* fall through */
4156
4157 default:
4158 conn->state = BT_CLOSED;
4159 break;
4160 }
4161
4162 hci_connect_cfm(conn, ev->status);
4163 if (ev->status)
4164 hci_conn_del(conn);
4165
4166unlock:
4167 hci_dev_unlock(hdev);
4168}
4169
4170static inline size_t eir_get_length(u8 *eir, size_t eir_len)
4171{
4172 size_t parsed = 0;
4173
4174 while (parsed < eir_len) {
4175 u8 field_len = eir[0];
4176
4177 if (field_len == 0)
4178 return parsed;
4179
4180 parsed += field_len + 1;
4181 eir += field_len + 1;
4182 }
4183
4184 return eir_len;
4185}
4186
4187static void hci_extended_inquiry_result_evt(struct hci_dev *hdev,
4188 struct sk_buff *skb)
4189{
4190 struct inquiry_data data;
4191 struct extended_inquiry_info *info = (void *) (skb->data + 1);
4192 int num_rsp = *((__u8 *) skb->data);
4193 size_t eir_len;
4194
4195 BT_DBG("%s num_rsp %d", hdev->name, num_rsp);
4196
4197 if (!num_rsp)
4198 return;
4199
4200 if (hci_dev_test_flag(hdev, HCI_PERIODIC_INQ))
4201 return;
4202
4203 hci_dev_lock(hdev);
4204
4205 for (; num_rsp; num_rsp--, info++) {
4206 u32 flags;
4207 bool name_known;
4208
4209 bacpy(&data.bdaddr, &info->bdaddr);
4210 data.pscan_rep_mode = info->pscan_rep_mode;
4211 data.pscan_period_mode = info->pscan_period_mode;
4212 data.pscan_mode = 0x00;
4213 memcpy(data.dev_class, info->dev_class, 3);
4214 data.clock_offset = info->clock_offset;
4215 data.rssi = info->rssi;
4216 data.ssp_mode = 0x01;
4217
4218 if (hci_dev_test_flag(hdev, HCI_MGMT))
4219 name_known = eir_get_data(info->data,
4220 sizeof(info->data),
4221 EIR_NAME_COMPLETE, NULL);
4222 else
4223 name_known = true;
4224
4225 flags = hci_inquiry_cache_update(hdev, &data, name_known);
4226
4227 eir_len = eir_get_length(info->data, sizeof(info->data));
4228
4229 mgmt_device_found(hdev, &info->bdaddr, ACL_LINK, 0x00,
4230 info->dev_class, info->rssi,
4231 flags, info->data, eir_len, NULL, 0);
4232 }
4233
4234 hci_dev_unlock(hdev);
4235}
4236
4237static void hci_key_refresh_complete_evt(struct hci_dev *hdev,
4238 struct sk_buff *skb)
4239{
4240 struct hci_ev_key_refresh_complete *ev = (void *) skb->data;
4241 struct hci_conn *conn;
4242
4243 BT_DBG("%s status 0x%2.2x handle 0x%4.4x", hdev->name, ev->status,
4244 __le16_to_cpu(ev->handle));
4245
4246 hci_dev_lock(hdev);
4247
4248 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
4249 if (!conn)
4250 goto unlock;
4251
4252 /* For BR/EDR the necessary steps are taken through the
4253 * auth_complete event.
4254 */
4255 if (conn->type != LE_LINK)
4256 goto unlock;
4257
4258 if (!ev->status)
4259 conn->sec_level = conn->pending_sec_level;
4260
4261 clear_bit(HCI_CONN_ENCRYPT_PEND, &conn->flags);
4262
4263 if (ev->status && conn->state == BT_CONNECTED) {
4264 hci_disconnect(conn, HCI_ERROR_AUTH_FAILURE);
4265 hci_conn_drop(conn);
4266 goto unlock;
4267 }
4268
4269 if (conn->state == BT_CONFIG) {
4270 if (!ev->status)
4271 conn->state = BT_CONNECTED;
4272
4273 hci_connect_cfm(conn, ev->status);
4274 hci_conn_drop(conn);
4275 } else {
4276 hci_auth_cfm(conn, ev->status);
4277
4278 hci_conn_hold(conn);
4279 conn->disc_timeout = HCI_DISCONN_TIMEOUT;
4280 hci_conn_drop(conn);
4281 }
4282
4283unlock:
4284 hci_dev_unlock(hdev);
4285}
4286
4287static u8 hci_get_auth_req(struct hci_conn *conn)
4288{
4289 /* If remote requests no-bonding follow that lead */
4290 if (conn->remote_auth == HCI_AT_NO_BONDING ||
4291 conn->remote_auth == HCI_AT_NO_BONDING_MITM)
4292 return conn->remote_auth | (conn->auth_type & 0x01);
4293
4294 /* If both remote and local have enough IO capabilities, require
4295 * MITM protection
4296 */
4297 if (conn->remote_cap != HCI_IO_NO_INPUT_OUTPUT &&
4298 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT)
4299 return conn->remote_auth | 0x01;
4300
4301 /* No MITM protection possible so ignore remote requirement */
4302 return (conn->remote_auth & ~0x01) | (conn->auth_type & 0x01);
4303}
4304
4305static u8 bredr_oob_data_present(struct hci_conn *conn)
4306{
4307 struct hci_dev *hdev = conn->hdev;
4308 struct oob_data *data;
4309
4310 data = hci_find_remote_oob_data(hdev, &conn->dst, BDADDR_BREDR);
4311 if (!data)
4312 return 0x00;
4313
4314 if (bredr_sc_enabled(hdev)) {
4315 /* When Secure Connections is enabled, then just
4316 * return the present value stored with the OOB
4317 * data. The stored value contains the right present
4318 * information. However it can only be trusted when
4319 * not in Secure Connection Only mode.
4320 */
4321 if (!hci_dev_test_flag(hdev, HCI_SC_ONLY))
4322 return data->present;
4323
4324 /* When Secure Connections Only mode is enabled, then
4325 * the P-256 values are required. If they are not
4326 * available, then do not declare that OOB data is
4327 * present.
4328 */
4329 if (!memcmp(data->rand256, ZERO_KEY, 16) ||
4330 !memcmp(data->hash256, ZERO_KEY, 16))
4331 return 0x00;
4332
4333 return 0x02;
4334 }
4335
4336 /* When Secure Connections is not enabled or actually
4337 * not supported by the hardware, then check that if
4338 * P-192 data values are present.
4339 */
4340 if (!memcmp(data->rand192, ZERO_KEY, 16) ||
4341 !memcmp(data->hash192, ZERO_KEY, 16))
4342 return 0x00;
4343
4344 return 0x01;
4345}
4346
4347static void hci_io_capa_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
4348{
4349 struct hci_ev_io_capa_request *ev = (void *) skb->data;
4350 struct hci_conn *conn;
4351
4352 BT_DBG("%s", hdev->name);
4353
4354 hci_dev_lock(hdev);
4355
4356 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4357 if (!conn)
4358 goto unlock;
4359
4360 hci_conn_hold(conn);
4361
4362 if (!hci_dev_test_flag(hdev, HCI_MGMT))
4363 goto unlock;
4364
4365 /* Allow pairing if we're pairable, the initiators of the
4366 * pairing or if the remote is not requesting bonding.
4367 */
4368 if (hci_dev_test_flag(hdev, HCI_BONDABLE) ||
4369 test_bit(HCI_CONN_AUTH_INITIATOR, &conn->flags) ||
4370 (conn->remote_auth & ~0x01) == HCI_AT_NO_BONDING) {
4371 struct hci_cp_io_capability_reply cp;
4372
4373 bacpy(&cp.bdaddr, &ev->bdaddr);
4374 /* Change the IO capability from KeyboardDisplay
4375 * to DisplayYesNo as it is not supported by BT spec. */
4376 cp.capability = (conn->io_capability == 0x04) ?
4377 HCI_IO_DISPLAY_YESNO : conn->io_capability;
4378
4379 /* If we are initiators, there is no remote information yet */
4380 if (conn->remote_auth == 0xff) {
4381 /* Request MITM protection if our IO caps allow it
4382 * except for the no-bonding case.
4383 */
4384 if (conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
4385 conn->auth_type != HCI_AT_NO_BONDING)
4386 conn->auth_type |= 0x01;
4387 } else {
4388 conn->auth_type = hci_get_auth_req(conn);
4389 }
4390
4391 /* If we're not bondable, force one of the non-bondable
4392 * authentication requirement values.
4393 */
4394 if (!hci_dev_test_flag(hdev, HCI_BONDABLE))
4395 conn->auth_type &= HCI_AT_NO_BONDING_MITM;
4396
4397 cp.authentication = conn->auth_type;
4398 cp.oob_data = bredr_oob_data_present(conn);
4399
4400 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_REPLY,
4401 sizeof(cp), &cp);
4402 } else {
4403 struct hci_cp_io_capability_neg_reply cp;
4404
4405 bacpy(&cp.bdaddr, &ev->bdaddr);
4406 cp.reason = HCI_ERROR_PAIRING_NOT_ALLOWED;
4407
4408 hci_send_cmd(hdev, HCI_OP_IO_CAPABILITY_NEG_REPLY,
4409 sizeof(cp), &cp);
4410 }
4411
4412unlock:
4413 hci_dev_unlock(hdev);
4414}
4415
4416static void hci_io_capa_reply_evt(struct hci_dev *hdev, struct sk_buff *skb)
4417{
4418 struct hci_ev_io_capa_reply *ev = (void *) skb->data;
4419 struct hci_conn *conn;
4420
4421 BT_DBG("%s", hdev->name);
4422
4423 hci_dev_lock(hdev);
4424
4425 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4426 if (!conn)
4427 goto unlock;
4428
4429 conn->remote_cap = ev->capability;
4430 conn->remote_auth = ev->authentication;
4431
4432unlock:
4433 hci_dev_unlock(hdev);
4434}
4435
4436static void hci_user_confirm_request_evt(struct hci_dev *hdev,
4437 struct sk_buff *skb)
4438{
4439 struct hci_ev_user_confirm_req *ev = (void *) skb->data;
4440 int loc_mitm, rem_mitm, confirm_hint = 0;
4441 struct hci_conn *conn;
4442
4443 BT_DBG("%s", hdev->name);
4444
4445 hci_dev_lock(hdev);
4446
4447 if (!hci_dev_test_flag(hdev, HCI_MGMT))
4448 goto unlock;
4449
4450 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4451 if (!conn)
4452 goto unlock;
4453
4454 loc_mitm = (conn->auth_type & 0x01);
4455 rem_mitm = (conn->remote_auth & 0x01);
4456
4457 /* If we require MITM but the remote device can't provide that
4458 * (it has NoInputNoOutput) then reject the confirmation
4459 * request. We check the security level here since it doesn't
4460 * necessarily match conn->auth_type.
4461 */
4462 if (conn->pending_sec_level > BT_SECURITY_MEDIUM &&
4463 conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) {
4464 BT_DBG("Rejecting request: remote device can't provide MITM");
4465 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_NEG_REPLY,
4466 sizeof(ev->bdaddr), &ev->bdaddr);
4467 goto unlock;
4468 }
4469
4470 /* If no side requires MITM protection; auto-accept */
4471 if ((!loc_mitm || conn->remote_cap == HCI_IO_NO_INPUT_OUTPUT) &&
4472 (!rem_mitm || conn->io_capability == HCI_IO_NO_INPUT_OUTPUT)) {
4473
4474 /* If we're not the initiators request authorization to
4475 * proceed from user space (mgmt_user_confirm with
4476 * confirm_hint set to 1). The exception is if neither
4477 * side had MITM or if the local IO capability is
4478 * NoInputNoOutput, in which case we do auto-accept
4479 */
4480 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) &&
4481 conn->io_capability != HCI_IO_NO_INPUT_OUTPUT &&
4482 (loc_mitm || rem_mitm)) {
4483 BT_DBG("Confirming auto-accept as acceptor");
4484 confirm_hint = 1;
4485 goto confirm;
4486 }
4487
4488 BT_DBG("Auto-accept of user confirmation with %ums delay",
4489 hdev->auto_accept_delay);
4490
4491 if (hdev->auto_accept_delay > 0) {
4492 int delay = msecs_to_jiffies(hdev->auto_accept_delay);
4493 queue_delayed_work(conn->hdev->workqueue,
4494 &conn->auto_accept_work, delay);
4495 goto unlock;
4496 }
4497
4498 hci_send_cmd(hdev, HCI_OP_USER_CONFIRM_REPLY,
4499 sizeof(ev->bdaddr), &ev->bdaddr);
4500 goto unlock;
4501 }
4502
4503confirm:
4504 mgmt_user_confirm_request(hdev, &ev->bdaddr, ACL_LINK, 0,
4505 le32_to_cpu(ev->passkey), confirm_hint);
4506
4507unlock:
4508 hci_dev_unlock(hdev);
4509}
4510
4511static void hci_user_passkey_request_evt(struct hci_dev *hdev,
4512 struct sk_buff *skb)
4513{
4514 struct hci_ev_user_passkey_req *ev = (void *) skb->data;
4515
4516 BT_DBG("%s", hdev->name);
4517
4518 if (hci_dev_test_flag(hdev, HCI_MGMT))
4519 mgmt_user_passkey_request(hdev, &ev->bdaddr, ACL_LINK, 0);
4520}
4521
4522static void hci_user_passkey_notify_evt(struct hci_dev *hdev,
4523 struct sk_buff *skb)
4524{
4525 struct hci_ev_user_passkey_notify *ev = (void *) skb->data;
4526 struct hci_conn *conn;
4527
4528 BT_DBG("%s", hdev->name);
4529
4530 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4531 if (!conn)
4532 return;
4533
4534 conn->passkey_notify = __le32_to_cpu(ev->passkey);
4535 conn->passkey_entered = 0;
4536
4537 if (hci_dev_test_flag(hdev, HCI_MGMT))
4538 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
4539 conn->dst_type, conn->passkey_notify,
4540 conn->passkey_entered);
4541}
4542
4543static void hci_keypress_notify_evt(struct hci_dev *hdev, struct sk_buff *skb)
4544{
4545 struct hci_ev_keypress_notify *ev = (void *) skb->data;
4546 struct hci_conn *conn;
4547
4548 BT_DBG("%s", hdev->name);
4549
4550 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4551 if (!conn)
4552 return;
4553
4554 switch (ev->type) {
4555 case HCI_KEYPRESS_STARTED:
4556 conn->passkey_entered = 0;
4557 return;
4558
4559 case HCI_KEYPRESS_ENTERED:
4560 conn->passkey_entered++;
4561 break;
4562
4563 case HCI_KEYPRESS_ERASED:
4564 conn->passkey_entered--;
4565 break;
4566
4567 case HCI_KEYPRESS_CLEARED:
4568 conn->passkey_entered = 0;
4569 break;
4570
4571 case HCI_KEYPRESS_COMPLETED:
4572 return;
4573 }
4574
4575 if (hci_dev_test_flag(hdev, HCI_MGMT))
4576 mgmt_user_passkey_notify(hdev, &conn->dst, conn->type,
4577 conn->dst_type, conn->passkey_notify,
4578 conn->passkey_entered);
4579}
4580
4581static void hci_simple_pair_complete_evt(struct hci_dev *hdev,
4582 struct sk_buff *skb)
4583{
4584 struct hci_ev_simple_pair_complete *ev = (void *) skb->data;
4585 struct hci_conn *conn;
4586
4587 BT_DBG("%s", hdev->name);
4588
4589 hci_dev_lock(hdev);
4590
4591 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4592 if (!conn)
4593 goto unlock;
4594
4595 /* Reset the authentication requirement to unknown */
4596 conn->remote_auth = 0xff;
4597
4598 /* To avoid duplicate auth_failed events to user space we check
4599 * the HCI_CONN_AUTH_PEND flag which will be set if we
4600 * initiated the authentication. A traditional auth_complete
4601 * event gets always produced as initiator and is also mapped to
4602 * the mgmt_auth_failed event */
4603 if (!test_bit(HCI_CONN_AUTH_PEND, &conn->flags) && ev->status)
4604 mgmt_auth_failed(conn, ev->status);
4605
4606 hci_conn_drop(conn);
4607
4608unlock:
4609 hci_dev_unlock(hdev);
4610}
4611
4612static void hci_remote_host_features_evt(struct hci_dev *hdev,
4613 struct sk_buff *skb)
4614{
4615 struct hci_ev_remote_host_features *ev = (void *) skb->data;
4616 struct inquiry_entry *ie;
4617 struct hci_conn *conn;
4618
4619 BT_DBG("%s", hdev->name);
4620
4621 hci_dev_lock(hdev);
4622
4623 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &ev->bdaddr);
4624 if (conn)
4625 memcpy(conn->features[1], ev->features, 8);
4626
4627 ie = hci_inquiry_cache_lookup(hdev, &ev->bdaddr);
4628 if (ie)
4629 ie->data.ssp_mode = (ev->features[0] & LMP_HOST_SSP);
4630
4631 hci_dev_unlock(hdev);
4632}
4633
4634static void hci_remote_oob_data_request_evt(struct hci_dev *hdev,
4635 struct sk_buff *skb)
4636{
4637 struct hci_ev_remote_oob_data_request *ev = (void *) skb->data;
4638 struct oob_data *data;
4639
4640 BT_DBG("%s", hdev->name);
4641
4642 hci_dev_lock(hdev);
4643
4644 if (!hci_dev_test_flag(hdev, HCI_MGMT))
4645 goto unlock;
4646
4647 data = hci_find_remote_oob_data(hdev, &ev->bdaddr, BDADDR_BREDR);
4648 if (!data) {
4649 struct hci_cp_remote_oob_data_neg_reply cp;
4650
4651 bacpy(&cp.bdaddr, &ev->bdaddr);
4652 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_NEG_REPLY,
4653 sizeof(cp), &cp);
4654 goto unlock;
4655 }
4656
4657 if (bredr_sc_enabled(hdev)) {
4658 struct hci_cp_remote_oob_ext_data_reply cp;
4659
4660 bacpy(&cp.bdaddr, &ev->bdaddr);
4661 if (hci_dev_test_flag(hdev, HCI_SC_ONLY)) {
4662 memset(cp.hash192, 0, sizeof(cp.hash192));
4663 memset(cp.rand192, 0, sizeof(cp.rand192));
4664 } else {
4665 memcpy(cp.hash192, data->hash192, sizeof(cp.hash192));
4666 memcpy(cp.rand192, data->rand192, sizeof(cp.rand192));
4667 }
4668 memcpy(cp.hash256, data->hash256, sizeof(cp.hash256));
4669 memcpy(cp.rand256, data->rand256, sizeof(cp.rand256));
4670
4671 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_EXT_DATA_REPLY,
4672 sizeof(cp), &cp);
4673 } else {
4674 struct hci_cp_remote_oob_data_reply cp;
4675
4676 bacpy(&cp.bdaddr, &ev->bdaddr);
4677 memcpy(cp.hash, data->hash192, sizeof(cp.hash));
4678 memcpy(cp.rand, data->rand192, sizeof(cp.rand));
4679
4680 hci_send_cmd(hdev, HCI_OP_REMOTE_OOB_DATA_REPLY,
4681 sizeof(cp), &cp);
4682 }
4683
4684unlock:
4685 hci_dev_unlock(hdev);
4686}
4687
4688#if IS_ENABLED(CONFIG_BT_HS)
4689static void hci_chan_selected_evt(struct hci_dev *hdev, struct sk_buff *skb)
4690{
4691 struct hci_ev_channel_selected *ev = (void *)skb->data;
4692 struct hci_conn *hcon;
4693
4694 BT_DBG("%s handle 0x%2.2x", hdev->name, ev->phy_handle);
4695
4696 skb_pull(skb, sizeof(*ev));
4697
4698 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4699 if (!hcon)
4700 return;
4701
4702 amp_read_loc_assoc_final_data(hdev, hcon);
4703}
4704
4705static void hci_phy_link_complete_evt(struct hci_dev *hdev,
4706 struct sk_buff *skb)
4707{
4708 struct hci_ev_phy_link_complete *ev = (void *) skb->data;
4709 struct hci_conn *hcon, *bredr_hcon;
4710
4711 BT_DBG("%s handle 0x%2.2x status 0x%2.2x", hdev->name, ev->phy_handle,
4712 ev->status);
4713
4714 hci_dev_lock(hdev);
4715
4716 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4717 if (!hcon) {
4718 hci_dev_unlock(hdev);
4719 return;
4720 }
4721
4722 if (ev->status) {
4723 hci_conn_del(hcon);
4724 hci_dev_unlock(hdev);
4725 return;
4726 }
4727
4728 bredr_hcon = hcon->amp_mgr->l2cap_conn->hcon;
4729
4730 hcon->state = BT_CONNECTED;
4731 bacpy(&hcon->dst, &bredr_hcon->dst);
4732
4733 hci_conn_hold(hcon);
4734 hcon->disc_timeout = HCI_DISCONN_TIMEOUT;
4735 hci_conn_drop(hcon);
4736
4737 hci_debugfs_create_conn(hcon);
4738 hci_conn_add_sysfs(hcon);
4739
4740 amp_physical_cfm(bredr_hcon, hcon);
4741
4742 hci_dev_unlock(hdev);
4743}
4744
4745static void hci_loglink_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4746{
4747 struct hci_ev_logical_link_complete *ev = (void *) skb->data;
4748 struct hci_conn *hcon;
4749 struct hci_chan *hchan;
4750 struct amp_mgr *mgr;
4751
4752 BT_DBG("%s log_handle 0x%4.4x phy_handle 0x%2.2x status 0x%2.2x",
4753 hdev->name, le16_to_cpu(ev->handle), ev->phy_handle,
4754 ev->status);
4755
4756 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4757 if (!hcon)
4758 return;
4759
4760 /* Create AMP hchan */
4761 hchan = hci_chan_create(hcon);
4762 if (!hchan)
4763 return;
4764
4765 hchan->handle = le16_to_cpu(ev->handle);
4766
4767 BT_DBG("hcon %p mgr %p hchan %p", hcon, hcon->amp_mgr, hchan);
4768
4769 mgr = hcon->amp_mgr;
4770 if (mgr && mgr->bredr_chan) {
4771 struct l2cap_chan *bredr_chan = mgr->bredr_chan;
4772
4773 l2cap_chan_lock(bredr_chan);
4774
4775 bredr_chan->conn->mtu = hdev->block_mtu;
4776 l2cap_logical_cfm(bredr_chan, hchan, 0);
4777 hci_conn_hold(hcon);
4778
4779 l2cap_chan_unlock(bredr_chan);
4780 }
4781}
4782
4783static void hci_disconn_loglink_complete_evt(struct hci_dev *hdev,
4784 struct sk_buff *skb)
4785{
4786 struct hci_ev_disconn_logical_link_complete *ev = (void *) skb->data;
4787 struct hci_chan *hchan;
4788
4789 BT_DBG("%s log handle 0x%4.4x status 0x%2.2x", hdev->name,
4790 le16_to_cpu(ev->handle), ev->status);
4791
4792 if (ev->status)
4793 return;
4794
4795 hci_dev_lock(hdev);
4796
4797 hchan = hci_chan_lookup_handle(hdev, le16_to_cpu(ev->handle));
4798 if (!hchan)
4799 goto unlock;
4800
4801 amp_destroy_logical_link(hchan, ev->reason);
4802
4803unlock:
4804 hci_dev_unlock(hdev);
4805}
4806
4807static void hci_disconn_phylink_complete_evt(struct hci_dev *hdev,
4808 struct sk_buff *skb)
4809{
4810 struct hci_ev_disconn_phy_link_complete *ev = (void *) skb->data;
4811 struct hci_conn *hcon;
4812
4813 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4814
4815 if (ev->status)
4816 return;
4817
4818 hci_dev_lock(hdev);
4819
4820 hcon = hci_conn_hash_lookup_handle(hdev, ev->phy_handle);
4821 if (hcon) {
4822 hcon->state = BT_CLOSED;
4823 hci_conn_del(hcon);
4824 }
4825
4826 hci_dev_unlock(hdev);
4827}
4828#endif
4829
4830static void le_conn_complete_evt(struct hci_dev *hdev, u8 status,
4831 bdaddr_t *bdaddr, u8 bdaddr_type, u8 role, u16 handle,
4832 u16 interval, u16 latency, u16 supervision_timeout)
4833{
4834 struct hci_conn_params *params;
4835 struct hci_conn *conn;
4836 struct smp_irk *irk;
4837 u8 addr_type;
4838
4839 hci_dev_lock(hdev);
4840
4841 /* All controllers implicitly stop advertising in the event of a
4842 * connection, so ensure that the state bit is cleared.
4843 */
4844 hci_dev_clear_flag(hdev, HCI_LE_ADV);
4845
4846 conn = hci_lookup_le_connect(hdev);
4847 if (!conn) {
4848 conn = hci_conn_add(hdev, LE_LINK, bdaddr, role);
4849 if (!conn) {
4850 bt_dev_err(hdev, "no memory for new connection");
4851 goto unlock;
4852 }
4853
4854 conn->dst_type = bdaddr_type;
4855
4856 /* If we didn't have a hci_conn object previously
4857 * but we're in master role this must be something
4858 * initiated using a white list. Since white list based
4859 * connections are not "first class citizens" we don't
4860 * have full tracking of them. Therefore, we go ahead
4861 * with a "best effort" approach of determining the
4862 * initiator address based on the HCI_PRIVACY flag.
4863 */
4864 if (conn->out) {
4865 conn->resp_addr_type = bdaddr_type;
4866 bacpy(&conn->resp_addr, bdaddr);
4867 if (hci_dev_test_flag(hdev, HCI_PRIVACY)) {
4868 conn->init_addr_type = ADDR_LE_DEV_RANDOM;
4869 bacpy(&conn->init_addr, &hdev->rpa);
4870 } else {
4871 hci_copy_identity_address(hdev,
4872 &conn->init_addr,
4873 &conn->init_addr_type);
4874 }
4875 }
4876 } else {
4877 cancel_delayed_work(&conn->le_conn_timeout);
4878 }
4879
4880 if (!conn->out) {
4881 /* Set the responder (our side) address type based on
4882 * the advertising address type.
4883 */
4884 conn->resp_addr_type = hdev->adv_addr_type;
4885 if (hdev->adv_addr_type == ADDR_LE_DEV_RANDOM) {
4886 /* In case of ext adv, resp_addr will be updated in
4887 * Adv Terminated event.
4888 */
4889 if (!ext_adv_capable(hdev))
4890 bacpy(&conn->resp_addr, &hdev->random_addr);
4891 } else {
4892 bacpy(&conn->resp_addr, &hdev->bdaddr);
4893 }
4894
4895 conn->init_addr_type = bdaddr_type;
4896 bacpy(&conn->init_addr, bdaddr);
4897
4898 /* For incoming connections, set the default minimum
4899 * and maximum connection interval. They will be used
4900 * to check if the parameters are in range and if not
4901 * trigger the connection update procedure.
4902 */
4903 conn->le_conn_min_interval = hdev->le_conn_min_interval;
4904 conn->le_conn_max_interval = hdev->le_conn_max_interval;
4905 }
4906
4907 /* Lookup the identity address from the stored connection
4908 * address and address type.
4909 *
4910 * When establishing connections to an identity address, the
4911 * connection procedure will store the resolvable random
4912 * address first. Now if it can be converted back into the
4913 * identity address, start using the identity address from
4914 * now on.
4915 */
4916 irk = hci_get_irk(hdev, &conn->dst, conn->dst_type);
4917 if (irk) {
4918 bacpy(&conn->dst, &irk->bdaddr);
4919 conn->dst_type = irk->addr_type;
4920 }
4921
4922 if (status) {
4923 hci_le_conn_failed(conn, status);
4924 goto unlock;
4925 }
4926
4927 if (conn->dst_type == ADDR_LE_DEV_PUBLIC)
4928 addr_type = BDADDR_LE_PUBLIC;
4929 else
4930 addr_type = BDADDR_LE_RANDOM;
4931
4932 /* Drop the connection if the device is blocked */
4933 if (hci_bdaddr_list_lookup(&hdev->blacklist, &conn->dst, addr_type)) {
4934 hci_conn_drop(conn);
4935 goto unlock;
4936 }
4937
4938 if (!test_and_set_bit(HCI_CONN_MGMT_CONNECTED, &conn->flags))
4939 mgmt_device_connected(hdev, conn, 0, NULL, 0);
4940
4941 conn->sec_level = BT_SECURITY_LOW;
4942 conn->handle = handle;
4943 conn->state = BT_CONFIG;
4944
4945 conn->le_conn_interval = interval;
4946 conn->le_conn_latency = latency;
4947 conn->le_supv_timeout = supervision_timeout;
4948
4949 hci_debugfs_create_conn(conn);
4950 hci_conn_add_sysfs(conn);
4951
4952 /* The remote features procedure is defined for master
4953 * role only. So only in case of an initiated connection
4954 * request the remote features.
4955 *
4956 * If the local controller supports slave-initiated features
4957 * exchange, then requesting the remote features in slave
4958 * role is possible. Otherwise just transition into the
4959 * connected state without requesting the remote features.
4960 */
4961 if (conn->out ||
4962 (hdev->le_features[0] & HCI_LE_SLAVE_FEATURES)) {
4963 struct hci_cp_le_read_remote_features cp;
4964
4965 cp.handle = __cpu_to_le16(conn->handle);
4966
4967 hci_send_cmd(hdev, HCI_OP_LE_READ_REMOTE_FEATURES,
4968 sizeof(cp), &cp);
4969
4970 hci_conn_hold(conn);
4971 } else {
4972 conn->state = BT_CONNECTED;
4973 hci_connect_cfm(conn, status);
4974 }
4975
4976 params = hci_pend_le_action_lookup(&hdev->pend_le_conns, &conn->dst,
4977 conn->dst_type);
4978 if (params) {
4979 list_del_init(¶ms->action);
4980 if (params->conn) {
4981 hci_conn_drop(params->conn);
4982 hci_conn_put(params->conn);
4983 params->conn = NULL;
4984 }
4985 }
4986
4987unlock:
4988 hci_update_background_scan(hdev);
4989 hci_dev_unlock(hdev);
4990}
4991
4992static void hci_le_conn_complete_evt(struct hci_dev *hdev, struct sk_buff *skb)
4993{
4994 struct hci_ev_le_conn_complete *ev = (void *) skb->data;
4995
4996 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
4997
4998 le_conn_complete_evt(hdev, ev->status, &ev->bdaddr, ev->bdaddr_type,
4999 ev->role, le16_to_cpu(ev->handle),
5000 le16_to_cpu(ev->interval),
5001 le16_to_cpu(ev->latency),
5002 le16_to_cpu(ev->supervision_timeout));
5003}
5004
5005static void hci_le_enh_conn_complete_evt(struct hci_dev *hdev,
5006 struct sk_buff *skb)
5007{
5008 struct hci_ev_le_enh_conn_complete *ev = (void *) skb->data;
5009
5010 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
5011
5012 le_conn_complete_evt(hdev, ev->status, &ev->bdaddr, ev->bdaddr_type,
5013 ev->role, le16_to_cpu(ev->handle),
5014 le16_to_cpu(ev->interval),
5015 le16_to_cpu(ev->latency),
5016 le16_to_cpu(ev->supervision_timeout));
5017}
5018
5019static void hci_le_ext_adv_term_evt(struct hci_dev *hdev, struct sk_buff *skb)
5020{
5021 struct hci_evt_le_ext_adv_set_term *ev = (void *) skb->data;
5022 struct hci_conn *conn;
5023
5024 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
5025
5026 if (ev->status)
5027 return;
5028
5029 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->conn_handle));
5030 if (conn) {
5031 struct adv_info *adv_instance;
5032
5033 if (hdev->adv_addr_type != ADDR_LE_DEV_RANDOM)
5034 return;
5035
5036 if (!hdev->cur_adv_instance) {
5037 bacpy(&conn->resp_addr, &hdev->random_addr);
5038 return;
5039 }
5040
5041 adv_instance = hci_find_adv_instance(hdev, hdev->cur_adv_instance);
5042 if (adv_instance)
5043 bacpy(&conn->resp_addr, &adv_instance->random_addr);
5044 }
5045}
5046
5047static void hci_le_conn_update_complete_evt(struct hci_dev *hdev,
5048 struct sk_buff *skb)
5049{
5050 struct hci_ev_le_conn_update_complete *ev = (void *) skb->data;
5051 struct hci_conn *conn;
5052
5053 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
5054
5055 if (ev->status)
5056 return;
5057
5058 hci_dev_lock(hdev);
5059
5060 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
5061 if (conn) {
5062 conn->le_conn_interval = le16_to_cpu(ev->interval);
5063 conn->le_conn_latency = le16_to_cpu(ev->latency);
5064 conn->le_supv_timeout = le16_to_cpu(ev->supervision_timeout);
5065 }
5066
5067 hci_dev_unlock(hdev);
5068}
5069
5070/* This function requires the caller holds hdev->lock */
5071static struct hci_conn *check_pending_le_conn(struct hci_dev *hdev,
5072 bdaddr_t *addr,
5073 u8 addr_type, u8 adv_type,
5074 bdaddr_t *direct_rpa)
5075{
5076 struct hci_conn *conn;
5077 struct hci_conn_params *params;
5078
5079 /* If the event is not connectable don't proceed further */
5080 if (adv_type != LE_ADV_IND && adv_type != LE_ADV_DIRECT_IND)
5081 return NULL;
5082
5083 /* Ignore if the device is blocked */
5084 if (hci_bdaddr_list_lookup(&hdev->blacklist, addr, addr_type))
5085 return NULL;
5086
5087 /* Most controller will fail if we try to create new connections
5088 * while we have an existing one in slave role.
5089 */
5090 if (hdev->conn_hash.le_num_slave > 0)
5091 return NULL;
5092
5093 /* If we're not connectable only connect devices that we have in
5094 * our pend_le_conns list.
5095 */
5096 params = hci_pend_le_action_lookup(&hdev->pend_le_conns, addr,
5097 addr_type);
5098 if (!params)
5099 return NULL;
5100
5101 if (!params->explicit_connect) {
5102 switch (params->auto_connect) {
5103 case HCI_AUTO_CONN_DIRECT:
5104 /* Only devices advertising with ADV_DIRECT_IND are
5105 * triggering a connection attempt. This is allowing
5106 * incoming connections from slave devices.
5107 */
5108 if (adv_type != LE_ADV_DIRECT_IND)
5109 return NULL;
5110 break;
5111 case HCI_AUTO_CONN_ALWAYS:
5112 /* Devices advertising with ADV_IND or ADV_DIRECT_IND
5113 * are triggering a connection attempt. This means
5114 * that incoming connectioms from slave device are
5115 * accepted and also outgoing connections to slave
5116 * devices are established when found.
5117 */
5118 break;
5119 default:
5120 return NULL;
5121 }
5122 }
5123
5124 conn = hci_connect_le(hdev, addr, addr_type, BT_SECURITY_LOW,
5125 HCI_LE_AUTOCONN_TIMEOUT, HCI_ROLE_MASTER,
5126 direct_rpa);
5127 if (!IS_ERR(conn)) {
5128 /* If HCI_AUTO_CONN_EXPLICIT is set, conn is already owned
5129 * by higher layer that tried to connect, if no then
5130 * store the pointer since we don't really have any
5131 * other owner of the object besides the params that
5132 * triggered it. This way we can abort the connection if
5133 * the parameters get removed and keep the reference
5134 * count consistent once the connection is established.
5135 */
5136
5137 if (!params->explicit_connect)
5138 params->conn = hci_conn_get(conn);
5139
5140 return conn;
5141 }
5142
5143 switch (PTR_ERR(conn)) {
5144 case -EBUSY:
5145 /* If hci_connect() returns -EBUSY it means there is already
5146 * an LE connection attempt going on. Since controllers don't
5147 * support more than one connection attempt at the time, we
5148 * don't consider this an error case.
5149 */
5150 break;
5151 default:
5152 BT_DBG("Failed to connect: err %ld", PTR_ERR(conn));
5153 return NULL;
5154 }
5155
5156 return NULL;
5157}
5158
5159static void process_adv_report(struct hci_dev *hdev, u8 type, bdaddr_t *bdaddr,
5160 u8 bdaddr_type, bdaddr_t *direct_addr,
5161 u8 direct_addr_type, s8 rssi, u8 *data, u8 len)
5162{
5163 struct discovery_state *d = &hdev->discovery;
5164 struct smp_irk *irk;
5165 struct hci_conn *conn;
5166 bool match;
5167 u32 flags;
5168 u8 *ptr, real_len;
5169
5170 switch (type) {
5171 case LE_ADV_IND:
5172 case LE_ADV_DIRECT_IND:
5173 case LE_ADV_SCAN_IND:
5174 case LE_ADV_NONCONN_IND:
5175 case LE_ADV_SCAN_RSP:
5176 break;
5177 default:
5178 bt_dev_err_ratelimited(hdev, "unknown advertising packet "
5179 "type: 0x%02x", type);
5180 return;
5181 }
5182
5183 /* Find the end of the data in case the report contains padded zero
5184 * bytes at the end causing an invalid length value.
5185 *
5186 * When data is NULL, len is 0 so there is no need for extra ptr
5187 * check as 'ptr < data + 0' is already false in such case.
5188 */
5189 for (ptr = data; ptr < data + len && *ptr; ptr += *ptr + 1) {
5190 if (ptr + 1 + *ptr > data + len)
5191 break;
5192 }
5193
5194 real_len = ptr - data;
5195
5196 /* Adjust for actual length */
5197 if (len != real_len) {
5198 bt_dev_err_ratelimited(hdev, "advertising data len corrected");
5199 len = real_len;
5200 }
5201
5202 /* If the direct address is present, then this report is from
5203 * a LE Direct Advertising Report event. In that case it is
5204 * important to see if the address is matching the local
5205 * controller address.
5206 */
5207 if (direct_addr) {
5208 /* Only resolvable random addresses are valid for these
5209 * kind of reports and others can be ignored.
5210 */
5211 if (!hci_bdaddr_is_rpa(direct_addr, direct_addr_type))
5212 return;
5213
5214 /* If the controller is not using resolvable random
5215 * addresses, then this report can be ignored.
5216 */
5217 if (!hci_dev_test_flag(hdev, HCI_PRIVACY))
5218 return;
5219
5220 /* If the local IRK of the controller does not match
5221 * with the resolvable random address provided, then
5222 * this report can be ignored.
5223 */
5224 if (!smp_irk_matches(hdev, hdev->irk, direct_addr))
5225 return;
5226 }
5227
5228 /* Check if we need to convert to identity address */
5229 irk = hci_get_irk(hdev, bdaddr, bdaddr_type);
5230 if (irk) {
5231 bdaddr = &irk->bdaddr;
5232 bdaddr_type = irk->addr_type;
5233 }
5234
5235 /* Check if we have been requested to connect to this device.
5236 *
5237 * direct_addr is set only for directed advertising reports (it is NULL
5238 * for advertising reports) and is already verified to be RPA above.
5239 */
5240 conn = check_pending_le_conn(hdev, bdaddr, bdaddr_type, type,
5241 direct_addr);
5242 if (conn && type == LE_ADV_IND) {
5243 /* Store report for later inclusion by
5244 * mgmt_device_connected
5245 */
5246 memcpy(conn->le_adv_data, data, len);
5247 conn->le_adv_data_len = len;
5248 }
5249
5250 /* Passive scanning shouldn't trigger any device found events,
5251 * except for devices marked as CONN_REPORT for which we do send
5252 * device found events.
5253 */
5254 if (hdev->le_scan_type == LE_SCAN_PASSIVE) {
5255 if (type == LE_ADV_DIRECT_IND)
5256 return;
5257
5258 if (!hci_pend_le_action_lookup(&hdev->pend_le_reports,
5259 bdaddr, bdaddr_type))
5260 return;
5261
5262 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND)
5263 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
5264 else
5265 flags = 0;
5266 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
5267 rssi, flags, data, len, NULL, 0);
5268 return;
5269 }
5270
5271 /* When receiving non-connectable or scannable undirected
5272 * advertising reports, this means that the remote device is
5273 * not connectable and then clearly indicate this in the
5274 * device found event.
5275 *
5276 * When receiving a scan response, then there is no way to
5277 * know if the remote device is connectable or not. However
5278 * since scan responses are merged with a previously seen
5279 * advertising report, the flags field from that report
5280 * will be used.
5281 *
5282 * In the really unlikely case that a controller get confused
5283 * and just sends a scan response event, then it is marked as
5284 * not connectable as well.
5285 */
5286 if (type == LE_ADV_NONCONN_IND || type == LE_ADV_SCAN_IND ||
5287 type == LE_ADV_SCAN_RSP)
5288 flags = MGMT_DEV_FOUND_NOT_CONNECTABLE;
5289 else
5290 flags = 0;
5291
5292 /* If there's nothing pending either store the data from this
5293 * event or send an immediate device found event if the data
5294 * should not be stored for later.
5295 */
5296 if (!has_pending_adv_report(hdev)) {
5297 /* If the report will trigger a SCAN_REQ store it for
5298 * later merging.
5299 */
5300 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
5301 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
5302 rssi, flags, data, len);
5303 return;
5304 }
5305
5306 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
5307 rssi, flags, data, len, NULL, 0);
5308 return;
5309 }
5310
5311 /* Check if the pending report is for the same device as the new one */
5312 match = (!bacmp(bdaddr, &d->last_adv_addr) &&
5313 bdaddr_type == d->last_adv_addr_type);
5314
5315 /* If the pending data doesn't match this report or this isn't a
5316 * scan response (e.g. we got a duplicate ADV_IND) then force
5317 * sending of the pending data.
5318 */
5319 if (type != LE_ADV_SCAN_RSP || !match) {
5320 /* Send out whatever is in the cache, but skip duplicates */
5321 if (!match)
5322 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
5323 d->last_adv_addr_type, NULL,
5324 d->last_adv_rssi, d->last_adv_flags,
5325 d->last_adv_data,
5326 d->last_adv_data_len, NULL, 0);
5327
5328 /* If the new report will trigger a SCAN_REQ store it for
5329 * later merging.
5330 */
5331 if (type == LE_ADV_IND || type == LE_ADV_SCAN_IND) {
5332 store_pending_adv_report(hdev, bdaddr, bdaddr_type,
5333 rssi, flags, data, len);
5334 return;
5335 }
5336
5337 /* The advertising reports cannot be merged, so clear
5338 * the pending report and send out a device found event.
5339 */
5340 clear_pending_adv_report(hdev);
5341 mgmt_device_found(hdev, bdaddr, LE_LINK, bdaddr_type, NULL,
5342 rssi, flags, data, len, NULL, 0);
5343 return;
5344 }
5345
5346 /* If we get here we've got a pending ADV_IND or ADV_SCAN_IND and
5347 * the new event is a SCAN_RSP. We can therefore proceed with
5348 * sending a merged device found event.
5349 */
5350 mgmt_device_found(hdev, &d->last_adv_addr, LE_LINK,
5351 d->last_adv_addr_type, NULL, rssi, d->last_adv_flags,
5352 d->last_adv_data, d->last_adv_data_len, data, len);
5353 clear_pending_adv_report(hdev);
5354}
5355
5356static void hci_le_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb)
5357{
5358 u8 num_reports = skb->data[0];
5359 void *ptr = &skb->data[1];
5360
5361 hci_dev_lock(hdev);
5362
5363 while (num_reports--) {
5364 struct hci_ev_le_advertising_info *ev = ptr;
5365 s8 rssi;
5366
5367 if (ev->length <= HCI_MAX_AD_LENGTH) {
5368 rssi = ev->data[ev->length];
5369 process_adv_report(hdev, ev->evt_type, &ev->bdaddr,
5370 ev->bdaddr_type, NULL, 0, rssi,
5371 ev->data, ev->length);
5372 } else {
5373 bt_dev_err(hdev, "Dropping invalid advertising data");
5374 }
5375
5376 ptr += sizeof(*ev) + ev->length + 1;
5377 }
5378
5379 hci_dev_unlock(hdev);
5380}
5381
5382static u8 ext_evt_type_to_legacy(u16 evt_type)
5383{
5384 if (evt_type & LE_EXT_ADV_LEGACY_PDU) {
5385 switch (evt_type) {
5386 case LE_LEGACY_ADV_IND:
5387 return LE_ADV_IND;
5388 case LE_LEGACY_ADV_DIRECT_IND:
5389 return LE_ADV_DIRECT_IND;
5390 case LE_LEGACY_ADV_SCAN_IND:
5391 return LE_ADV_SCAN_IND;
5392 case LE_LEGACY_NONCONN_IND:
5393 return LE_ADV_NONCONN_IND;
5394 case LE_LEGACY_SCAN_RSP_ADV:
5395 case LE_LEGACY_SCAN_RSP_ADV_SCAN:
5396 return LE_ADV_SCAN_RSP;
5397 }
5398
5399 BT_ERR_RATELIMITED("Unknown advertising packet type: 0x%02x",
5400 evt_type);
5401
5402 return LE_ADV_INVALID;
5403 }
5404
5405 if (evt_type & LE_EXT_ADV_CONN_IND) {
5406 if (evt_type & LE_EXT_ADV_DIRECT_IND)
5407 return LE_ADV_DIRECT_IND;
5408
5409 return LE_ADV_IND;
5410 }
5411
5412 if (evt_type & LE_EXT_ADV_SCAN_RSP)
5413 return LE_ADV_SCAN_RSP;
5414
5415 if (evt_type & LE_EXT_ADV_SCAN_IND)
5416 return LE_ADV_SCAN_IND;
5417
5418 if (evt_type == LE_EXT_ADV_NON_CONN_IND ||
5419 evt_type & LE_EXT_ADV_DIRECT_IND)
5420 return LE_ADV_NONCONN_IND;
5421
5422 BT_ERR_RATELIMITED("Unknown advertising packet type: 0x%02x",
5423 evt_type);
5424
5425 return LE_ADV_INVALID;
5426}
5427
5428static void hci_le_ext_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb)
5429{
5430 u8 num_reports = skb->data[0];
5431 void *ptr = &skb->data[1];
5432
5433 hci_dev_lock(hdev);
5434
5435 while (num_reports--) {
5436 struct hci_ev_le_ext_adv_report *ev = ptr;
5437 u8 legacy_evt_type;
5438 u16 evt_type;
5439
5440 evt_type = __le16_to_cpu(ev->evt_type);
5441 legacy_evt_type = ext_evt_type_to_legacy(evt_type);
5442 if (legacy_evt_type != LE_ADV_INVALID) {
5443 process_adv_report(hdev, legacy_evt_type, &ev->bdaddr,
5444 ev->bdaddr_type, NULL, 0, ev->rssi,
5445 ev->data, ev->length);
5446 }
5447
5448 ptr += sizeof(*ev) + ev->length;
5449 }
5450
5451 hci_dev_unlock(hdev);
5452}
5453
5454static void hci_le_remote_feat_complete_evt(struct hci_dev *hdev,
5455 struct sk_buff *skb)
5456{
5457 struct hci_ev_le_remote_feat_complete *ev = (void *)skb->data;
5458 struct hci_conn *conn;
5459
5460 BT_DBG("%s status 0x%2.2x", hdev->name, ev->status);
5461
5462 hci_dev_lock(hdev);
5463
5464 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
5465 if (conn) {
5466 if (!ev->status)
5467 memcpy(conn->features[0], ev->features, 8);
5468
5469 if (conn->state == BT_CONFIG) {
5470 __u8 status;
5471
5472 /* If the local controller supports slave-initiated
5473 * features exchange, but the remote controller does
5474 * not, then it is possible that the error code 0x1a
5475 * for unsupported remote feature gets returned.
5476 *
5477 * In this specific case, allow the connection to
5478 * transition into connected state and mark it as
5479 * successful.
5480 */
5481 if ((hdev->le_features[0] & HCI_LE_SLAVE_FEATURES) &&
5482 !conn->out && ev->status == 0x1a)
5483 status = 0x00;
5484 else
5485 status = ev->status;
5486
5487 conn->state = BT_CONNECTED;
5488 hci_connect_cfm(conn, status);
5489 hci_conn_drop(conn);
5490 }
5491 }
5492
5493 hci_dev_unlock(hdev);
5494}
5495
5496static void hci_le_ltk_request_evt(struct hci_dev *hdev, struct sk_buff *skb)
5497{
5498 struct hci_ev_le_ltk_req *ev = (void *) skb->data;
5499 struct hci_cp_le_ltk_reply cp;
5500 struct hci_cp_le_ltk_neg_reply neg;
5501 struct hci_conn *conn;
5502 struct smp_ltk *ltk;
5503
5504 BT_DBG("%s handle 0x%4.4x", hdev->name, __le16_to_cpu(ev->handle));
5505
5506 hci_dev_lock(hdev);
5507
5508 conn = hci_conn_hash_lookup_handle(hdev, __le16_to_cpu(ev->handle));
5509 if (conn == NULL)
5510 goto not_found;
5511
5512 ltk = hci_find_ltk(hdev, &conn->dst, conn->dst_type, conn->role);
5513 if (!ltk)
5514 goto not_found;
5515
5516 if (smp_ltk_is_sc(ltk)) {
5517 /* With SC both EDiv and Rand are set to zero */
5518 if (ev->ediv || ev->rand)
5519 goto not_found;
5520 } else {
5521 /* For non-SC keys check that EDiv and Rand match */
5522 if (ev->ediv != ltk->ediv || ev->rand != ltk->rand)
5523 goto not_found;
5524 }
5525
5526 memcpy(cp.ltk, ltk->val, ltk->enc_size);
5527 memset(cp.ltk + ltk->enc_size, 0, sizeof(cp.ltk) - ltk->enc_size);
5528 cp.handle = cpu_to_le16(conn->handle);
5529
5530 conn->pending_sec_level = smp_ltk_sec_level(ltk);
5531
5532 conn->enc_key_size = ltk->enc_size;
5533
5534 hci_send_cmd(hdev, HCI_OP_LE_LTK_REPLY, sizeof(cp), &cp);
5535
5536 /* Ref. Bluetooth Core SPEC pages 1975 and 2004. STK is a
5537 * temporary key used to encrypt a connection following
5538 * pairing. It is used during the Encrypted Session Setup to
5539 * distribute the keys. Later, security can be re-established
5540 * using a distributed LTK.
5541 */
5542 if (ltk->type == SMP_STK) {
5543 set_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
5544 list_del_rcu(<k->list);
5545 kfree_rcu(ltk, rcu);
5546 } else {
5547 clear_bit(HCI_CONN_STK_ENCRYPT, &conn->flags);
5548 }
5549
5550 hci_dev_unlock(hdev);
5551
5552 return;
5553
5554not_found:
5555 neg.handle = ev->handle;
5556 hci_send_cmd(hdev, HCI_OP_LE_LTK_NEG_REPLY, sizeof(neg), &neg);
5557 hci_dev_unlock(hdev);
5558}
5559
5560static void send_conn_param_neg_reply(struct hci_dev *hdev, u16 handle,
5561 u8 reason)
5562{
5563 struct hci_cp_le_conn_param_req_neg_reply cp;
5564
5565 cp.handle = cpu_to_le16(handle);
5566 cp.reason = reason;
5567
5568 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_NEG_REPLY, sizeof(cp),
5569 &cp);
5570}
5571
5572static void hci_le_remote_conn_param_req_evt(struct hci_dev *hdev,
5573 struct sk_buff *skb)
5574{
5575 struct hci_ev_le_remote_conn_param_req *ev = (void *) skb->data;
5576 struct hci_cp_le_conn_param_req_reply cp;
5577 struct hci_conn *hcon;
5578 u16 handle, min, max, latency, timeout;
5579
5580 handle = le16_to_cpu(ev->handle);
5581 min = le16_to_cpu(ev->interval_min);
5582 max = le16_to_cpu(ev->interval_max);
5583 latency = le16_to_cpu(ev->latency);
5584 timeout = le16_to_cpu(ev->timeout);
5585
5586 hcon = hci_conn_hash_lookup_handle(hdev, handle);
5587 if (!hcon || hcon->state != BT_CONNECTED)
5588 return send_conn_param_neg_reply(hdev, handle,
5589 HCI_ERROR_UNKNOWN_CONN_ID);
5590
5591 if (hci_check_conn_params(min, max, latency, timeout))
5592 return send_conn_param_neg_reply(hdev, handle,
5593 HCI_ERROR_INVALID_LL_PARAMS);
5594
5595 if (hcon->role == HCI_ROLE_MASTER) {
5596 struct hci_conn_params *params;
5597 u8 store_hint;
5598
5599 hci_dev_lock(hdev);
5600
5601 params = hci_conn_params_lookup(hdev, &hcon->dst,
5602 hcon->dst_type);
5603 if (params) {
5604 params->conn_min_interval = min;
5605 params->conn_max_interval = max;
5606 params->conn_latency = latency;
5607 params->supervision_timeout = timeout;
5608 store_hint = 0x01;
5609 } else{
5610 store_hint = 0x00;
5611 }
5612
5613 hci_dev_unlock(hdev);
5614
5615 mgmt_new_conn_param(hdev, &hcon->dst, hcon->dst_type,
5616 store_hint, min, max, latency, timeout);
5617 }
5618
5619 cp.handle = ev->handle;
5620 cp.interval_min = ev->interval_min;
5621 cp.interval_max = ev->interval_max;
5622 cp.latency = ev->latency;
5623 cp.timeout = ev->timeout;
5624 cp.min_ce_len = 0;
5625 cp.max_ce_len = 0;
5626
5627 hci_send_cmd(hdev, HCI_OP_LE_CONN_PARAM_REQ_REPLY, sizeof(cp), &cp);
5628}
5629
5630static void hci_le_direct_adv_report_evt(struct hci_dev *hdev,
5631 struct sk_buff *skb)
5632{
5633 u8 num_reports = skb->data[0];
5634 void *ptr = &skb->data[1];
5635
5636 hci_dev_lock(hdev);
5637
5638 while (num_reports--) {
5639 struct hci_ev_le_direct_adv_info *ev = ptr;
5640
5641 process_adv_report(hdev, ev->evt_type, &ev->bdaddr,
5642 ev->bdaddr_type, &ev->direct_addr,
5643 ev->direct_addr_type, ev->rssi, NULL, 0);
5644
5645 ptr += sizeof(*ev);
5646 }
5647
5648 hci_dev_unlock(hdev);
5649}
5650
5651static void hci_le_meta_evt(struct hci_dev *hdev, struct sk_buff *skb)
5652{
5653 struct hci_ev_le_meta *le_ev = (void *) skb->data;
5654
5655 skb_pull(skb, sizeof(*le_ev));
5656
5657 switch (le_ev->subevent) {
5658 case HCI_EV_LE_CONN_COMPLETE:
5659 hci_le_conn_complete_evt(hdev, skb);
5660 break;
5661
5662 case HCI_EV_LE_CONN_UPDATE_COMPLETE:
5663 hci_le_conn_update_complete_evt(hdev, skb);
5664 break;
5665
5666 case HCI_EV_LE_ADVERTISING_REPORT:
5667 hci_le_adv_report_evt(hdev, skb);
5668 break;
5669
5670 case HCI_EV_LE_REMOTE_FEAT_COMPLETE:
5671 hci_le_remote_feat_complete_evt(hdev, skb);
5672 break;
5673
5674 case HCI_EV_LE_LTK_REQ:
5675 hci_le_ltk_request_evt(hdev, skb);
5676 break;
5677
5678 case HCI_EV_LE_REMOTE_CONN_PARAM_REQ:
5679 hci_le_remote_conn_param_req_evt(hdev, skb);
5680 break;
5681
5682 case HCI_EV_LE_DIRECT_ADV_REPORT:
5683 hci_le_direct_adv_report_evt(hdev, skb);
5684 break;
5685
5686 case HCI_EV_LE_EXT_ADV_REPORT:
5687 hci_le_ext_adv_report_evt(hdev, skb);
5688 break;
5689
5690 case HCI_EV_LE_ENHANCED_CONN_COMPLETE:
5691 hci_le_enh_conn_complete_evt(hdev, skb);
5692 break;
5693
5694 case HCI_EV_LE_EXT_ADV_SET_TERM:
5695 hci_le_ext_adv_term_evt(hdev, skb);
5696 break;
5697
5698 default:
5699 break;
5700 }
5701}
5702
5703static bool hci_get_cmd_complete(struct hci_dev *hdev, u16 opcode,
5704 u8 event, struct sk_buff *skb)
5705{
5706 struct hci_ev_cmd_complete *ev;
5707 struct hci_event_hdr *hdr;
5708
5709 if (!skb)
5710 return false;
5711
5712 if (skb->len < sizeof(*hdr)) {
5713 bt_dev_err(hdev, "too short HCI event");
5714 return false;
5715 }
5716
5717 hdr = (void *) skb->data;
5718 skb_pull(skb, HCI_EVENT_HDR_SIZE);
5719
5720 if (event) {
5721 if (hdr->evt != event)
5722 return false;
5723 return true;
5724 }
5725
5726 /* Check if request ended in Command Status - no way to retreive
5727 * any extra parameters in this case.
5728 */
5729 if (hdr->evt == HCI_EV_CMD_STATUS)
5730 return false;
5731
5732 if (hdr->evt != HCI_EV_CMD_COMPLETE) {
5733 bt_dev_err(hdev, "last event is not cmd complete (0x%2.2x)",
5734 hdr->evt);
5735 return false;
5736 }
5737
5738 if (skb->len < sizeof(*ev)) {
5739 bt_dev_err(hdev, "too short cmd_complete event");
5740 return false;
5741 }
5742
5743 ev = (void *) skb->data;
5744 skb_pull(skb, sizeof(*ev));
5745
5746 if (opcode != __le16_to_cpu(ev->opcode)) {
5747 BT_DBG("opcode doesn't match (0x%2.2x != 0x%2.2x)", opcode,
5748 __le16_to_cpu(ev->opcode));
5749 return false;
5750 }
5751
5752 return true;
5753}
5754
5755void hci_event_packet(struct hci_dev *hdev, struct sk_buff *skb)
5756{
5757 struct hci_event_hdr *hdr = (void *) skb->data;
5758 hci_req_complete_t req_complete = NULL;
5759 hci_req_complete_skb_t req_complete_skb = NULL;
5760 struct sk_buff *orig_skb = NULL;
5761 u8 status = 0, event = hdr->evt, req_evt = 0;
5762 u16 opcode = HCI_OP_NOP;
5763
5764 if (hdev->sent_cmd && bt_cb(hdev->sent_cmd)->hci.req_event == event) {
5765 struct hci_command_hdr *cmd_hdr = (void *) hdev->sent_cmd->data;
5766 opcode = __le16_to_cpu(cmd_hdr->opcode);
5767 hci_req_cmd_complete(hdev, opcode, status, &req_complete,
5768 &req_complete_skb);
5769 req_evt = event;
5770 }
5771
5772 /* If it looks like we might end up having to call
5773 * req_complete_skb, store a pristine copy of the skb since the
5774 * various handlers may modify the original one through
5775 * skb_pull() calls, etc.
5776 */
5777 if (req_complete_skb || event == HCI_EV_CMD_STATUS ||
5778 event == HCI_EV_CMD_COMPLETE)
5779 orig_skb = skb_clone(skb, GFP_KERNEL);
5780
5781 skb_pull(skb, HCI_EVENT_HDR_SIZE);
5782
5783 switch (event) {
5784 case HCI_EV_INQUIRY_COMPLETE:
5785 hci_inquiry_complete_evt(hdev, skb);
5786 break;
5787
5788 case HCI_EV_INQUIRY_RESULT:
5789 hci_inquiry_result_evt(hdev, skb);
5790 break;
5791
5792 case HCI_EV_CONN_COMPLETE:
5793 hci_conn_complete_evt(hdev, skb);
5794 break;
5795
5796 case HCI_EV_CONN_REQUEST:
5797 hci_conn_request_evt(hdev, skb);
5798 break;
5799
5800 case HCI_EV_DISCONN_COMPLETE:
5801 hci_disconn_complete_evt(hdev, skb);
5802 break;
5803
5804 case HCI_EV_AUTH_COMPLETE:
5805 hci_auth_complete_evt(hdev, skb);
5806 break;
5807
5808 case HCI_EV_REMOTE_NAME:
5809 hci_remote_name_evt(hdev, skb);
5810 break;
5811
5812 case HCI_EV_ENCRYPT_CHANGE:
5813 hci_encrypt_change_evt(hdev, skb);
5814 break;
5815
5816 case HCI_EV_CHANGE_LINK_KEY_COMPLETE:
5817 hci_change_link_key_complete_evt(hdev, skb);
5818 break;
5819
5820 case HCI_EV_REMOTE_FEATURES:
5821 hci_remote_features_evt(hdev, skb);
5822 break;
5823
5824 case HCI_EV_CMD_COMPLETE:
5825 hci_cmd_complete_evt(hdev, skb, &opcode, &status,
5826 &req_complete, &req_complete_skb);
5827 break;
5828
5829 case HCI_EV_CMD_STATUS:
5830 hci_cmd_status_evt(hdev, skb, &opcode, &status, &req_complete,
5831 &req_complete_skb);
5832 break;
5833
5834 case HCI_EV_HARDWARE_ERROR:
5835 hci_hardware_error_evt(hdev, skb);
5836 break;
5837
5838 case HCI_EV_ROLE_CHANGE:
5839 hci_role_change_evt(hdev, skb);
5840 break;
5841
5842 case HCI_EV_NUM_COMP_PKTS:
5843 hci_num_comp_pkts_evt(hdev, skb);
5844 break;
5845
5846 case HCI_EV_MODE_CHANGE:
5847 hci_mode_change_evt(hdev, skb);
5848 break;
5849
5850 case HCI_EV_PIN_CODE_REQ:
5851 hci_pin_code_request_evt(hdev, skb);
5852 break;
5853
5854 case HCI_EV_LINK_KEY_REQ:
5855 hci_link_key_request_evt(hdev, skb);
5856 break;
5857
5858 case HCI_EV_LINK_KEY_NOTIFY:
5859 hci_link_key_notify_evt(hdev, skb);
5860 break;
5861
5862 case HCI_EV_CLOCK_OFFSET:
5863 hci_clock_offset_evt(hdev, skb);
5864 break;
5865
5866 case HCI_EV_PKT_TYPE_CHANGE:
5867 hci_pkt_type_change_evt(hdev, skb);
5868 break;
5869
5870 case HCI_EV_PSCAN_REP_MODE:
5871 hci_pscan_rep_mode_evt(hdev, skb);
5872 break;
5873
5874 case HCI_EV_INQUIRY_RESULT_WITH_RSSI:
5875 hci_inquiry_result_with_rssi_evt(hdev, skb);
5876 break;
5877
5878 case HCI_EV_REMOTE_EXT_FEATURES:
5879 hci_remote_ext_features_evt(hdev, skb);
5880 break;
5881
5882 case HCI_EV_SYNC_CONN_COMPLETE:
5883 hci_sync_conn_complete_evt(hdev, skb);
5884 break;
5885
5886 case HCI_EV_EXTENDED_INQUIRY_RESULT:
5887 hci_extended_inquiry_result_evt(hdev, skb);
5888 break;
5889
5890 case HCI_EV_KEY_REFRESH_COMPLETE:
5891 hci_key_refresh_complete_evt(hdev, skb);
5892 break;
5893
5894 case HCI_EV_IO_CAPA_REQUEST:
5895 hci_io_capa_request_evt(hdev, skb);
5896 break;
5897
5898 case HCI_EV_IO_CAPA_REPLY:
5899 hci_io_capa_reply_evt(hdev, skb);
5900 break;
5901
5902 case HCI_EV_USER_CONFIRM_REQUEST:
5903 hci_user_confirm_request_evt(hdev, skb);
5904 break;
5905
5906 case HCI_EV_USER_PASSKEY_REQUEST:
5907 hci_user_passkey_request_evt(hdev, skb);
5908 break;
5909
5910 case HCI_EV_USER_PASSKEY_NOTIFY:
5911 hci_user_passkey_notify_evt(hdev, skb);
5912 break;
5913
5914 case HCI_EV_KEYPRESS_NOTIFY:
5915 hci_keypress_notify_evt(hdev, skb);
5916 break;
5917
5918 case HCI_EV_SIMPLE_PAIR_COMPLETE:
5919 hci_simple_pair_complete_evt(hdev, skb);
5920 break;
5921
5922 case HCI_EV_REMOTE_HOST_FEATURES:
5923 hci_remote_host_features_evt(hdev, skb);
5924 break;
5925
5926 case HCI_EV_LE_META:
5927 hci_le_meta_evt(hdev, skb);
5928 break;
5929
5930 case HCI_EV_REMOTE_OOB_DATA_REQUEST:
5931 hci_remote_oob_data_request_evt(hdev, skb);
5932 break;
5933
5934#if IS_ENABLED(CONFIG_BT_HS)
5935 case HCI_EV_CHANNEL_SELECTED:
5936 hci_chan_selected_evt(hdev, skb);
5937 break;
5938
5939 case HCI_EV_PHY_LINK_COMPLETE:
5940 hci_phy_link_complete_evt(hdev, skb);
5941 break;
5942
5943 case HCI_EV_LOGICAL_LINK_COMPLETE:
5944 hci_loglink_complete_evt(hdev, skb);
5945 break;
5946
5947 case HCI_EV_DISCONN_LOGICAL_LINK_COMPLETE:
5948 hci_disconn_loglink_complete_evt(hdev, skb);
5949 break;
5950
5951 case HCI_EV_DISCONN_PHY_LINK_COMPLETE:
5952 hci_disconn_phylink_complete_evt(hdev, skb);
5953 break;
5954#endif
5955
5956 case HCI_EV_NUM_COMP_BLOCKS:
5957 hci_num_comp_blocks_evt(hdev, skb);
5958 break;
5959
5960 default:
5961 BT_DBG("%s event 0x%2.2x", hdev->name, event);
5962 break;
5963 }
5964
5965 if (req_complete) {
5966 req_complete(hdev, status, opcode);
5967 } else if (req_complete_skb) {
5968 if (!hci_get_cmd_complete(hdev, opcode, req_evt, orig_skb)) {
5969 kfree_skb(orig_skb);
5970 orig_skb = NULL;
5971 }
5972 req_complete_skb(hdev, status, opcode, orig_skb);
5973 }
5974
5975 kfree_skb(orig_skb);
5976 kfree_skb(skb);
5977 hdev->stat.evt_rx++;
5978}