inet: move inet->recverr_rfc4884 to inet->inet_flags
[platform/kernel/linux-rpi.git] / net / bluetooth / mgmt.c
1 /*
2    BlueZ - Bluetooth protocol stack for Linux
3
4    Copyright (C) 2010  Nokia Corporation
5    Copyright (C) 2011-2012 Intel Corporation
6
7    This program is free software; you can redistribute it and/or modify
8    it under the terms of the GNU General Public License version 2 as
9    published by the Free Software Foundation;
10
11    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
12    OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
13    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
14    IN NO EVENT SHALL THE COPYRIGHT HOLDER(S) AND AUTHOR(S) BE LIABLE FOR ANY
15    CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR ANY DAMAGES
16    WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
17    ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
18    OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
19
20    ALL LIABILITY, INCLUDING LIABILITY FOR INFRINGEMENT OF ANY PATENTS,
21    COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS, RELATING TO USE OF THIS
22    SOFTWARE IS DISCLAIMED.
23 */
24
25 /* Bluetooth HCI Management interface */
26
27 #include <linux/module.h>
28 #include <asm/unaligned.h>
29
30 #include <net/bluetooth/bluetooth.h>
31 #include <net/bluetooth/hci_core.h>
32 #include <net/bluetooth/hci_sock.h>
33 #include <net/bluetooth/l2cap.h>
34 #include <net/bluetooth/mgmt.h>
35
36 #include "hci_request.h"
37 #include "smp.h"
38 #include "mgmt_util.h"
39 #include "mgmt_config.h"
40 #include "msft.h"
41 #include "eir.h"
42 #include "aosp.h"
43
44 #define MGMT_VERSION    1
45 #define MGMT_REVISION   22
46
47 static const u16 mgmt_commands[] = {
48         MGMT_OP_READ_INDEX_LIST,
49         MGMT_OP_READ_INFO,
50         MGMT_OP_SET_POWERED,
51         MGMT_OP_SET_DISCOVERABLE,
52         MGMT_OP_SET_CONNECTABLE,
53         MGMT_OP_SET_FAST_CONNECTABLE,
54         MGMT_OP_SET_BONDABLE,
55         MGMT_OP_SET_LINK_SECURITY,
56         MGMT_OP_SET_SSP,
57         MGMT_OP_SET_HS,
58         MGMT_OP_SET_LE,
59         MGMT_OP_SET_DEV_CLASS,
60         MGMT_OP_SET_LOCAL_NAME,
61         MGMT_OP_ADD_UUID,
62         MGMT_OP_REMOVE_UUID,
63         MGMT_OP_LOAD_LINK_KEYS,
64         MGMT_OP_LOAD_LONG_TERM_KEYS,
65         MGMT_OP_DISCONNECT,
66         MGMT_OP_GET_CONNECTIONS,
67         MGMT_OP_PIN_CODE_REPLY,
68         MGMT_OP_PIN_CODE_NEG_REPLY,
69         MGMT_OP_SET_IO_CAPABILITY,
70         MGMT_OP_PAIR_DEVICE,
71         MGMT_OP_CANCEL_PAIR_DEVICE,
72         MGMT_OP_UNPAIR_DEVICE,
73         MGMT_OP_USER_CONFIRM_REPLY,
74         MGMT_OP_USER_CONFIRM_NEG_REPLY,
75         MGMT_OP_USER_PASSKEY_REPLY,
76         MGMT_OP_USER_PASSKEY_NEG_REPLY,
77         MGMT_OP_READ_LOCAL_OOB_DATA,
78         MGMT_OP_ADD_REMOTE_OOB_DATA,
79         MGMT_OP_REMOVE_REMOTE_OOB_DATA,
80         MGMT_OP_START_DISCOVERY,
81         MGMT_OP_STOP_DISCOVERY,
82         MGMT_OP_CONFIRM_NAME,
83         MGMT_OP_BLOCK_DEVICE,
84         MGMT_OP_UNBLOCK_DEVICE,
85         MGMT_OP_SET_DEVICE_ID,
86         MGMT_OP_SET_ADVERTISING,
87         MGMT_OP_SET_BREDR,
88         MGMT_OP_SET_STATIC_ADDRESS,
89         MGMT_OP_SET_SCAN_PARAMS,
90         MGMT_OP_SET_SECURE_CONN,
91         MGMT_OP_SET_DEBUG_KEYS,
92         MGMT_OP_SET_PRIVACY,
93         MGMT_OP_LOAD_IRKS,
94         MGMT_OP_GET_CONN_INFO,
95         MGMT_OP_GET_CLOCK_INFO,
96         MGMT_OP_ADD_DEVICE,
97         MGMT_OP_REMOVE_DEVICE,
98         MGMT_OP_LOAD_CONN_PARAM,
99         MGMT_OP_READ_UNCONF_INDEX_LIST,
100         MGMT_OP_READ_CONFIG_INFO,
101         MGMT_OP_SET_EXTERNAL_CONFIG,
102         MGMT_OP_SET_PUBLIC_ADDRESS,
103         MGMT_OP_START_SERVICE_DISCOVERY,
104         MGMT_OP_READ_LOCAL_OOB_EXT_DATA,
105         MGMT_OP_READ_EXT_INDEX_LIST,
106         MGMT_OP_READ_ADV_FEATURES,
107         MGMT_OP_ADD_ADVERTISING,
108         MGMT_OP_REMOVE_ADVERTISING,
109         MGMT_OP_GET_ADV_SIZE_INFO,
110         MGMT_OP_START_LIMITED_DISCOVERY,
111         MGMT_OP_READ_EXT_INFO,
112         MGMT_OP_SET_APPEARANCE,
113         MGMT_OP_GET_PHY_CONFIGURATION,
114         MGMT_OP_SET_PHY_CONFIGURATION,
115         MGMT_OP_SET_BLOCKED_KEYS,
116         MGMT_OP_SET_WIDEBAND_SPEECH,
117         MGMT_OP_READ_CONTROLLER_CAP,
118         MGMT_OP_READ_EXP_FEATURES_INFO,
119         MGMT_OP_SET_EXP_FEATURE,
120         MGMT_OP_READ_DEF_SYSTEM_CONFIG,
121         MGMT_OP_SET_DEF_SYSTEM_CONFIG,
122         MGMT_OP_READ_DEF_RUNTIME_CONFIG,
123         MGMT_OP_SET_DEF_RUNTIME_CONFIG,
124         MGMT_OP_GET_DEVICE_FLAGS,
125         MGMT_OP_SET_DEVICE_FLAGS,
126         MGMT_OP_READ_ADV_MONITOR_FEATURES,
127         MGMT_OP_ADD_ADV_PATTERNS_MONITOR,
128         MGMT_OP_REMOVE_ADV_MONITOR,
129         MGMT_OP_ADD_EXT_ADV_PARAMS,
130         MGMT_OP_ADD_EXT_ADV_DATA,
131         MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI,
132         MGMT_OP_SET_MESH_RECEIVER,
133         MGMT_OP_MESH_READ_FEATURES,
134         MGMT_OP_MESH_SEND,
135         MGMT_OP_MESH_SEND_CANCEL,
136 };
137
138 static const u16 mgmt_events[] = {
139         MGMT_EV_CONTROLLER_ERROR,
140         MGMT_EV_INDEX_ADDED,
141         MGMT_EV_INDEX_REMOVED,
142         MGMT_EV_NEW_SETTINGS,
143         MGMT_EV_CLASS_OF_DEV_CHANGED,
144         MGMT_EV_LOCAL_NAME_CHANGED,
145         MGMT_EV_NEW_LINK_KEY,
146         MGMT_EV_NEW_LONG_TERM_KEY,
147         MGMT_EV_DEVICE_CONNECTED,
148         MGMT_EV_DEVICE_DISCONNECTED,
149         MGMT_EV_CONNECT_FAILED,
150         MGMT_EV_PIN_CODE_REQUEST,
151         MGMT_EV_USER_CONFIRM_REQUEST,
152         MGMT_EV_USER_PASSKEY_REQUEST,
153         MGMT_EV_AUTH_FAILED,
154         MGMT_EV_DEVICE_FOUND,
155         MGMT_EV_DISCOVERING,
156         MGMT_EV_DEVICE_BLOCKED,
157         MGMT_EV_DEVICE_UNBLOCKED,
158         MGMT_EV_DEVICE_UNPAIRED,
159         MGMT_EV_PASSKEY_NOTIFY,
160         MGMT_EV_NEW_IRK,
161         MGMT_EV_NEW_CSRK,
162         MGMT_EV_DEVICE_ADDED,
163         MGMT_EV_DEVICE_REMOVED,
164         MGMT_EV_NEW_CONN_PARAM,
165         MGMT_EV_UNCONF_INDEX_ADDED,
166         MGMT_EV_UNCONF_INDEX_REMOVED,
167         MGMT_EV_NEW_CONFIG_OPTIONS,
168         MGMT_EV_EXT_INDEX_ADDED,
169         MGMT_EV_EXT_INDEX_REMOVED,
170         MGMT_EV_LOCAL_OOB_DATA_UPDATED,
171         MGMT_EV_ADVERTISING_ADDED,
172         MGMT_EV_ADVERTISING_REMOVED,
173         MGMT_EV_EXT_INFO_CHANGED,
174         MGMT_EV_PHY_CONFIGURATION_CHANGED,
175         MGMT_EV_EXP_FEATURE_CHANGED,
176         MGMT_EV_DEVICE_FLAGS_CHANGED,
177         MGMT_EV_ADV_MONITOR_ADDED,
178         MGMT_EV_ADV_MONITOR_REMOVED,
179         MGMT_EV_CONTROLLER_SUSPEND,
180         MGMT_EV_CONTROLLER_RESUME,
181         MGMT_EV_ADV_MONITOR_DEVICE_FOUND,
182         MGMT_EV_ADV_MONITOR_DEVICE_LOST,
183 };
184
185 static const u16 mgmt_untrusted_commands[] = {
186         MGMT_OP_READ_INDEX_LIST,
187         MGMT_OP_READ_INFO,
188         MGMT_OP_READ_UNCONF_INDEX_LIST,
189         MGMT_OP_READ_CONFIG_INFO,
190         MGMT_OP_READ_EXT_INDEX_LIST,
191         MGMT_OP_READ_EXT_INFO,
192         MGMT_OP_READ_CONTROLLER_CAP,
193         MGMT_OP_READ_EXP_FEATURES_INFO,
194         MGMT_OP_READ_DEF_SYSTEM_CONFIG,
195         MGMT_OP_READ_DEF_RUNTIME_CONFIG,
196 };
197
198 static const u16 mgmt_untrusted_events[] = {
199         MGMT_EV_INDEX_ADDED,
200         MGMT_EV_INDEX_REMOVED,
201         MGMT_EV_NEW_SETTINGS,
202         MGMT_EV_CLASS_OF_DEV_CHANGED,
203         MGMT_EV_LOCAL_NAME_CHANGED,
204         MGMT_EV_UNCONF_INDEX_ADDED,
205         MGMT_EV_UNCONF_INDEX_REMOVED,
206         MGMT_EV_NEW_CONFIG_OPTIONS,
207         MGMT_EV_EXT_INDEX_ADDED,
208         MGMT_EV_EXT_INDEX_REMOVED,
209         MGMT_EV_EXT_INFO_CHANGED,
210         MGMT_EV_EXP_FEATURE_CHANGED,
211 };
212
213 #define CACHE_TIMEOUT   msecs_to_jiffies(2 * 1000)
214
215 #define ZERO_KEY "\x00\x00\x00\x00\x00\x00\x00\x00" \
216                  "\x00\x00\x00\x00\x00\x00\x00\x00"
217
218 /* HCI to MGMT error code conversion table */
219 static const u8 mgmt_status_table[] = {
220         MGMT_STATUS_SUCCESS,
221         MGMT_STATUS_UNKNOWN_COMMAND,    /* Unknown Command */
222         MGMT_STATUS_NOT_CONNECTED,      /* No Connection */
223         MGMT_STATUS_FAILED,             /* Hardware Failure */
224         MGMT_STATUS_CONNECT_FAILED,     /* Page Timeout */
225         MGMT_STATUS_AUTH_FAILED,        /* Authentication Failed */
226         MGMT_STATUS_AUTH_FAILED,        /* PIN or Key Missing */
227         MGMT_STATUS_NO_RESOURCES,       /* Memory Full */
228         MGMT_STATUS_TIMEOUT,            /* Connection Timeout */
229         MGMT_STATUS_NO_RESOURCES,       /* Max Number of Connections */
230         MGMT_STATUS_NO_RESOURCES,       /* Max Number of SCO Connections */
231         MGMT_STATUS_ALREADY_CONNECTED,  /* ACL Connection Exists */
232         MGMT_STATUS_BUSY,               /* Command Disallowed */
233         MGMT_STATUS_NO_RESOURCES,       /* Rejected Limited Resources */
234         MGMT_STATUS_REJECTED,           /* Rejected Security */
235         MGMT_STATUS_REJECTED,           /* Rejected Personal */
236         MGMT_STATUS_TIMEOUT,            /* Host Timeout */
237         MGMT_STATUS_NOT_SUPPORTED,      /* Unsupported Feature */
238         MGMT_STATUS_INVALID_PARAMS,     /* Invalid Parameters */
239         MGMT_STATUS_DISCONNECTED,       /* OE User Ended Connection */
240         MGMT_STATUS_NO_RESOURCES,       /* OE Low Resources */
241         MGMT_STATUS_DISCONNECTED,       /* OE Power Off */
242         MGMT_STATUS_DISCONNECTED,       /* Connection Terminated */
243         MGMT_STATUS_BUSY,               /* Repeated Attempts */
244         MGMT_STATUS_REJECTED,           /* Pairing Not Allowed */
245         MGMT_STATUS_FAILED,             /* Unknown LMP PDU */
246         MGMT_STATUS_NOT_SUPPORTED,      /* Unsupported Remote Feature */
247         MGMT_STATUS_REJECTED,           /* SCO Offset Rejected */
248         MGMT_STATUS_REJECTED,           /* SCO Interval Rejected */
249         MGMT_STATUS_REJECTED,           /* Air Mode Rejected */
250         MGMT_STATUS_INVALID_PARAMS,     /* Invalid LMP Parameters */
251         MGMT_STATUS_FAILED,             /* Unspecified Error */
252         MGMT_STATUS_NOT_SUPPORTED,      /* Unsupported LMP Parameter Value */
253         MGMT_STATUS_FAILED,             /* Role Change Not Allowed */
254         MGMT_STATUS_TIMEOUT,            /* LMP Response Timeout */
255         MGMT_STATUS_FAILED,             /* LMP Error Transaction Collision */
256         MGMT_STATUS_FAILED,             /* LMP PDU Not Allowed */
257         MGMT_STATUS_REJECTED,           /* Encryption Mode Not Accepted */
258         MGMT_STATUS_FAILED,             /* Unit Link Key Used */
259         MGMT_STATUS_NOT_SUPPORTED,      /* QoS Not Supported */
260         MGMT_STATUS_TIMEOUT,            /* Instant Passed */
261         MGMT_STATUS_NOT_SUPPORTED,      /* Pairing Not Supported */
262         MGMT_STATUS_FAILED,             /* Transaction Collision */
263         MGMT_STATUS_FAILED,             /* Reserved for future use */
264         MGMT_STATUS_INVALID_PARAMS,     /* Unacceptable Parameter */
265         MGMT_STATUS_REJECTED,           /* QoS Rejected */
266         MGMT_STATUS_NOT_SUPPORTED,      /* Classification Not Supported */
267         MGMT_STATUS_REJECTED,           /* Insufficient Security */
268         MGMT_STATUS_INVALID_PARAMS,     /* Parameter Out Of Range */
269         MGMT_STATUS_FAILED,             /* Reserved for future use */
270         MGMT_STATUS_BUSY,               /* Role Switch Pending */
271         MGMT_STATUS_FAILED,             /* Reserved for future use */
272         MGMT_STATUS_FAILED,             /* Slot Violation */
273         MGMT_STATUS_FAILED,             /* Role Switch Failed */
274         MGMT_STATUS_INVALID_PARAMS,     /* EIR Too Large */
275         MGMT_STATUS_NOT_SUPPORTED,      /* Simple Pairing Not Supported */
276         MGMT_STATUS_BUSY,               /* Host Busy Pairing */
277         MGMT_STATUS_REJECTED,           /* Rejected, No Suitable Channel */
278         MGMT_STATUS_BUSY,               /* Controller Busy */
279         MGMT_STATUS_INVALID_PARAMS,     /* Unsuitable Connection Interval */
280         MGMT_STATUS_TIMEOUT,            /* Directed Advertising Timeout */
281         MGMT_STATUS_AUTH_FAILED,        /* Terminated Due to MIC Failure */
282         MGMT_STATUS_CONNECT_FAILED,     /* Connection Establishment Failed */
283         MGMT_STATUS_CONNECT_FAILED,     /* MAC Connection Failed */
284 };
285
286 static u8 mgmt_errno_status(int err)
287 {
288         switch (err) {
289         case 0:
290                 return MGMT_STATUS_SUCCESS;
291         case -EPERM:
292                 return MGMT_STATUS_REJECTED;
293         case -EINVAL:
294                 return MGMT_STATUS_INVALID_PARAMS;
295         case -EOPNOTSUPP:
296                 return MGMT_STATUS_NOT_SUPPORTED;
297         case -EBUSY:
298                 return MGMT_STATUS_BUSY;
299         case -ETIMEDOUT:
300                 return MGMT_STATUS_AUTH_FAILED;
301         case -ENOMEM:
302                 return MGMT_STATUS_NO_RESOURCES;
303         case -EISCONN:
304                 return MGMT_STATUS_ALREADY_CONNECTED;
305         case -ENOTCONN:
306                 return MGMT_STATUS_DISCONNECTED;
307         }
308
309         return MGMT_STATUS_FAILED;
310 }
311
312 static u8 mgmt_status(int err)
313 {
314         if (err < 0)
315                 return mgmt_errno_status(err);
316
317         if (err < ARRAY_SIZE(mgmt_status_table))
318                 return mgmt_status_table[err];
319
320         return MGMT_STATUS_FAILED;
321 }
322
323 static int mgmt_index_event(u16 event, struct hci_dev *hdev, void *data,
324                             u16 len, int flag)
325 {
326         return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len,
327                                flag, NULL);
328 }
329
330 static int mgmt_limited_event(u16 event, struct hci_dev *hdev, void *data,
331                               u16 len, int flag, struct sock *skip_sk)
332 {
333         return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len,
334                                flag, skip_sk);
335 }
336
337 static int mgmt_event(u16 event, struct hci_dev *hdev, void *data, u16 len,
338                       struct sock *skip_sk)
339 {
340         return mgmt_send_event(event, hdev, HCI_CHANNEL_CONTROL, data, len,
341                                HCI_SOCK_TRUSTED, skip_sk);
342 }
343
344 static int mgmt_event_skb(struct sk_buff *skb, struct sock *skip_sk)
345 {
346         return mgmt_send_event_skb(HCI_CHANNEL_CONTROL, skb, HCI_SOCK_TRUSTED,
347                                    skip_sk);
348 }
349
350 static u8 le_addr_type(u8 mgmt_addr_type)
351 {
352         if (mgmt_addr_type == BDADDR_LE_PUBLIC)
353                 return ADDR_LE_DEV_PUBLIC;
354         else
355                 return ADDR_LE_DEV_RANDOM;
356 }
357
358 void mgmt_fill_version_info(void *ver)
359 {
360         struct mgmt_rp_read_version *rp = ver;
361
362         rp->version = MGMT_VERSION;
363         rp->revision = cpu_to_le16(MGMT_REVISION);
364 }
365
366 static int read_version(struct sock *sk, struct hci_dev *hdev, void *data,
367                         u16 data_len)
368 {
369         struct mgmt_rp_read_version rp;
370
371         bt_dev_dbg(hdev, "sock %p", sk);
372
373         mgmt_fill_version_info(&rp);
374
375         return mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_VERSION, 0,
376                                  &rp, sizeof(rp));
377 }
378
379 static int read_commands(struct sock *sk, struct hci_dev *hdev, void *data,
380                          u16 data_len)
381 {
382         struct mgmt_rp_read_commands *rp;
383         u16 num_commands, num_events;
384         size_t rp_size;
385         int i, err;
386
387         bt_dev_dbg(hdev, "sock %p", sk);
388
389         if (hci_sock_test_flag(sk, HCI_SOCK_TRUSTED)) {
390                 num_commands = ARRAY_SIZE(mgmt_commands);
391                 num_events = ARRAY_SIZE(mgmt_events);
392         } else {
393                 num_commands = ARRAY_SIZE(mgmt_untrusted_commands);
394                 num_events = ARRAY_SIZE(mgmt_untrusted_events);
395         }
396
397         rp_size = sizeof(*rp) + ((num_commands + num_events) * sizeof(u16));
398
399         rp = kmalloc(rp_size, GFP_KERNEL);
400         if (!rp)
401                 return -ENOMEM;
402
403         rp->num_commands = cpu_to_le16(num_commands);
404         rp->num_events = cpu_to_le16(num_events);
405
406         if (hci_sock_test_flag(sk, HCI_SOCK_TRUSTED)) {
407                 __le16 *opcode = rp->opcodes;
408
409                 for (i = 0; i < num_commands; i++, opcode++)
410                         put_unaligned_le16(mgmt_commands[i], opcode);
411
412                 for (i = 0; i < num_events; i++, opcode++)
413                         put_unaligned_le16(mgmt_events[i], opcode);
414         } else {
415                 __le16 *opcode = rp->opcodes;
416
417                 for (i = 0; i < num_commands; i++, opcode++)
418                         put_unaligned_le16(mgmt_untrusted_commands[i], opcode);
419
420                 for (i = 0; i < num_events; i++, opcode++)
421                         put_unaligned_le16(mgmt_untrusted_events[i], opcode);
422         }
423
424         err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_COMMANDS, 0,
425                                 rp, rp_size);
426         kfree(rp);
427
428         return err;
429 }
430
431 static int read_index_list(struct sock *sk, struct hci_dev *hdev, void *data,
432                            u16 data_len)
433 {
434         struct mgmt_rp_read_index_list *rp;
435         struct hci_dev *d;
436         size_t rp_len;
437         u16 count;
438         int err;
439
440         bt_dev_dbg(hdev, "sock %p", sk);
441
442         read_lock(&hci_dev_list_lock);
443
444         count = 0;
445         list_for_each_entry(d, &hci_dev_list, list) {
446                 if (d->dev_type == HCI_PRIMARY &&
447                     !hci_dev_test_flag(d, HCI_UNCONFIGURED))
448                         count++;
449         }
450
451         rp_len = sizeof(*rp) + (2 * count);
452         rp = kmalloc(rp_len, GFP_ATOMIC);
453         if (!rp) {
454                 read_unlock(&hci_dev_list_lock);
455                 return -ENOMEM;
456         }
457
458         count = 0;
459         list_for_each_entry(d, &hci_dev_list, list) {
460                 if (hci_dev_test_flag(d, HCI_SETUP) ||
461                     hci_dev_test_flag(d, HCI_CONFIG) ||
462                     hci_dev_test_flag(d, HCI_USER_CHANNEL))
463                         continue;
464
465                 /* Devices marked as raw-only are neither configured
466                  * nor unconfigured controllers.
467                  */
468                 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks))
469                         continue;
470
471                 if (d->dev_type == HCI_PRIMARY &&
472                     !hci_dev_test_flag(d, HCI_UNCONFIGURED)) {
473                         rp->index[count++] = cpu_to_le16(d->id);
474                         bt_dev_dbg(hdev, "Added hci%u", d->id);
475                 }
476         }
477
478         rp->num_controllers = cpu_to_le16(count);
479         rp_len = sizeof(*rp) + (2 * count);
480
481         read_unlock(&hci_dev_list_lock);
482
483         err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, MGMT_OP_READ_INDEX_LIST,
484                                 0, rp, rp_len);
485
486         kfree(rp);
487
488         return err;
489 }
490
491 static int read_unconf_index_list(struct sock *sk, struct hci_dev *hdev,
492                                   void *data, u16 data_len)
493 {
494         struct mgmt_rp_read_unconf_index_list *rp;
495         struct hci_dev *d;
496         size_t rp_len;
497         u16 count;
498         int err;
499
500         bt_dev_dbg(hdev, "sock %p", sk);
501
502         read_lock(&hci_dev_list_lock);
503
504         count = 0;
505         list_for_each_entry(d, &hci_dev_list, list) {
506                 if (d->dev_type == HCI_PRIMARY &&
507                     hci_dev_test_flag(d, HCI_UNCONFIGURED))
508                         count++;
509         }
510
511         rp_len = sizeof(*rp) + (2 * count);
512         rp = kmalloc(rp_len, GFP_ATOMIC);
513         if (!rp) {
514                 read_unlock(&hci_dev_list_lock);
515                 return -ENOMEM;
516         }
517
518         count = 0;
519         list_for_each_entry(d, &hci_dev_list, list) {
520                 if (hci_dev_test_flag(d, HCI_SETUP) ||
521                     hci_dev_test_flag(d, HCI_CONFIG) ||
522                     hci_dev_test_flag(d, HCI_USER_CHANNEL))
523                         continue;
524
525                 /* Devices marked as raw-only are neither configured
526                  * nor unconfigured controllers.
527                  */
528                 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks))
529                         continue;
530
531                 if (d->dev_type == HCI_PRIMARY &&
532                     hci_dev_test_flag(d, HCI_UNCONFIGURED)) {
533                         rp->index[count++] = cpu_to_le16(d->id);
534                         bt_dev_dbg(hdev, "Added hci%u", d->id);
535                 }
536         }
537
538         rp->num_controllers = cpu_to_le16(count);
539         rp_len = sizeof(*rp) + (2 * count);
540
541         read_unlock(&hci_dev_list_lock);
542
543         err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE,
544                                 MGMT_OP_READ_UNCONF_INDEX_LIST, 0, rp, rp_len);
545
546         kfree(rp);
547
548         return err;
549 }
550
551 static int read_ext_index_list(struct sock *sk, struct hci_dev *hdev,
552                                void *data, u16 data_len)
553 {
554         struct mgmt_rp_read_ext_index_list *rp;
555         struct hci_dev *d;
556         u16 count;
557         int err;
558
559         bt_dev_dbg(hdev, "sock %p", sk);
560
561         read_lock(&hci_dev_list_lock);
562
563         count = 0;
564         list_for_each_entry(d, &hci_dev_list, list) {
565                 if (d->dev_type == HCI_PRIMARY || d->dev_type == HCI_AMP)
566                         count++;
567         }
568
569         rp = kmalloc(struct_size(rp, entry, count), GFP_ATOMIC);
570         if (!rp) {
571                 read_unlock(&hci_dev_list_lock);
572                 return -ENOMEM;
573         }
574
575         count = 0;
576         list_for_each_entry(d, &hci_dev_list, list) {
577                 if (hci_dev_test_flag(d, HCI_SETUP) ||
578                     hci_dev_test_flag(d, HCI_CONFIG) ||
579                     hci_dev_test_flag(d, HCI_USER_CHANNEL))
580                         continue;
581
582                 /* Devices marked as raw-only are neither configured
583                  * nor unconfigured controllers.
584                  */
585                 if (test_bit(HCI_QUIRK_RAW_DEVICE, &d->quirks))
586                         continue;
587
588                 if (d->dev_type == HCI_PRIMARY) {
589                         if (hci_dev_test_flag(d, HCI_UNCONFIGURED))
590                                 rp->entry[count].type = 0x01;
591                         else
592                                 rp->entry[count].type = 0x00;
593                 } else if (d->dev_type == HCI_AMP) {
594                         rp->entry[count].type = 0x02;
595                 } else {
596                         continue;
597                 }
598
599                 rp->entry[count].bus = d->bus;
600                 rp->entry[count++].index = cpu_to_le16(d->id);
601                 bt_dev_dbg(hdev, "Added hci%u", d->id);
602         }
603
604         rp->num_controllers = cpu_to_le16(count);
605
606         read_unlock(&hci_dev_list_lock);
607
608         /* If this command is called at least once, then all the
609          * default index and unconfigured index events are disabled
610          * and from now on only extended index events are used.
611          */
612         hci_sock_set_flag(sk, HCI_MGMT_EXT_INDEX_EVENTS);
613         hci_sock_clear_flag(sk, HCI_MGMT_INDEX_EVENTS);
614         hci_sock_clear_flag(sk, HCI_MGMT_UNCONF_INDEX_EVENTS);
615
616         err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE,
617                                 MGMT_OP_READ_EXT_INDEX_LIST, 0, rp,
618                                 struct_size(rp, entry, count));
619
620         kfree(rp);
621
622         return err;
623 }
624
625 static bool is_configured(struct hci_dev *hdev)
626 {
627         if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) &&
628             !hci_dev_test_flag(hdev, HCI_EXT_CONFIGURED))
629                 return false;
630
631         if ((test_bit(HCI_QUIRK_INVALID_BDADDR, &hdev->quirks) ||
632              test_bit(HCI_QUIRK_USE_BDADDR_PROPERTY, &hdev->quirks)) &&
633             !bacmp(&hdev->public_addr, BDADDR_ANY))
634                 return false;
635
636         return true;
637 }
638
639 static __le32 get_missing_options(struct hci_dev *hdev)
640 {
641         u32 options = 0;
642
643         if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) &&
644             !hci_dev_test_flag(hdev, HCI_EXT_CONFIGURED))
645                 options |= MGMT_OPTION_EXTERNAL_CONFIG;
646
647         if ((test_bit(HCI_QUIRK_INVALID_BDADDR, &hdev->quirks) ||
648              test_bit(HCI_QUIRK_USE_BDADDR_PROPERTY, &hdev->quirks)) &&
649             !bacmp(&hdev->public_addr, BDADDR_ANY))
650                 options |= MGMT_OPTION_PUBLIC_ADDRESS;
651
652         return cpu_to_le32(options);
653 }
654
655 static int new_options(struct hci_dev *hdev, struct sock *skip)
656 {
657         __le32 options = get_missing_options(hdev);
658
659         return mgmt_limited_event(MGMT_EV_NEW_CONFIG_OPTIONS, hdev, &options,
660                                   sizeof(options), HCI_MGMT_OPTION_EVENTS, skip);
661 }
662
663 static int send_options_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev)
664 {
665         __le32 options = get_missing_options(hdev);
666
667         return mgmt_cmd_complete(sk, hdev->id, opcode, 0, &options,
668                                  sizeof(options));
669 }
670
671 static int read_config_info(struct sock *sk, struct hci_dev *hdev,
672                             void *data, u16 data_len)
673 {
674         struct mgmt_rp_read_config_info rp;
675         u32 options = 0;
676
677         bt_dev_dbg(hdev, "sock %p", sk);
678
679         hci_dev_lock(hdev);
680
681         memset(&rp, 0, sizeof(rp));
682         rp.manufacturer = cpu_to_le16(hdev->manufacturer);
683
684         if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks))
685                 options |= MGMT_OPTION_EXTERNAL_CONFIG;
686
687         if (hdev->set_bdaddr)
688                 options |= MGMT_OPTION_PUBLIC_ADDRESS;
689
690         rp.supported_options = cpu_to_le32(options);
691         rp.missing_options = get_missing_options(hdev);
692
693         hci_dev_unlock(hdev);
694
695         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_CONFIG_INFO, 0,
696                                  &rp, sizeof(rp));
697 }
698
699 static u32 get_supported_phys(struct hci_dev *hdev)
700 {
701         u32 supported_phys = 0;
702
703         if (lmp_bredr_capable(hdev)) {
704                 supported_phys |= MGMT_PHY_BR_1M_1SLOT;
705
706                 if (hdev->features[0][0] & LMP_3SLOT)
707                         supported_phys |= MGMT_PHY_BR_1M_3SLOT;
708
709                 if (hdev->features[0][0] & LMP_5SLOT)
710                         supported_phys |= MGMT_PHY_BR_1M_5SLOT;
711
712                 if (lmp_edr_2m_capable(hdev)) {
713                         supported_phys |= MGMT_PHY_EDR_2M_1SLOT;
714
715                         if (lmp_edr_3slot_capable(hdev))
716                                 supported_phys |= MGMT_PHY_EDR_2M_3SLOT;
717
718                         if (lmp_edr_5slot_capable(hdev))
719                                 supported_phys |= MGMT_PHY_EDR_2M_5SLOT;
720
721                         if (lmp_edr_3m_capable(hdev)) {
722                                 supported_phys |= MGMT_PHY_EDR_3M_1SLOT;
723
724                                 if (lmp_edr_3slot_capable(hdev))
725                                         supported_phys |= MGMT_PHY_EDR_3M_3SLOT;
726
727                                 if (lmp_edr_5slot_capable(hdev))
728                                         supported_phys |= MGMT_PHY_EDR_3M_5SLOT;
729                         }
730                 }
731         }
732
733         if (lmp_le_capable(hdev)) {
734                 supported_phys |= MGMT_PHY_LE_1M_TX;
735                 supported_phys |= MGMT_PHY_LE_1M_RX;
736
737                 if (hdev->le_features[1] & HCI_LE_PHY_2M) {
738                         supported_phys |= MGMT_PHY_LE_2M_TX;
739                         supported_phys |= MGMT_PHY_LE_2M_RX;
740                 }
741
742                 if (hdev->le_features[1] & HCI_LE_PHY_CODED) {
743                         supported_phys |= MGMT_PHY_LE_CODED_TX;
744                         supported_phys |= MGMT_PHY_LE_CODED_RX;
745                 }
746         }
747
748         return supported_phys;
749 }
750
751 static u32 get_selected_phys(struct hci_dev *hdev)
752 {
753         u32 selected_phys = 0;
754
755         if (lmp_bredr_capable(hdev)) {
756                 selected_phys |= MGMT_PHY_BR_1M_1SLOT;
757
758                 if (hdev->pkt_type & (HCI_DM3 | HCI_DH3))
759                         selected_phys |= MGMT_PHY_BR_1M_3SLOT;
760
761                 if (hdev->pkt_type & (HCI_DM5 | HCI_DH5))
762                         selected_phys |= MGMT_PHY_BR_1M_5SLOT;
763
764                 if (lmp_edr_2m_capable(hdev)) {
765                         if (!(hdev->pkt_type & HCI_2DH1))
766                                 selected_phys |= MGMT_PHY_EDR_2M_1SLOT;
767
768                         if (lmp_edr_3slot_capable(hdev) &&
769                             !(hdev->pkt_type & HCI_2DH3))
770                                 selected_phys |= MGMT_PHY_EDR_2M_3SLOT;
771
772                         if (lmp_edr_5slot_capable(hdev) &&
773                             !(hdev->pkt_type & HCI_2DH5))
774                                 selected_phys |= MGMT_PHY_EDR_2M_5SLOT;
775
776                         if (lmp_edr_3m_capable(hdev)) {
777                                 if (!(hdev->pkt_type & HCI_3DH1))
778                                         selected_phys |= MGMT_PHY_EDR_3M_1SLOT;
779
780                                 if (lmp_edr_3slot_capable(hdev) &&
781                                     !(hdev->pkt_type & HCI_3DH3))
782                                         selected_phys |= MGMT_PHY_EDR_3M_3SLOT;
783
784                                 if (lmp_edr_5slot_capable(hdev) &&
785                                     !(hdev->pkt_type & HCI_3DH5))
786                                         selected_phys |= MGMT_PHY_EDR_3M_5SLOT;
787                         }
788                 }
789         }
790
791         if (lmp_le_capable(hdev)) {
792                 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_1M)
793                         selected_phys |= MGMT_PHY_LE_1M_TX;
794
795                 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_1M)
796                         selected_phys |= MGMT_PHY_LE_1M_RX;
797
798                 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_2M)
799                         selected_phys |= MGMT_PHY_LE_2M_TX;
800
801                 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_2M)
802                         selected_phys |= MGMT_PHY_LE_2M_RX;
803
804                 if (hdev->le_tx_def_phys & HCI_LE_SET_PHY_CODED)
805                         selected_phys |= MGMT_PHY_LE_CODED_TX;
806
807                 if (hdev->le_rx_def_phys & HCI_LE_SET_PHY_CODED)
808                         selected_phys |= MGMT_PHY_LE_CODED_RX;
809         }
810
811         return selected_phys;
812 }
813
814 static u32 get_configurable_phys(struct hci_dev *hdev)
815 {
816         return (get_supported_phys(hdev) & ~MGMT_PHY_BR_1M_1SLOT &
817                 ~MGMT_PHY_LE_1M_TX & ~MGMT_PHY_LE_1M_RX);
818 }
819
820 static u32 get_supported_settings(struct hci_dev *hdev)
821 {
822         u32 settings = 0;
823
824         settings |= MGMT_SETTING_POWERED;
825         settings |= MGMT_SETTING_BONDABLE;
826         settings |= MGMT_SETTING_DEBUG_KEYS;
827         settings |= MGMT_SETTING_CONNECTABLE;
828         settings |= MGMT_SETTING_DISCOVERABLE;
829
830         if (lmp_bredr_capable(hdev)) {
831                 if (hdev->hci_ver >= BLUETOOTH_VER_1_2)
832                         settings |= MGMT_SETTING_FAST_CONNECTABLE;
833                 settings |= MGMT_SETTING_BREDR;
834                 settings |= MGMT_SETTING_LINK_SECURITY;
835
836                 if (lmp_ssp_capable(hdev)) {
837                         settings |= MGMT_SETTING_SSP;
838                         if (IS_ENABLED(CONFIG_BT_HS))
839                                 settings |= MGMT_SETTING_HS;
840                 }
841
842                 if (lmp_sc_capable(hdev))
843                         settings |= MGMT_SETTING_SECURE_CONN;
844
845                 if (test_bit(HCI_QUIRK_WIDEBAND_SPEECH_SUPPORTED,
846                              &hdev->quirks))
847                         settings |= MGMT_SETTING_WIDEBAND_SPEECH;
848         }
849
850         if (lmp_le_capable(hdev)) {
851                 settings |= MGMT_SETTING_LE;
852                 settings |= MGMT_SETTING_SECURE_CONN;
853                 settings |= MGMT_SETTING_PRIVACY;
854                 settings |= MGMT_SETTING_STATIC_ADDRESS;
855                 settings |= MGMT_SETTING_ADVERTISING;
856         }
857
858         if (test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks) ||
859             hdev->set_bdaddr)
860                 settings |= MGMT_SETTING_CONFIGURATION;
861
862         if (cis_central_capable(hdev))
863                 settings |= MGMT_SETTING_CIS_CENTRAL;
864
865         if (cis_peripheral_capable(hdev))
866                 settings |= MGMT_SETTING_CIS_PERIPHERAL;
867
868         settings |= MGMT_SETTING_PHY_CONFIGURATION;
869
870         return settings;
871 }
872
873 static u32 get_current_settings(struct hci_dev *hdev)
874 {
875         u32 settings = 0;
876
877         if (hdev_is_powered(hdev))
878                 settings |= MGMT_SETTING_POWERED;
879
880         if (hci_dev_test_flag(hdev, HCI_CONNECTABLE))
881                 settings |= MGMT_SETTING_CONNECTABLE;
882
883         if (hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE))
884                 settings |= MGMT_SETTING_FAST_CONNECTABLE;
885
886         if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
887                 settings |= MGMT_SETTING_DISCOVERABLE;
888
889         if (hci_dev_test_flag(hdev, HCI_BONDABLE))
890                 settings |= MGMT_SETTING_BONDABLE;
891
892         if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
893                 settings |= MGMT_SETTING_BREDR;
894
895         if (hci_dev_test_flag(hdev, HCI_LE_ENABLED))
896                 settings |= MGMT_SETTING_LE;
897
898         if (hci_dev_test_flag(hdev, HCI_LINK_SECURITY))
899                 settings |= MGMT_SETTING_LINK_SECURITY;
900
901         if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED))
902                 settings |= MGMT_SETTING_SSP;
903
904         if (hci_dev_test_flag(hdev, HCI_HS_ENABLED))
905                 settings |= MGMT_SETTING_HS;
906
907         if (hci_dev_test_flag(hdev, HCI_ADVERTISING))
908                 settings |= MGMT_SETTING_ADVERTISING;
909
910         if (hci_dev_test_flag(hdev, HCI_SC_ENABLED))
911                 settings |= MGMT_SETTING_SECURE_CONN;
912
913         if (hci_dev_test_flag(hdev, HCI_KEEP_DEBUG_KEYS))
914                 settings |= MGMT_SETTING_DEBUG_KEYS;
915
916         if (hci_dev_test_flag(hdev, HCI_PRIVACY))
917                 settings |= MGMT_SETTING_PRIVACY;
918
919         /* The current setting for static address has two purposes. The
920          * first is to indicate if the static address will be used and
921          * the second is to indicate if it is actually set.
922          *
923          * This means if the static address is not configured, this flag
924          * will never be set. If the address is configured, then if the
925          * address is actually used decides if the flag is set or not.
926          *
927          * For single mode LE only controllers and dual-mode controllers
928          * with BR/EDR disabled, the existence of the static address will
929          * be evaluated.
930          */
931         if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) ||
932             !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) ||
933             !bacmp(&hdev->bdaddr, BDADDR_ANY)) {
934                 if (bacmp(&hdev->static_addr, BDADDR_ANY))
935                         settings |= MGMT_SETTING_STATIC_ADDRESS;
936         }
937
938         if (hci_dev_test_flag(hdev, HCI_WIDEBAND_SPEECH_ENABLED))
939                 settings |= MGMT_SETTING_WIDEBAND_SPEECH;
940
941         if (cis_central_capable(hdev))
942                 settings |= MGMT_SETTING_CIS_CENTRAL;
943
944         if (cis_peripheral_capable(hdev))
945                 settings |= MGMT_SETTING_CIS_PERIPHERAL;
946
947         if (bis_capable(hdev))
948                 settings |= MGMT_SETTING_ISO_BROADCASTER;
949
950         if (sync_recv_capable(hdev))
951                 settings |= MGMT_SETTING_ISO_SYNC_RECEIVER;
952
953         return settings;
954 }
955
956 static struct mgmt_pending_cmd *pending_find(u16 opcode, struct hci_dev *hdev)
957 {
958         return mgmt_pending_find(HCI_CHANNEL_CONTROL, opcode, hdev);
959 }
960
961 u8 mgmt_get_adv_discov_flags(struct hci_dev *hdev)
962 {
963         struct mgmt_pending_cmd *cmd;
964
965         /* If there's a pending mgmt command the flags will not yet have
966          * their final values, so check for this first.
967          */
968         cmd = pending_find(MGMT_OP_SET_DISCOVERABLE, hdev);
969         if (cmd) {
970                 struct mgmt_mode *cp = cmd->param;
971                 if (cp->val == 0x01)
972                         return LE_AD_GENERAL;
973                 else if (cp->val == 0x02)
974                         return LE_AD_LIMITED;
975         } else {
976                 if (hci_dev_test_flag(hdev, HCI_LIMITED_DISCOVERABLE))
977                         return LE_AD_LIMITED;
978                 else if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE))
979                         return LE_AD_GENERAL;
980         }
981
982         return 0;
983 }
984
985 bool mgmt_get_connectable(struct hci_dev *hdev)
986 {
987         struct mgmt_pending_cmd *cmd;
988
989         /* If there's a pending mgmt command the flag will not yet have
990          * it's final value, so check for this first.
991          */
992         cmd = pending_find(MGMT_OP_SET_CONNECTABLE, hdev);
993         if (cmd) {
994                 struct mgmt_mode *cp = cmd->param;
995
996                 return cp->val;
997         }
998
999         return hci_dev_test_flag(hdev, HCI_CONNECTABLE);
1000 }
1001
1002 static int service_cache_sync(struct hci_dev *hdev, void *data)
1003 {
1004         hci_update_eir_sync(hdev);
1005         hci_update_class_sync(hdev);
1006
1007         return 0;
1008 }
1009
1010 static void service_cache_off(struct work_struct *work)
1011 {
1012         struct hci_dev *hdev = container_of(work, struct hci_dev,
1013                                             service_cache.work);
1014
1015         if (!hci_dev_test_and_clear_flag(hdev, HCI_SERVICE_CACHE))
1016                 return;
1017
1018         hci_cmd_sync_queue(hdev, service_cache_sync, NULL, NULL);
1019 }
1020
1021 static int rpa_expired_sync(struct hci_dev *hdev, void *data)
1022 {
1023         /* The generation of a new RPA and programming it into the
1024          * controller happens in the hci_req_enable_advertising()
1025          * function.
1026          */
1027         if (ext_adv_capable(hdev))
1028                 return hci_start_ext_adv_sync(hdev, hdev->cur_adv_instance);
1029         else
1030                 return hci_enable_advertising_sync(hdev);
1031 }
1032
1033 static void rpa_expired(struct work_struct *work)
1034 {
1035         struct hci_dev *hdev = container_of(work, struct hci_dev,
1036                                             rpa_expired.work);
1037
1038         bt_dev_dbg(hdev, "");
1039
1040         hci_dev_set_flag(hdev, HCI_RPA_EXPIRED);
1041
1042         if (!hci_dev_test_flag(hdev, HCI_ADVERTISING))
1043                 return;
1044
1045         hci_cmd_sync_queue(hdev, rpa_expired_sync, NULL, NULL);
1046 }
1047
1048 static void discov_off(struct work_struct *work)
1049 {
1050         struct hci_dev *hdev = container_of(work, struct hci_dev,
1051                                             discov_off.work);
1052
1053         bt_dev_dbg(hdev, "");
1054
1055         hci_dev_lock(hdev);
1056
1057         /* When discoverable timeout triggers, then just make sure
1058          * the limited discoverable flag is cleared. Even in the case
1059          * of a timeout triggered from general discoverable, it is
1060          * safe to unconditionally clear the flag.
1061          */
1062         hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE);
1063         hci_dev_clear_flag(hdev, HCI_DISCOVERABLE);
1064         hdev->discov_timeout = 0;
1065
1066         hci_update_discoverable(hdev);
1067
1068         mgmt_new_settings(hdev);
1069
1070         hci_dev_unlock(hdev);
1071 }
1072
1073 static int send_settings_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev);
1074
1075 static void mesh_send_complete(struct hci_dev *hdev,
1076                                struct mgmt_mesh_tx *mesh_tx, bool silent)
1077 {
1078         u8 handle = mesh_tx->handle;
1079
1080         if (!silent)
1081                 mgmt_event(MGMT_EV_MESH_PACKET_CMPLT, hdev, &handle,
1082                            sizeof(handle), NULL);
1083
1084         mgmt_mesh_remove(mesh_tx);
1085 }
1086
1087 static int mesh_send_done_sync(struct hci_dev *hdev, void *data)
1088 {
1089         struct mgmt_mesh_tx *mesh_tx;
1090
1091         hci_dev_clear_flag(hdev, HCI_MESH_SENDING);
1092         hci_disable_advertising_sync(hdev);
1093         mesh_tx = mgmt_mesh_next(hdev, NULL);
1094
1095         if (mesh_tx)
1096                 mesh_send_complete(hdev, mesh_tx, false);
1097
1098         return 0;
1099 }
1100
1101 static int mesh_send_sync(struct hci_dev *hdev, void *data);
1102 static void mesh_send_start_complete(struct hci_dev *hdev, void *data, int err);
1103 static void mesh_next(struct hci_dev *hdev, void *data, int err)
1104 {
1105         struct mgmt_mesh_tx *mesh_tx = mgmt_mesh_next(hdev, NULL);
1106
1107         if (!mesh_tx)
1108                 return;
1109
1110         err = hci_cmd_sync_queue(hdev, mesh_send_sync, mesh_tx,
1111                                  mesh_send_start_complete);
1112
1113         if (err < 0)
1114                 mesh_send_complete(hdev, mesh_tx, false);
1115         else
1116                 hci_dev_set_flag(hdev, HCI_MESH_SENDING);
1117 }
1118
1119 static void mesh_send_done(struct work_struct *work)
1120 {
1121         struct hci_dev *hdev = container_of(work, struct hci_dev,
1122                                             mesh_send_done.work);
1123
1124         if (!hci_dev_test_flag(hdev, HCI_MESH_SENDING))
1125                 return;
1126
1127         hci_cmd_sync_queue(hdev, mesh_send_done_sync, NULL, mesh_next);
1128 }
1129
1130 static void mgmt_init_hdev(struct sock *sk, struct hci_dev *hdev)
1131 {
1132         if (hci_dev_test_flag(hdev, HCI_MGMT))
1133                 return;
1134
1135         BT_INFO("MGMT ver %d.%d", MGMT_VERSION, MGMT_REVISION);
1136
1137         INIT_DELAYED_WORK(&hdev->discov_off, discov_off);
1138         INIT_DELAYED_WORK(&hdev->service_cache, service_cache_off);
1139         INIT_DELAYED_WORK(&hdev->rpa_expired, rpa_expired);
1140         INIT_DELAYED_WORK(&hdev->mesh_send_done, mesh_send_done);
1141
1142         /* Non-mgmt controlled devices get this bit set
1143          * implicitly so that pairing works for them, however
1144          * for mgmt we require user-space to explicitly enable
1145          * it
1146          */
1147         hci_dev_clear_flag(hdev, HCI_BONDABLE);
1148
1149         hci_dev_set_flag(hdev, HCI_MGMT);
1150 }
1151
1152 static int read_controller_info(struct sock *sk, struct hci_dev *hdev,
1153                                 void *data, u16 data_len)
1154 {
1155         struct mgmt_rp_read_info rp;
1156
1157         bt_dev_dbg(hdev, "sock %p", sk);
1158
1159         hci_dev_lock(hdev);
1160
1161         memset(&rp, 0, sizeof(rp));
1162
1163         bacpy(&rp.bdaddr, &hdev->bdaddr);
1164
1165         rp.version = hdev->hci_ver;
1166         rp.manufacturer = cpu_to_le16(hdev->manufacturer);
1167
1168         rp.supported_settings = cpu_to_le32(get_supported_settings(hdev));
1169         rp.current_settings = cpu_to_le32(get_current_settings(hdev));
1170
1171         memcpy(rp.dev_class, hdev->dev_class, 3);
1172
1173         memcpy(rp.name, hdev->dev_name, sizeof(hdev->dev_name));
1174         memcpy(rp.short_name, hdev->short_name, sizeof(hdev->short_name));
1175
1176         hci_dev_unlock(hdev);
1177
1178         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_INFO, 0, &rp,
1179                                  sizeof(rp));
1180 }
1181
1182 static u16 append_eir_data_to_buf(struct hci_dev *hdev, u8 *eir)
1183 {
1184         u16 eir_len = 0;
1185         size_t name_len;
1186
1187         if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
1188                 eir_len = eir_append_data(eir, eir_len, EIR_CLASS_OF_DEV,
1189                                           hdev->dev_class, 3);
1190
1191         if (hci_dev_test_flag(hdev, HCI_LE_ENABLED))
1192                 eir_len = eir_append_le16(eir, eir_len, EIR_APPEARANCE,
1193                                           hdev->appearance);
1194
1195         name_len = strnlen(hdev->dev_name, sizeof(hdev->dev_name));
1196         eir_len = eir_append_data(eir, eir_len, EIR_NAME_COMPLETE,
1197                                   hdev->dev_name, name_len);
1198
1199         name_len = strnlen(hdev->short_name, sizeof(hdev->short_name));
1200         eir_len = eir_append_data(eir, eir_len, EIR_NAME_SHORT,
1201                                   hdev->short_name, name_len);
1202
1203         return eir_len;
1204 }
1205
1206 static int read_ext_controller_info(struct sock *sk, struct hci_dev *hdev,
1207                                     void *data, u16 data_len)
1208 {
1209         char buf[512];
1210         struct mgmt_rp_read_ext_info *rp = (void *)buf;
1211         u16 eir_len;
1212
1213         bt_dev_dbg(hdev, "sock %p", sk);
1214
1215         memset(&buf, 0, sizeof(buf));
1216
1217         hci_dev_lock(hdev);
1218
1219         bacpy(&rp->bdaddr, &hdev->bdaddr);
1220
1221         rp->version = hdev->hci_ver;
1222         rp->manufacturer = cpu_to_le16(hdev->manufacturer);
1223
1224         rp->supported_settings = cpu_to_le32(get_supported_settings(hdev));
1225         rp->current_settings = cpu_to_le32(get_current_settings(hdev));
1226
1227
1228         eir_len = append_eir_data_to_buf(hdev, rp->eir);
1229         rp->eir_len = cpu_to_le16(eir_len);
1230
1231         hci_dev_unlock(hdev);
1232
1233         /* If this command is called at least once, then the events
1234          * for class of device and local name changes are disabled
1235          * and only the new extended controller information event
1236          * is used.
1237          */
1238         hci_sock_set_flag(sk, HCI_MGMT_EXT_INFO_EVENTS);
1239         hci_sock_clear_flag(sk, HCI_MGMT_DEV_CLASS_EVENTS);
1240         hci_sock_clear_flag(sk, HCI_MGMT_LOCAL_NAME_EVENTS);
1241
1242         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_EXT_INFO, 0, rp,
1243                                  sizeof(*rp) + eir_len);
1244 }
1245
1246 static int ext_info_changed(struct hci_dev *hdev, struct sock *skip)
1247 {
1248         char buf[512];
1249         struct mgmt_ev_ext_info_changed *ev = (void *)buf;
1250         u16 eir_len;
1251
1252         memset(buf, 0, sizeof(buf));
1253
1254         eir_len = append_eir_data_to_buf(hdev, ev->eir);
1255         ev->eir_len = cpu_to_le16(eir_len);
1256
1257         return mgmt_limited_event(MGMT_EV_EXT_INFO_CHANGED, hdev, ev,
1258                                   sizeof(*ev) + eir_len,
1259                                   HCI_MGMT_EXT_INFO_EVENTS, skip);
1260 }
1261
1262 static int send_settings_rsp(struct sock *sk, u16 opcode, struct hci_dev *hdev)
1263 {
1264         __le32 settings = cpu_to_le32(get_current_settings(hdev));
1265
1266         return mgmt_cmd_complete(sk, hdev->id, opcode, 0, &settings,
1267                                  sizeof(settings));
1268 }
1269
1270 void mgmt_advertising_added(struct sock *sk, struct hci_dev *hdev, u8 instance)
1271 {
1272         struct mgmt_ev_advertising_added ev;
1273
1274         ev.instance = instance;
1275
1276         mgmt_event(MGMT_EV_ADVERTISING_ADDED, hdev, &ev, sizeof(ev), sk);
1277 }
1278
1279 void mgmt_advertising_removed(struct sock *sk, struct hci_dev *hdev,
1280                               u8 instance)
1281 {
1282         struct mgmt_ev_advertising_removed ev;
1283
1284         ev.instance = instance;
1285
1286         mgmt_event(MGMT_EV_ADVERTISING_REMOVED, hdev, &ev, sizeof(ev), sk);
1287 }
1288
1289 static void cancel_adv_timeout(struct hci_dev *hdev)
1290 {
1291         if (hdev->adv_instance_timeout) {
1292                 hdev->adv_instance_timeout = 0;
1293                 cancel_delayed_work(&hdev->adv_instance_expire);
1294         }
1295 }
1296
1297 /* This function requires the caller holds hdev->lock */
1298 static void restart_le_actions(struct hci_dev *hdev)
1299 {
1300         struct hci_conn_params *p;
1301
1302         list_for_each_entry(p, &hdev->le_conn_params, list) {
1303                 /* Needed for AUTO_OFF case where might not "really"
1304                  * have been powered off.
1305                  */
1306                 hci_pend_le_list_del_init(p);
1307
1308                 switch (p->auto_connect) {
1309                 case HCI_AUTO_CONN_DIRECT:
1310                 case HCI_AUTO_CONN_ALWAYS:
1311                         hci_pend_le_list_add(p, &hdev->pend_le_conns);
1312                         break;
1313                 case HCI_AUTO_CONN_REPORT:
1314                         hci_pend_le_list_add(p, &hdev->pend_le_reports);
1315                         break;
1316                 default:
1317                         break;
1318                 }
1319         }
1320 }
1321
1322 static int new_settings(struct hci_dev *hdev, struct sock *skip)
1323 {
1324         __le32 ev = cpu_to_le32(get_current_settings(hdev));
1325
1326         return mgmt_limited_event(MGMT_EV_NEW_SETTINGS, hdev, &ev,
1327                                   sizeof(ev), HCI_MGMT_SETTING_EVENTS, skip);
1328 }
1329
1330 static void mgmt_set_powered_complete(struct hci_dev *hdev, void *data, int err)
1331 {
1332         struct mgmt_pending_cmd *cmd = data;
1333         struct mgmt_mode *cp;
1334
1335         /* Make sure cmd still outstanding. */
1336         if (cmd != pending_find(MGMT_OP_SET_POWERED, hdev))
1337                 return;
1338
1339         cp = cmd->param;
1340
1341         bt_dev_dbg(hdev, "err %d", err);
1342
1343         if (!err) {
1344                 if (cp->val) {
1345                         hci_dev_lock(hdev);
1346                         restart_le_actions(hdev);
1347                         hci_update_passive_scan(hdev);
1348                         hci_dev_unlock(hdev);
1349                 }
1350
1351                 send_settings_rsp(cmd->sk, cmd->opcode, hdev);
1352
1353                 /* Only call new_setting for power on as power off is deferred
1354                  * to hdev->power_off work which does call hci_dev_do_close.
1355                  */
1356                 if (cp->val)
1357                         new_settings(hdev, cmd->sk);
1358         } else {
1359                 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_POWERED,
1360                                 mgmt_status(err));
1361         }
1362
1363         mgmt_pending_remove(cmd);
1364 }
1365
1366 static int set_powered_sync(struct hci_dev *hdev, void *data)
1367 {
1368         struct mgmt_pending_cmd *cmd = data;
1369         struct mgmt_mode *cp = cmd->param;
1370
1371         BT_DBG("%s", hdev->name);
1372
1373         return hci_set_powered_sync(hdev, cp->val);
1374 }
1375
1376 static int set_powered(struct sock *sk, struct hci_dev *hdev, void *data,
1377                        u16 len)
1378 {
1379         struct mgmt_mode *cp = data;
1380         struct mgmt_pending_cmd *cmd;
1381         int err;
1382
1383         bt_dev_dbg(hdev, "sock %p", sk);
1384
1385         if (cp->val != 0x00 && cp->val != 0x01)
1386                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_POWERED,
1387                                        MGMT_STATUS_INVALID_PARAMS);
1388
1389         hci_dev_lock(hdev);
1390
1391         if (pending_find(MGMT_OP_SET_POWERED, hdev)) {
1392                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_POWERED,
1393                                       MGMT_STATUS_BUSY);
1394                 goto failed;
1395         }
1396
1397         if (!!cp->val == hdev_is_powered(hdev)) {
1398                 err = send_settings_rsp(sk, MGMT_OP_SET_POWERED, hdev);
1399                 goto failed;
1400         }
1401
1402         cmd = mgmt_pending_add(sk, MGMT_OP_SET_POWERED, hdev, data, len);
1403         if (!cmd) {
1404                 err = -ENOMEM;
1405                 goto failed;
1406         }
1407
1408         /* Cancel potentially blocking sync operation before power off */
1409         if (cp->val == 0x00) {
1410                 __hci_cmd_sync_cancel(hdev, -EHOSTDOWN);
1411                 err = hci_cmd_sync_queue(hdev, set_powered_sync, cmd,
1412                                          mgmt_set_powered_complete);
1413         } else {
1414                 /* Use hci_cmd_sync_submit since hdev might not be running */
1415                 err = hci_cmd_sync_submit(hdev, set_powered_sync, cmd,
1416                                           mgmt_set_powered_complete);
1417         }
1418
1419         if (err < 0)
1420                 mgmt_pending_remove(cmd);
1421
1422 failed:
1423         hci_dev_unlock(hdev);
1424         return err;
1425 }
1426
1427 int mgmt_new_settings(struct hci_dev *hdev)
1428 {
1429         return new_settings(hdev, NULL);
1430 }
1431
1432 struct cmd_lookup {
1433         struct sock *sk;
1434         struct hci_dev *hdev;
1435         u8 mgmt_status;
1436 };
1437
1438 static void settings_rsp(struct mgmt_pending_cmd *cmd, void *data)
1439 {
1440         struct cmd_lookup *match = data;
1441
1442         send_settings_rsp(cmd->sk, cmd->opcode, match->hdev);
1443
1444         list_del(&cmd->list);
1445
1446         if (match->sk == NULL) {
1447                 match->sk = cmd->sk;
1448                 sock_hold(match->sk);
1449         }
1450
1451         mgmt_pending_free(cmd);
1452 }
1453
1454 static void cmd_status_rsp(struct mgmt_pending_cmd *cmd, void *data)
1455 {
1456         u8 *status = data;
1457
1458         mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, *status);
1459         mgmt_pending_remove(cmd);
1460 }
1461
1462 static void cmd_complete_rsp(struct mgmt_pending_cmd *cmd, void *data)
1463 {
1464         if (cmd->cmd_complete) {
1465                 u8 *status = data;
1466
1467                 cmd->cmd_complete(cmd, *status);
1468                 mgmt_pending_remove(cmd);
1469
1470                 return;
1471         }
1472
1473         cmd_status_rsp(cmd, data);
1474 }
1475
1476 static int generic_cmd_complete(struct mgmt_pending_cmd *cmd, u8 status)
1477 {
1478         return mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status,
1479                                  cmd->param, cmd->param_len);
1480 }
1481
1482 static int addr_cmd_complete(struct mgmt_pending_cmd *cmd, u8 status)
1483 {
1484         return mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status,
1485                                  cmd->param, sizeof(struct mgmt_addr_info));
1486 }
1487
1488 static u8 mgmt_bredr_support(struct hci_dev *hdev)
1489 {
1490         if (!lmp_bredr_capable(hdev))
1491                 return MGMT_STATUS_NOT_SUPPORTED;
1492         else if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
1493                 return MGMT_STATUS_REJECTED;
1494         else
1495                 return MGMT_STATUS_SUCCESS;
1496 }
1497
1498 static u8 mgmt_le_support(struct hci_dev *hdev)
1499 {
1500         if (!lmp_le_capable(hdev))
1501                 return MGMT_STATUS_NOT_SUPPORTED;
1502         else if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
1503                 return MGMT_STATUS_REJECTED;
1504         else
1505                 return MGMT_STATUS_SUCCESS;
1506 }
1507
1508 static void mgmt_set_discoverable_complete(struct hci_dev *hdev, void *data,
1509                                            int err)
1510 {
1511         struct mgmt_pending_cmd *cmd = data;
1512
1513         bt_dev_dbg(hdev, "err %d", err);
1514
1515         /* Make sure cmd still outstanding. */
1516         if (cmd != pending_find(MGMT_OP_SET_DISCOVERABLE, hdev))
1517                 return;
1518
1519         hci_dev_lock(hdev);
1520
1521         if (err) {
1522                 u8 mgmt_err = mgmt_status(err);
1523                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err);
1524                 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE);
1525                 goto done;
1526         }
1527
1528         if (hci_dev_test_flag(hdev, HCI_DISCOVERABLE) &&
1529             hdev->discov_timeout > 0) {
1530                 int to = msecs_to_jiffies(hdev->discov_timeout * 1000);
1531                 queue_delayed_work(hdev->req_workqueue, &hdev->discov_off, to);
1532         }
1533
1534         send_settings_rsp(cmd->sk, MGMT_OP_SET_DISCOVERABLE, hdev);
1535         new_settings(hdev, cmd->sk);
1536
1537 done:
1538         mgmt_pending_remove(cmd);
1539         hci_dev_unlock(hdev);
1540 }
1541
1542 static int set_discoverable_sync(struct hci_dev *hdev, void *data)
1543 {
1544         BT_DBG("%s", hdev->name);
1545
1546         return hci_update_discoverable_sync(hdev);
1547 }
1548
1549 static int set_discoverable(struct sock *sk, struct hci_dev *hdev, void *data,
1550                             u16 len)
1551 {
1552         struct mgmt_cp_set_discoverable *cp = data;
1553         struct mgmt_pending_cmd *cmd;
1554         u16 timeout;
1555         int err;
1556
1557         bt_dev_dbg(hdev, "sock %p", sk);
1558
1559         if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) &&
1560             !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
1561                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE,
1562                                        MGMT_STATUS_REJECTED);
1563
1564         if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02)
1565                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE,
1566                                        MGMT_STATUS_INVALID_PARAMS);
1567
1568         timeout = __le16_to_cpu(cp->timeout);
1569
1570         /* Disabling discoverable requires that no timeout is set,
1571          * and enabling limited discoverable requires a timeout.
1572          */
1573         if ((cp->val == 0x00 && timeout > 0) ||
1574             (cp->val == 0x02 && timeout == 0))
1575                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE,
1576                                        MGMT_STATUS_INVALID_PARAMS);
1577
1578         hci_dev_lock(hdev);
1579
1580         if (!hdev_is_powered(hdev) && timeout > 0) {
1581                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE,
1582                                       MGMT_STATUS_NOT_POWERED);
1583                 goto failed;
1584         }
1585
1586         if (pending_find(MGMT_OP_SET_DISCOVERABLE, hdev) ||
1587             pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) {
1588                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE,
1589                                       MGMT_STATUS_BUSY);
1590                 goto failed;
1591         }
1592
1593         if (!hci_dev_test_flag(hdev, HCI_CONNECTABLE)) {
1594                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE,
1595                                       MGMT_STATUS_REJECTED);
1596                 goto failed;
1597         }
1598
1599         if (hdev->advertising_paused) {
1600                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DISCOVERABLE,
1601                                       MGMT_STATUS_BUSY);
1602                 goto failed;
1603         }
1604
1605         if (!hdev_is_powered(hdev)) {
1606                 bool changed = false;
1607
1608                 /* Setting limited discoverable when powered off is
1609                  * not a valid operation since it requires a timeout
1610                  * and so no need to check HCI_LIMITED_DISCOVERABLE.
1611                  */
1612                 if (!!cp->val != hci_dev_test_flag(hdev, HCI_DISCOVERABLE)) {
1613                         hci_dev_change_flag(hdev, HCI_DISCOVERABLE);
1614                         changed = true;
1615                 }
1616
1617                 err = send_settings_rsp(sk, MGMT_OP_SET_DISCOVERABLE, hdev);
1618                 if (err < 0)
1619                         goto failed;
1620
1621                 if (changed)
1622                         err = new_settings(hdev, sk);
1623
1624                 goto failed;
1625         }
1626
1627         /* If the current mode is the same, then just update the timeout
1628          * value with the new value. And if only the timeout gets updated,
1629          * then no need for any HCI transactions.
1630          */
1631         if (!!cp->val == hci_dev_test_flag(hdev, HCI_DISCOVERABLE) &&
1632             (cp->val == 0x02) == hci_dev_test_flag(hdev,
1633                                                    HCI_LIMITED_DISCOVERABLE)) {
1634                 cancel_delayed_work(&hdev->discov_off);
1635                 hdev->discov_timeout = timeout;
1636
1637                 if (cp->val && hdev->discov_timeout > 0) {
1638                         int to = msecs_to_jiffies(hdev->discov_timeout * 1000);
1639                         queue_delayed_work(hdev->req_workqueue,
1640                                            &hdev->discov_off, to);
1641                 }
1642
1643                 err = send_settings_rsp(sk, MGMT_OP_SET_DISCOVERABLE, hdev);
1644                 goto failed;
1645         }
1646
1647         cmd = mgmt_pending_add(sk, MGMT_OP_SET_DISCOVERABLE, hdev, data, len);
1648         if (!cmd) {
1649                 err = -ENOMEM;
1650                 goto failed;
1651         }
1652
1653         /* Cancel any potential discoverable timeout that might be
1654          * still active and store new timeout value. The arming of
1655          * the timeout happens in the complete handler.
1656          */
1657         cancel_delayed_work(&hdev->discov_off);
1658         hdev->discov_timeout = timeout;
1659
1660         if (cp->val)
1661                 hci_dev_set_flag(hdev, HCI_DISCOVERABLE);
1662         else
1663                 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE);
1664
1665         /* Limited discoverable mode */
1666         if (cp->val == 0x02)
1667                 hci_dev_set_flag(hdev, HCI_LIMITED_DISCOVERABLE);
1668         else
1669                 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE);
1670
1671         err = hci_cmd_sync_queue(hdev, set_discoverable_sync, cmd,
1672                                  mgmt_set_discoverable_complete);
1673
1674         if (err < 0)
1675                 mgmt_pending_remove(cmd);
1676
1677 failed:
1678         hci_dev_unlock(hdev);
1679         return err;
1680 }
1681
1682 static void mgmt_set_connectable_complete(struct hci_dev *hdev, void *data,
1683                                           int err)
1684 {
1685         struct mgmt_pending_cmd *cmd = data;
1686
1687         bt_dev_dbg(hdev, "err %d", err);
1688
1689         /* Make sure cmd still outstanding. */
1690         if (cmd != pending_find(MGMT_OP_SET_CONNECTABLE, hdev))
1691                 return;
1692
1693         hci_dev_lock(hdev);
1694
1695         if (err) {
1696                 u8 mgmt_err = mgmt_status(err);
1697                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err);
1698                 goto done;
1699         }
1700
1701         send_settings_rsp(cmd->sk, MGMT_OP_SET_CONNECTABLE, hdev);
1702         new_settings(hdev, cmd->sk);
1703
1704 done:
1705         if (cmd)
1706                 mgmt_pending_remove(cmd);
1707
1708         hci_dev_unlock(hdev);
1709 }
1710
1711 static int set_connectable_update_settings(struct hci_dev *hdev,
1712                                            struct sock *sk, u8 val)
1713 {
1714         bool changed = false;
1715         int err;
1716
1717         if (!!val != hci_dev_test_flag(hdev, HCI_CONNECTABLE))
1718                 changed = true;
1719
1720         if (val) {
1721                 hci_dev_set_flag(hdev, HCI_CONNECTABLE);
1722         } else {
1723                 hci_dev_clear_flag(hdev, HCI_CONNECTABLE);
1724                 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE);
1725         }
1726
1727         err = send_settings_rsp(sk, MGMT_OP_SET_CONNECTABLE, hdev);
1728         if (err < 0)
1729                 return err;
1730
1731         if (changed) {
1732                 hci_update_scan(hdev);
1733                 hci_update_passive_scan(hdev);
1734                 return new_settings(hdev, sk);
1735         }
1736
1737         return 0;
1738 }
1739
1740 static int set_connectable_sync(struct hci_dev *hdev, void *data)
1741 {
1742         BT_DBG("%s", hdev->name);
1743
1744         return hci_update_connectable_sync(hdev);
1745 }
1746
1747 static int set_connectable(struct sock *sk, struct hci_dev *hdev, void *data,
1748                            u16 len)
1749 {
1750         struct mgmt_mode *cp = data;
1751         struct mgmt_pending_cmd *cmd;
1752         int err;
1753
1754         bt_dev_dbg(hdev, "sock %p", sk);
1755
1756         if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) &&
1757             !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
1758                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE,
1759                                        MGMT_STATUS_REJECTED);
1760
1761         if (cp->val != 0x00 && cp->val != 0x01)
1762                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE,
1763                                        MGMT_STATUS_INVALID_PARAMS);
1764
1765         hci_dev_lock(hdev);
1766
1767         if (!hdev_is_powered(hdev)) {
1768                 err = set_connectable_update_settings(hdev, sk, cp->val);
1769                 goto failed;
1770         }
1771
1772         if (pending_find(MGMT_OP_SET_DISCOVERABLE, hdev) ||
1773             pending_find(MGMT_OP_SET_CONNECTABLE, hdev)) {
1774                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_CONNECTABLE,
1775                                       MGMT_STATUS_BUSY);
1776                 goto failed;
1777         }
1778
1779         cmd = mgmt_pending_add(sk, MGMT_OP_SET_CONNECTABLE, hdev, data, len);
1780         if (!cmd) {
1781                 err = -ENOMEM;
1782                 goto failed;
1783         }
1784
1785         if (cp->val) {
1786                 hci_dev_set_flag(hdev, HCI_CONNECTABLE);
1787         } else {
1788                 if (hdev->discov_timeout > 0)
1789                         cancel_delayed_work(&hdev->discov_off);
1790
1791                 hci_dev_clear_flag(hdev, HCI_LIMITED_DISCOVERABLE);
1792                 hci_dev_clear_flag(hdev, HCI_DISCOVERABLE);
1793                 hci_dev_clear_flag(hdev, HCI_CONNECTABLE);
1794         }
1795
1796         err = hci_cmd_sync_queue(hdev, set_connectable_sync, cmd,
1797                                  mgmt_set_connectable_complete);
1798
1799         if (err < 0)
1800                 mgmt_pending_remove(cmd);
1801
1802 failed:
1803         hci_dev_unlock(hdev);
1804         return err;
1805 }
1806
1807 static int set_bondable(struct sock *sk, struct hci_dev *hdev, void *data,
1808                         u16 len)
1809 {
1810         struct mgmt_mode *cp = data;
1811         bool changed;
1812         int err;
1813
1814         bt_dev_dbg(hdev, "sock %p", sk);
1815
1816         if (cp->val != 0x00 && cp->val != 0x01)
1817                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BONDABLE,
1818                                        MGMT_STATUS_INVALID_PARAMS);
1819
1820         hci_dev_lock(hdev);
1821
1822         if (cp->val)
1823                 changed = !hci_dev_test_and_set_flag(hdev, HCI_BONDABLE);
1824         else
1825                 changed = hci_dev_test_and_clear_flag(hdev, HCI_BONDABLE);
1826
1827         err = send_settings_rsp(sk, MGMT_OP_SET_BONDABLE, hdev);
1828         if (err < 0)
1829                 goto unlock;
1830
1831         if (changed) {
1832                 /* In limited privacy mode the change of bondable mode
1833                  * may affect the local advertising address.
1834                  */
1835                 hci_update_discoverable(hdev);
1836
1837                 err = new_settings(hdev, sk);
1838         }
1839
1840 unlock:
1841         hci_dev_unlock(hdev);
1842         return err;
1843 }
1844
1845 static int set_link_security(struct sock *sk, struct hci_dev *hdev, void *data,
1846                              u16 len)
1847 {
1848         struct mgmt_mode *cp = data;
1849         struct mgmt_pending_cmd *cmd;
1850         u8 val, status;
1851         int err;
1852
1853         bt_dev_dbg(hdev, "sock %p", sk);
1854
1855         status = mgmt_bredr_support(hdev);
1856         if (status)
1857                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY,
1858                                        status);
1859
1860         if (cp->val != 0x00 && cp->val != 0x01)
1861                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY,
1862                                        MGMT_STATUS_INVALID_PARAMS);
1863
1864         hci_dev_lock(hdev);
1865
1866         if (!hdev_is_powered(hdev)) {
1867                 bool changed = false;
1868
1869                 if (!!cp->val != hci_dev_test_flag(hdev, HCI_LINK_SECURITY)) {
1870                         hci_dev_change_flag(hdev, HCI_LINK_SECURITY);
1871                         changed = true;
1872                 }
1873
1874                 err = send_settings_rsp(sk, MGMT_OP_SET_LINK_SECURITY, hdev);
1875                 if (err < 0)
1876                         goto failed;
1877
1878                 if (changed)
1879                         err = new_settings(hdev, sk);
1880
1881                 goto failed;
1882         }
1883
1884         if (pending_find(MGMT_OP_SET_LINK_SECURITY, hdev)) {
1885                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LINK_SECURITY,
1886                                       MGMT_STATUS_BUSY);
1887                 goto failed;
1888         }
1889
1890         val = !!cp->val;
1891
1892         if (test_bit(HCI_AUTH, &hdev->flags) == val) {
1893                 err = send_settings_rsp(sk, MGMT_OP_SET_LINK_SECURITY, hdev);
1894                 goto failed;
1895         }
1896
1897         cmd = mgmt_pending_add(sk, MGMT_OP_SET_LINK_SECURITY, hdev, data, len);
1898         if (!cmd) {
1899                 err = -ENOMEM;
1900                 goto failed;
1901         }
1902
1903         err = hci_send_cmd(hdev, HCI_OP_WRITE_AUTH_ENABLE, sizeof(val), &val);
1904         if (err < 0) {
1905                 mgmt_pending_remove(cmd);
1906                 goto failed;
1907         }
1908
1909 failed:
1910         hci_dev_unlock(hdev);
1911         return err;
1912 }
1913
1914 static void set_ssp_complete(struct hci_dev *hdev, void *data, int err)
1915 {
1916         struct cmd_lookup match = { NULL, hdev };
1917         struct mgmt_pending_cmd *cmd = data;
1918         struct mgmt_mode *cp = cmd->param;
1919         u8 enable = cp->val;
1920         bool changed;
1921
1922         /* Make sure cmd still outstanding. */
1923         if (cmd != pending_find(MGMT_OP_SET_SSP, hdev))
1924                 return;
1925
1926         if (err) {
1927                 u8 mgmt_err = mgmt_status(err);
1928
1929                 if (enable && hci_dev_test_and_clear_flag(hdev,
1930                                                           HCI_SSP_ENABLED)) {
1931                         hci_dev_clear_flag(hdev, HCI_HS_ENABLED);
1932                         new_settings(hdev, NULL);
1933                 }
1934
1935                 mgmt_pending_foreach(MGMT_OP_SET_SSP, hdev, cmd_status_rsp,
1936                                      &mgmt_err);
1937                 return;
1938         }
1939
1940         if (enable) {
1941                 changed = !hci_dev_test_and_set_flag(hdev, HCI_SSP_ENABLED);
1942         } else {
1943                 changed = hci_dev_test_and_clear_flag(hdev, HCI_SSP_ENABLED);
1944
1945                 if (!changed)
1946                         changed = hci_dev_test_and_clear_flag(hdev,
1947                                                               HCI_HS_ENABLED);
1948                 else
1949                         hci_dev_clear_flag(hdev, HCI_HS_ENABLED);
1950         }
1951
1952         mgmt_pending_foreach(MGMT_OP_SET_SSP, hdev, settings_rsp, &match);
1953
1954         if (changed)
1955                 new_settings(hdev, match.sk);
1956
1957         if (match.sk)
1958                 sock_put(match.sk);
1959
1960         hci_update_eir_sync(hdev);
1961 }
1962
1963 static int set_ssp_sync(struct hci_dev *hdev, void *data)
1964 {
1965         struct mgmt_pending_cmd *cmd = data;
1966         struct mgmt_mode *cp = cmd->param;
1967         bool changed = false;
1968         int err;
1969
1970         if (cp->val)
1971                 changed = !hci_dev_test_and_set_flag(hdev, HCI_SSP_ENABLED);
1972
1973         err = hci_write_ssp_mode_sync(hdev, cp->val);
1974
1975         if (!err && changed)
1976                 hci_dev_clear_flag(hdev, HCI_SSP_ENABLED);
1977
1978         return err;
1979 }
1980
1981 static int set_ssp(struct sock *sk, struct hci_dev *hdev, void *data, u16 len)
1982 {
1983         struct mgmt_mode *cp = data;
1984         struct mgmt_pending_cmd *cmd;
1985         u8 status;
1986         int err;
1987
1988         bt_dev_dbg(hdev, "sock %p", sk);
1989
1990         status = mgmt_bredr_support(hdev);
1991         if (status)
1992                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP, status);
1993
1994         if (!lmp_ssp_capable(hdev))
1995                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP,
1996                                        MGMT_STATUS_NOT_SUPPORTED);
1997
1998         if (cp->val != 0x00 && cp->val != 0x01)
1999                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP,
2000                                        MGMT_STATUS_INVALID_PARAMS);
2001
2002         hci_dev_lock(hdev);
2003
2004         if (!hdev_is_powered(hdev)) {
2005                 bool changed;
2006
2007                 if (cp->val) {
2008                         changed = !hci_dev_test_and_set_flag(hdev,
2009                                                              HCI_SSP_ENABLED);
2010                 } else {
2011                         changed = hci_dev_test_and_clear_flag(hdev,
2012                                                               HCI_SSP_ENABLED);
2013                         if (!changed)
2014                                 changed = hci_dev_test_and_clear_flag(hdev,
2015                                                                       HCI_HS_ENABLED);
2016                         else
2017                                 hci_dev_clear_flag(hdev, HCI_HS_ENABLED);
2018                 }
2019
2020                 err = send_settings_rsp(sk, MGMT_OP_SET_SSP, hdev);
2021                 if (err < 0)
2022                         goto failed;
2023
2024                 if (changed)
2025                         err = new_settings(hdev, sk);
2026
2027                 goto failed;
2028         }
2029
2030         if (pending_find(MGMT_OP_SET_SSP, hdev)) {
2031                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP,
2032                                       MGMT_STATUS_BUSY);
2033                 goto failed;
2034         }
2035
2036         if (!!cp->val == hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) {
2037                 err = send_settings_rsp(sk, MGMT_OP_SET_SSP, hdev);
2038                 goto failed;
2039         }
2040
2041         cmd = mgmt_pending_add(sk, MGMT_OP_SET_SSP, hdev, data, len);
2042         if (!cmd)
2043                 err = -ENOMEM;
2044         else
2045                 err = hci_cmd_sync_queue(hdev, set_ssp_sync, cmd,
2046                                          set_ssp_complete);
2047
2048         if (err < 0) {
2049                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SSP,
2050                                       MGMT_STATUS_FAILED);
2051
2052                 if (cmd)
2053                         mgmt_pending_remove(cmd);
2054         }
2055
2056 failed:
2057         hci_dev_unlock(hdev);
2058         return err;
2059 }
2060
2061 static int set_hs(struct sock *sk, struct hci_dev *hdev, void *data, u16 len)
2062 {
2063         struct mgmt_mode *cp = data;
2064         bool changed;
2065         u8 status;
2066         int err;
2067
2068         bt_dev_dbg(hdev, "sock %p", sk);
2069
2070         if (!IS_ENABLED(CONFIG_BT_HS))
2071                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS,
2072                                        MGMT_STATUS_NOT_SUPPORTED);
2073
2074         status = mgmt_bredr_support(hdev);
2075         if (status)
2076                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS, status);
2077
2078         if (!lmp_ssp_capable(hdev))
2079                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS,
2080                                        MGMT_STATUS_NOT_SUPPORTED);
2081
2082         if (!hci_dev_test_flag(hdev, HCI_SSP_ENABLED))
2083                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS,
2084                                        MGMT_STATUS_REJECTED);
2085
2086         if (cp->val != 0x00 && cp->val != 0x01)
2087                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS,
2088                                        MGMT_STATUS_INVALID_PARAMS);
2089
2090         hci_dev_lock(hdev);
2091
2092         if (pending_find(MGMT_OP_SET_SSP, hdev)) {
2093                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS,
2094                                       MGMT_STATUS_BUSY);
2095                 goto unlock;
2096         }
2097
2098         if (cp->val) {
2099                 changed = !hci_dev_test_and_set_flag(hdev, HCI_HS_ENABLED);
2100         } else {
2101                 if (hdev_is_powered(hdev)) {
2102                         err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_HS,
2103                                               MGMT_STATUS_REJECTED);
2104                         goto unlock;
2105                 }
2106
2107                 changed = hci_dev_test_and_clear_flag(hdev, HCI_HS_ENABLED);
2108         }
2109
2110         err = send_settings_rsp(sk, MGMT_OP_SET_HS, hdev);
2111         if (err < 0)
2112                 goto unlock;
2113
2114         if (changed)
2115                 err = new_settings(hdev, sk);
2116
2117 unlock:
2118         hci_dev_unlock(hdev);
2119         return err;
2120 }
2121
2122 static void set_le_complete(struct hci_dev *hdev, void *data, int err)
2123 {
2124         struct cmd_lookup match = { NULL, hdev };
2125         u8 status = mgmt_status(err);
2126
2127         bt_dev_dbg(hdev, "err %d", err);
2128
2129         if (status) {
2130                 mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, cmd_status_rsp,
2131                                                         &status);
2132                 return;
2133         }
2134
2135         mgmt_pending_foreach(MGMT_OP_SET_LE, hdev, settings_rsp, &match);
2136
2137         new_settings(hdev, match.sk);
2138
2139         if (match.sk)
2140                 sock_put(match.sk);
2141 }
2142
2143 static int set_le_sync(struct hci_dev *hdev, void *data)
2144 {
2145         struct mgmt_pending_cmd *cmd = data;
2146         struct mgmt_mode *cp = cmd->param;
2147         u8 val = !!cp->val;
2148         int err;
2149
2150         if (!val) {
2151                 hci_clear_adv_instance_sync(hdev, NULL, 0x00, true);
2152
2153                 if (hci_dev_test_flag(hdev, HCI_LE_ADV))
2154                         hci_disable_advertising_sync(hdev);
2155
2156                 if (ext_adv_capable(hdev))
2157                         hci_remove_ext_adv_instance_sync(hdev, 0, cmd->sk);
2158         } else {
2159                 hci_dev_set_flag(hdev, HCI_LE_ENABLED);
2160         }
2161
2162         err = hci_write_le_host_supported_sync(hdev, val, 0);
2163
2164         /* Make sure the controller has a good default for
2165          * advertising data. Restrict the update to when LE
2166          * has actually been enabled. During power on, the
2167          * update in powered_update_hci will take care of it.
2168          */
2169         if (!err && hci_dev_test_flag(hdev, HCI_LE_ENABLED)) {
2170                 if (ext_adv_capable(hdev)) {
2171                         int status;
2172
2173                         status = hci_setup_ext_adv_instance_sync(hdev, 0x00);
2174                         if (!status)
2175                                 hci_update_scan_rsp_data_sync(hdev, 0x00);
2176                 } else {
2177                         hci_update_adv_data_sync(hdev, 0x00);
2178                         hci_update_scan_rsp_data_sync(hdev, 0x00);
2179                 }
2180
2181                 hci_update_passive_scan(hdev);
2182         }
2183
2184         return err;
2185 }
2186
2187 static void set_mesh_complete(struct hci_dev *hdev, void *data, int err)
2188 {
2189         struct mgmt_pending_cmd *cmd = data;
2190         u8 status = mgmt_status(err);
2191         struct sock *sk = cmd->sk;
2192
2193         if (status) {
2194                 mgmt_pending_foreach(MGMT_OP_SET_MESH_RECEIVER, hdev,
2195                                      cmd_status_rsp, &status);
2196                 return;
2197         }
2198
2199         mgmt_pending_remove(cmd);
2200         mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER, 0, NULL, 0);
2201 }
2202
2203 static int set_mesh_sync(struct hci_dev *hdev, void *data)
2204 {
2205         struct mgmt_pending_cmd *cmd = data;
2206         struct mgmt_cp_set_mesh *cp = cmd->param;
2207         size_t len = cmd->param_len;
2208
2209         memset(hdev->mesh_ad_types, 0, sizeof(hdev->mesh_ad_types));
2210
2211         if (cp->enable)
2212                 hci_dev_set_flag(hdev, HCI_MESH);
2213         else
2214                 hci_dev_clear_flag(hdev, HCI_MESH);
2215
2216         len -= sizeof(*cp);
2217
2218         /* If filters don't fit, forward all adv pkts */
2219         if (len <= sizeof(hdev->mesh_ad_types))
2220                 memcpy(hdev->mesh_ad_types, cp->ad_types, len);
2221
2222         hci_update_passive_scan_sync(hdev);
2223         return 0;
2224 }
2225
2226 static int set_mesh(struct sock *sk, struct hci_dev *hdev, void *data, u16 len)
2227 {
2228         struct mgmt_cp_set_mesh *cp = data;
2229         struct mgmt_pending_cmd *cmd;
2230         int err = 0;
2231
2232         bt_dev_dbg(hdev, "sock %p", sk);
2233
2234         if (!lmp_le_capable(hdev) ||
2235             !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL))
2236                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER,
2237                                        MGMT_STATUS_NOT_SUPPORTED);
2238
2239         if (cp->enable != 0x00 && cp->enable != 0x01)
2240                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER,
2241                                        MGMT_STATUS_INVALID_PARAMS);
2242
2243         hci_dev_lock(hdev);
2244
2245         cmd = mgmt_pending_add(sk, MGMT_OP_SET_MESH_RECEIVER, hdev, data, len);
2246         if (!cmd)
2247                 err = -ENOMEM;
2248         else
2249                 err = hci_cmd_sync_queue(hdev, set_mesh_sync, cmd,
2250                                          set_mesh_complete);
2251
2252         if (err < 0) {
2253                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_MESH_RECEIVER,
2254                                       MGMT_STATUS_FAILED);
2255
2256                 if (cmd)
2257                         mgmt_pending_remove(cmd);
2258         }
2259
2260         hci_dev_unlock(hdev);
2261         return err;
2262 }
2263
2264 static void mesh_send_start_complete(struct hci_dev *hdev, void *data, int err)
2265 {
2266         struct mgmt_mesh_tx *mesh_tx = data;
2267         struct mgmt_cp_mesh_send *send = (void *)mesh_tx->param;
2268         unsigned long mesh_send_interval;
2269         u8 mgmt_err = mgmt_status(err);
2270
2271         /* Report any errors here, but don't report completion */
2272
2273         if (mgmt_err) {
2274                 hci_dev_clear_flag(hdev, HCI_MESH_SENDING);
2275                 /* Send Complete Error Code for handle */
2276                 mesh_send_complete(hdev, mesh_tx, false);
2277                 return;
2278         }
2279
2280         mesh_send_interval = msecs_to_jiffies((send->cnt) * 25);
2281         queue_delayed_work(hdev->req_workqueue, &hdev->mesh_send_done,
2282                            mesh_send_interval);
2283 }
2284
2285 static int mesh_send_sync(struct hci_dev *hdev, void *data)
2286 {
2287         struct mgmt_mesh_tx *mesh_tx = data;
2288         struct mgmt_cp_mesh_send *send = (void *)mesh_tx->param;
2289         struct adv_info *adv, *next_instance;
2290         u8 instance = hdev->le_num_of_adv_sets + 1;
2291         u16 timeout, duration;
2292         int err = 0;
2293
2294         if (hdev->le_num_of_adv_sets <= hdev->adv_instance_cnt)
2295                 return MGMT_STATUS_BUSY;
2296
2297         timeout = 1000;
2298         duration = send->cnt * INTERVAL_TO_MS(hdev->le_adv_max_interval);
2299         adv = hci_add_adv_instance(hdev, instance, 0,
2300                                    send->adv_data_len, send->adv_data,
2301                                    0, NULL,
2302                                    timeout, duration,
2303                                    HCI_ADV_TX_POWER_NO_PREFERENCE,
2304                                    hdev->le_adv_min_interval,
2305                                    hdev->le_adv_max_interval,
2306                                    mesh_tx->handle);
2307
2308         if (!IS_ERR(adv))
2309                 mesh_tx->instance = instance;
2310         else
2311                 err = PTR_ERR(adv);
2312
2313         if (hdev->cur_adv_instance == instance) {
2314                 /* If the currently advertised instance is being changed then
2315                  * cancel the current advertising and schedule the next
2316                  * instance. If there is only one instance then the overridden
2317                  * advertising data will be visible right away.
2318                  */
2319                 cancel_adv_timeout(hdev);
2320
2321                 next_instance = hci_get_next_instance(hdev, instance);
2322                 if (next_instance)
2323                         instance = next_instance->instance;
2324                 else
2325                         instance = 0;
2326         } else if (hdev->adv_instance_timeout) {
2327                 /* Immediately advertise the new instance if no other, or
2328                  * let it go naturally from queue if ADV is already happening
2329                  */
2330                 instance = 0;
2331         }
2332
2333         if (instance)
2334                 return hci_schedule_adv_instance_sync(hdev, instance, true);
2335
2336         return err;
2337 }
2338
2339 static void send_count(struct mgmt_mesh_tx *mesh_tx, void *data)
2340 {
2341         struct mgmt_rp_mesh_read_features *rp = data;
2342
2343         if (rp->used_handles >= rp->max_handles)
2344                 return;
2345
2346         rp->handles[rp->used_handles++] = mesh_tx->handle;
2347 }
2348
2349 static int mesh_features(struct sock *sk, struct hci_dev *hdev,
2350                          void *data, u16 len)
2351 {
2352         struct mgmt_rp_mesh_read_features rp;
2353
2354         if (!lmp_le_capable(hdev) ||
2355             !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL))
2356                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_READ_FEATURES,
2357                                        MGMT_STATUS_NOT_SUPPORTED);
2358
2359         memset(&rp, 0, sizeof(rp));
2360         rp.index = cpu_to_le16(hdev->id);
2361         if (hci_dev_test_flag(hdev, HCI_LE_ENABLED))
2362                 rp.max_handles = MESH_HANDLES_MAX;
2363
2364         hci_dev_lock(hdev);
2365
2366         if (rp.max_handles)
2367                 mgmt_mesh_foreach(hdev, send_count, &rp, sk);
2368
2369         mgmt_cmd_complete(sk, hdev->id, MGMT_OP_MESH_READ_FEATURES, 0, &rp,
2370                           rp.used_handles + sizeof(rp) - MESH_HANDLES_MAX);
2371
2372         hci_dev_unlock(hdev);
2373         return 0;
2374 }
2375
2376 static int send_cancel(struct hci_dev *hdev, void *data)
2377 {
2378         struct mgmt_pending_cmd *cmd = data;
2379         struct mgmt_cp_mesh_send_cancel *cancel = (void *)cmd->param;
2380         struct mgmt_mesh_tx *mesh_tx;
2381
2382         if (!cancel->handle) {
2383                 do {
2384                         mesh_tx = mgmt_mesh_next(hdev, cmd->sk);
2385
2386                         if (mesh_tx)
2387                                 mesh_send_complete(hdev, mesh_tx, false);
2388                 } while (mesh_tx);
2389         } else {
2390                 mesh_tx = mgmt_mesh_find(hdev, cancel->handle);
2391
2392                 if (mesh_tx && mesh_tx->sk == cmd->sk)
2393                         mesh_send_complete(hdev, mesh_tx, false);
2394         }
2395
2396         mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL,
2397                           0, NULL, 0);
2398         mgmt_pending_free(cmd);
2399
2400         return 0;
2401 }
2402
2403 static int mesh_send_cancel(struct sock *sk, struct hci_dev *hdev,
2404                             void *data, u16 len)
2405 {
2406         struct mgmt_pending_cmd *cmd;
2407         int err;
2408
2409         if (!lmp_le_capable(hdev) ||
2410             !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL))
2411                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL,
2412                                        MGMT_STATUS_NOT_SUPPORTED);
2413
2414         if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
2415                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL,
2416                                        MGMT_STATUS_REJECTED);
2417
2418         hci_dev_lock(hdev);
2419         cmd = mgmt_pending_new(sk, MGMT_OP_MESH_SEND_CANCEL, hdev, data, len);
2420         if (!cmd)
2421                 err = -ENOMEM;
2422         else
2423                 err = hci_cmd_sync_queue(hdev, send_cancel, cmd, NULL);
2424
2425         if (err < 0) {
2426                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND_CANCEL,
2427                                       MGMT_STATUS_FAILED);
2428
2429                 if (cmd)
2430                         mgmt_pending_free(cmd);
2431         }
2432
2433         hci_dev_unlock(hdev);
2434         return err;
2435 }
2436
2437 static int mesh_send(struct sock *sk, struct hci_dev *hdev, void *data, u16 len)
2438 {
2439         struct mgmt_mesh_tx *mesh_tx;
2440         struct mgmt_cp_mesh_send *send = data;
2441         struct mgmt_rp_mesh_read_features rp;
2442         bool sending;
2443         int err = 0;
2444
2445         if (!lmp_le_capable(hdev) ||
2446             !hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL))
2447                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND,
2448                                        MGMT_STATUS_NOT_SUPPORTED);
2449         if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED) ||
2450             len <= MGMT_MESH_SEND_SIZE ||
2451             len > (MGMT_MESH_SEND_SIZE + 31))
2452                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND,
2453                                        MGMT_STATUS_REJECTED);
2454
2455         hci_dev_lock(hdev);
2456
2457         memset(&rp, 0, sizeof(rp));
2458         rp.max_handles = MESH_HANDLES_MAX;
2459
2460         mgmt_mesh_foreach(hdev, send_count, &rp, sk);
2461
2462         if (rp.max_handles <= rp.used_handles) {
2463                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND,
2464                                       MGMT_STATUS_BUSY);
2465                 goto done;
2466         }
2467
2468         sending = hci_dev_test_flag(hdev, HCI_MESH_SENDING);
2469         mesh_tx = mgmt_mesh_add(sk, hdev, send, len);
2470
2471         if (!mesh_tx)
2472                 err = -ENOMEM;
2473         else if (!sending)
2474                 err = hci_cmd_sync_queue(hdev, mesh_send_sync, mesh_tx,
2475                                          mesh_send_start_complete);
2476
2477         if (err < 0) {
2478                 bt_dev_err(hdev, "Send Mesh Failed %d", err);
2479                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_MESH_SEND,
2480                                       MGMT_STATUS_FAILED);
2481
2482                 if (mesh_tx) {
2483                         if (sending)
2484                                 mgmt_mesh_remove(mesh_tx);
2485                 }
2486         } else {
2487                 hci_dev_set_flag(hdev, HCI_MESH_SENDING);
2488
2489                 mgmt_cmd_complete(sk, hdev->id, MGMT_OP_MESH_SEND, 0,
2490                                   &mesh_tx->handle, 1);
2491         }
2492
2493 done:
2494         hci_dev_unlock(hdev);
2495         return err;
2496 }
2497
2498 static int set_le(struct sock *sk, struct hci_dev *hdev, void *data, u16 len)
2499 {
2500         struct mgmt_mode *cp = data;
2501         struct mgmt_pending_cmd *cmd;
2502         int err;
2503         u8 val, enabled;
2504
2505         bt_dev_dbg(hdev, "sock %p", sk);
2506
2507         if (!lmp_le_capable(hdev))
2508                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE,
2509                                        MGMT_STATUS_NOT_SUPPORTED);
2510
2511         if (cp->val != 0x00 && cp->val != 0x01)
2512                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE,
2513                                        MGMT_STATUS_INVALID_PARAMS);
2514
2515         /* Bluetooth single mode LE only controllers or dual-mode
2516          * controllers configured as LE only devices, do not allow
2517          * switching LE off. These have either LE enabled explicitly
2518          * or BR/EDR has been previously switched off.
2519          *
2520          * When trying to enable an already enabled LE, then gracefully
2521          * send a positive response. Trying to disable it however will
2522          * result into rejection.
2523          */
2524         if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) {
2525                 if (cp->val == 0x01)
2526                         return send_settings_rsp(sk, MGMT_OP_SET_LE, hdev);
2527
2528                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE,
2529                                        MGMT_STATUS_REJECTED);
2530         }
2531
2532         hci_dev_lock(hdev);
2533
2534         val = !!cp->val;
2535         enabled = lmp_host_le_capable(hdev);
2536
2537         if (!hdev_is_powered(hdev) || val == enabled) {
2538                 bool changed = false;
2539
2540                 if (val != hci_dev_test_flag(hdev, HCI_LE_ENABLED)) {
2541                         hci_dev_change_flag(hdev, HCI_LE_ENABLED);
2542                         changed = true;
2543                 }
2544
2545                 if (!val && hci_dev_test_flag(hdev, HCI_ADVERTISING)) {
2546                         hci_dev_clear_flag(hdev, HCI_ADVERTISING);
2547                         changed = true;
2548                 }
2549
2550                 err = send_settings_rsp(sk, MGMT_OP_SET_LE, hdev);
2551                 if (err < 0)
2552                         goto unlock;
2553
2554                 if (changed)
2555                         err = new_settings(hdev, sk);
2556
2557                 goto unlock;
2558         }
2559
2560         if (pending_find(MGMT_OP_SET_LE, hdev) ||
2561             pending_find(MGMT_OP_SET_ADVERTISING, hdev)) {
2562                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE,
2563                                       MGMT_STATUS_BUSY);
2564                 goto unlock;
2565         }
2566
2567         cmd = mgmt_pending_add(sk, MGMT_OP_SET_LE, hdev, data, len);
2568         if (!cmd)
2569                 err = -ENOMEM;
2570         else
2571                 err = hci_cmd_sync_queue(hdev, set_le_sync, cmd,
2572                                          set_le_complete);
2573
2574         if (err < 0) {
2575                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LE,
2576                                       MGMT_STATUS_FAILED);
2577
2578                 if (cmd)
2579                         mgmt_pending_remove(cmd);
2580         }
2581
2582 unlock:
2583         hci_dev_unlock(hdev);
2584         return err;
2585 }
2586
2587 /* This is a helper function to test for pending mgmt commands that can
2588  * cause CoD or EIR HCI commands. We can only allow one such pending
2589  * mgmt command at a time since otherwise we cannot easily track what
2590  * the current values are, will be, and based on that calculate if a new
2591  * HCI command needs to be sent and if yes with what value.
2592  */
2593 static bool pending_eir_or_class(struct hci_dev *hdev)
2594 {
2595         struct mgmt_pending_cmd *cmd;
2596
2597         list_for_each_entry(cmd, &hdev->mgmt_pending, list) {
2598                 switch (cmd->opcode) {
2599                 case MGMT_OP_ADD_UUID:
2600                 case MGMT_OP_REMOVE_UUID:
2601                 case MGMT_OP_SET_DEV_CLASS:
2602                 case MGMT_OP_SET_POWERED:
2603                         return true;
2604                 }
2605         }
2606
2607         return false;
2608 }
2609
2610 static const u8 bluetooth_base_uuid[] = {
2611                         0xfb, 0x34, 0x9b, 0x5f, 0x80, 0x00, 0x00, 0x80,
2612                         0x00, 0x10, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
2613 };
2614
2615 static u8 get_uuid_size(const u8 *uuid)
2616 {
2617         u32 val;
2618
2619         if (memcmp(uuid, bluetooth_base_uuid, 12))
2620                 return 128;
2621
2622         val = get_unaligned_le32(&uuid[12]);
2623         if (val > 0xffff)
2624                 return 32;
2625
2626         return 16;
2627 }
2628
2629 static void mgmt_class_complete(struct hci_dev *hdev, void *data, int err)
2630 {
2631         struct mgmt_pending_cmd *cmd = data;
2632
2633         bt_dev_dbg(hdev, "err %d", err);
2634
2635         mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode,
2636                           mgmt_status(err), hdev->dev_class, 3);
2637
2638         mgmt_pending_free(cmd);
2639 }
2640
2641 static int add_uuid_sync(struct hci_dev *hdev, void *data)
2642 {
2643         int err;
2644
2645         err = hci_update_class_sync(hdev);
2646         if (err)
2647                 return err;
2648
2649         return hci_update_eir_sync(hdev);
2650 }
2651
2652 static int add_uuid(struct sock *sk, struct hci_dev *hdev, void *data, u16 len)
2653 {
2654         struct mgmt_cp_add_uuid *cp = data;
2655         struct mgmt_pending_cmd *cmd;
2656         struct bt_uuid *uuid;
2657         int err;
2658
2659         bt_dev_dbg(hdev, "sock %p", sk);
2660
2661         hci_dev_lock(hdev);
2662
2663         if (pending_eir_or_class(hdev)) {
2664                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_UUID,
2665                                       MGMT_STATUS_BUSY);
2666                 goto failed;
2667         }
2668
2669         uuid = kmalloc(sizeof(*uuid), GFP_KERNEL);
2670         if (!uuid) {
2671                 err = -ENOMEM;
2672                 goto failed;
2673         }
2674
2675         memcpy(uuid->uuid, cp->uuid, 16);
2676         uuid->svc_hint = cp->svc_hint;
2677         uuid->size = get_uuid_size(cp->uuid);
2678
2679         list_add_tail(&uuid->list, &hdev->uuids);
2680
2681         cmd = mgmt_pending_new(sk, MGMT_OP_ADD_UUID, hdev, data, len);
2682         if (!cmd) {
2683                 err = -ENOMEM;
2684                 goto failed;
2685         }
2686
2687         err = hci_cmd_sync_queue(hdev, add_uuid_sync, cmd, mgmt_class_complete);
2688         if (err < 0) {
2689                 mgmt_pending_free(cmd);
2690                 goto failed;
2691         }
2692
2693 failed:
2694         hci_dev_unlock(hdev);
2695         return err;
2696 }
2697
2698 static bool enable_service_cache(struct hci_dev *hdev)
2699 {
2700         if (!hdev_is_powered(hdev))
2701                 return false;
2702
2703         if (!hci_dev_test_and_set_flag(hdev, HCI_SERVICE_CACHE)) {
2704                 queue_delayed_work(hdev->workqueue, &hdev->service_cache,
2705                                    CACHE_TIMEOUT);
2706                 return true;
2707         }
2708
2709         return false;
2710 }
2711
2712 static int remove_uuid_sync(struct hci_dev *hdev, void *data)
2713 {
2714         int err;
2715
2716         err = hci_update_class_sync(hdev);
2717         if (err)
2718                 return err;
2719
2720         return hci_update_eir_sync(hdev);
2721 }
2722
2723 static int remove_uuid(struct sock *sk, struct hci_dev *hdev, void *data,
2724                        u16 len)
2725 {
2726         struct mgmt_cp_remove_uuid *cp = data;
2727         struct mgmt_pending_cmd *cmd;
2728         struct bt_uuid *match, *tmp;
2729         static const u8 bt_uuid_any[] = {
2730                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0
2731         };
2732         int err, found;
2733
2734         bt_dev_dbg(hdev, "sock %p", sk);
2735
2736         hci_dev_lock(hdev);
2737
2738         if (pending_eir_or_class(hdev)) {
2739                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_UUID,
2740                                       MGMT_STATUS_BUSY);
2741                 goto unlock;
2742         }
2743
2744         if (memcmp(cp->uuid, bt_uuid_any, 16) == 0) {
2745                 hci_uuids_clear(hdev);
2746
2747                 if (enable_service_cache(hdev)) {
2748                         err = mgmt_cmd_complete(sk, hdev->id,
2749                                                 MGMT_OP_REMOVE_UUID,
2750                                                 0, hdev->dev_class, 3);
2751                         goto unlock;
2752                 }
2753
2754                 goto update_class;
2755         }
2756
2757         found = 0;
2758
2759         list_for_each_entry_safe(match, tmp, &hdev->uuids, list) {
2760                 if (memcmp(match->uuid, cp->uuid, 16) != 0)
2761                         continue;
2762
2763                 list_del(&match->list);
2764                 kfree(match);
2765                 found++;
2766         }
2767
2768         if (found == 0) {
2769                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_UUID,
2770                                       MGMT_STATUS_INVALID_PARAMS);
2771                 goto unlock;
2772         }
2773
2774 update_class:
2775         cmd = mgmt_pending_new(sk, MGMT_OP_REMOVE_UUID, hdev, data, len);
2776         if (!cmd) {
2777                 err = -ENOMEM;
2778                 goto unlock;
2779         }
2780
2781         err = hci_cmd_sync_queue(hdev, remove_uuid_sync, cmd,
2782                                  mgmt_class_complete);
2783         if (err < 0)
2784                 mgmt_pending_free(cmd);
2785
2786 unlock:
2787         hci_dev_unlock(hdev);
2788         return err;
2789 }
2790
2791 static int set_class_sync(struct hci_dev *hdev, void *data)
2792 {
2793         int err = 0;
2794
2795         if (hci_dev_test_and_clear_flag(hdev, HCI_SERVICE_CACHE)) {
2796                 cancel_delayed_work_sync(&hdev->service_cache);
2797                 err = hci_update_eir_sync(hdev);
2798         }
2799
2800         if (err)
2801                 return err;
2802
2803         return hci_update_class_sync(hdev);
2804 }
2805
2806 static int set_dev_class(struct sock *sk, struct hci_dev *hdev, void *data,
2807                          u16 len)
2808 {
2809         struct mgmt_cp_set_dev_class *cp = data;
2810         struct mgmt_pending_cmd *cmd;
2811         int err;
2812
2813         bt_dev_dbg(hdev, "sock %p", sk);
2814
2815         if (!lmp_bredr_capable(hdev))
2816                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS,
2817                                        MGMT_STATUS_NOT_SUPPORTED);
2818
2819         hci_dev_lock(hdev);
2820
2821         if (pending_eir_or_class(hdev)) {
2822                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS,
2823                                       MGMT_STATUS_BUSY);
2824                 goto unlock;
2825         }
2826
2827         if ((cp->minor & 0x03) != 0 || (cp->major & 0xe0) != 0) {
2828                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEV_CLASS,
2829                                       MGMT_STATUS_INVALID_PARAMS);
2830                 goto unlock;
2831         }
2832
2833         hdev->major_class = cp->major;
2834         hdev->minor_class = cp->minor;
2835
2836         if (!hdev_is_powered(hdev)) {
2837                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEV_CLASS, 0,
2838                                         hdev->dev_class, 3);
2839                 goto unlock;
2840         }
2841
2842         cmd = mgmt_pending_new(sk, MGMT_OP_SET_DEV_CLASS, hdev, data, len);
2843         if (!cmd) {
2844                 err = -ENOMEM;
2845                 goto unlock;
2846         }
2847
2848         err = hci_cmd_sync_queue(hdev, set_class_sync, cmd,
2849                                  mgmt_class_complete);
2850         if (err < 0)
2851                 mgmt_pending_free(cmd);
2852
2853 unlock:
2854         hci_dev_unlock(hdev);
2855         return err;
2856 }
2857
2858 static int load_link_keys(struct sock *sk, struct hci_dev *hdev, void *data,
2859                           u16 len)
2860 {
2861         struct mgmt_cp_load_link_keys *cp = data;
2862         const u16 max_key_count = ((U16_MAX - sizeof(*cp)) /
2863                                    sizeof(struct mgmt_link_key_info));
2864         u16 key_count, expected_len;
2865         bool changed;
2866         int i;
2867
2868         bt_dev_dbg(hdev, "sock %p", sk);
2869
2870         if (!lmp_bredr_capable(hdev))
2871                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS,
2872                                        MGMT_STATUS_NOT_SUPPORTED);
2873
2874         key_count = __le16_to_cpu(cp->key_count);
2875         if (key_count > max_key_count) {
2876                 bt_dev_err(hdev, "load_link_keys: too big key_count value %u",
2877                            key_count);
2878                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS,
2879                                        MGMT_STATUS_INVALID_PARAMS);
2880         }
2881
2882         expected_len = struct_size(cp, keys, key_count);
2883         if (expected_len != len) {
2884                 bt_dev_err(hdev, "load_link_keys: expected %u bytes, got %u bytes",
2885                            expected_len, len);
2886                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS,
2887                                        MGMT_STATUS_INVALID_PARAMS);
2888         }
2889
2890         if (cp->debug_keys != 0x00 && cp->debug_keys != 0x01)
2891                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS,
2892                                        MGMT_STATUS_INVALID_PARAMS);
2893
2894         bt_dev_dbg(hdev, "debug_keys %u key_count %u", cp->debug_keys,
2895                    key_count);
2896
2897         for (i = 0; i < key_count; i++) {
2898                 struct mgmt_link_key_info *key = &cp->keys[i];
2899
2900                 if (key->addr.type != BDADDR_BREDR || key->type > 0x08)
2901                         return mgmt_cmd_status(sk, hdev->id,
2902                                                MGMT_OP_LOAD_LINK_KEYS,
2903                                                MGMT_STATUS_INVALID_PARAMS);
2904         }
2905
2906         hci_dev_lock(hdev);
2907
2908         hci_link_keys_clear(hdev);
2909
2910         if (cp->debug_keys)
2911                 changed = !hci_dev_test_and_set_flag(hdev, HCI_KEEP_DEBUG_KEYS);
2912         else
2913                 changed = hci_dev_test_and_clear_flag(hdev,
2914                                                       HCI_KEEP_DEBUG_KEYS);
2915
2916         if (changed)
2917                 new_settings(hdev, NULL);
2918
2919         for (i = 0; i < key_count; i++) {
2920                 struct mgmt_link_key_info *key = &cp->keys[i];
2921
2922                 if (hci_is_blocked_key(hdev,
2923                                        HCI_BLOCKED_KEY_TYPE_LINKKEY,
2924                                        key->val)) {
2925                         bt_dev_warn(hdev, "Skipping blocked link key for %pMR",
2926                                     &key->addr.bdaddr);
2927                         continue;
2928                 }
2929
2930                 /* Always ignore debug keys and require a new pairing if
2931                  * the user wants to use them.
2932                  */
2933                 if (key->type == HCI_LK_DEBUG_COMBINATION)
2934                         continue;
2935
2936                 hci_add_link_key(hdev, NULL, &key->addr.bdaddr, key->val,
2937                                  key->type, key->pin_len, NULL);
2938         }
2939
2940         mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_LINK_KEYS, 0, NULL, 0);
2941
2942         hci_dev_unlock(hdev);
2943
2944         return 0;
2945 }
2946
2947 static int device_unpaired(struct hci_dev *hdev, bdaddr_t *bdaddr,
2948                            u8 addr_type, struct sock *skip_sk)
2949 {
2950         struct mgmt_ev_device_unpaired ev;
2951
2952         bacpy(&ev.addr.bdaddr, bdaddr);
2953         ev.addr.type = addr_type;
2954
2955         return mgmt_event(MGMT_EV_DEVICE_UNPAIRED, hdev, &ev, sizeof(ev),
2956                           skip_sk);
2957 }
2958
2959 static void unpair_device_complete(struct hci_dev *hdev, void *data, int err)
2960 {
2961         struct mgmt_pending_cmd *cmd = data;
2962         struct mgmt_cp_unpair_device *cp = cmd->param;
2963
2964         if (!err)
2965                 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, cmd->sk);
2966
2967         cmd->cmd_complete(cmd, err);
2968         mgmt_pending_free(cmd);
2969 }
2970
2971 static int unpair_device_sync(struct hci_dev *hdev, void *data)
2972 {
2973         struct mgmt_pending_cmd *cmd = data;
2974         struct mgmt_cp_unpair_device *cp = cmd->param;
2975         struct hci_conn *conn;
2976
2977         if (cp->addr.type == BDADDR_BREDR)
2978                 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK,
2979                                                &cp->addr.bdaddr);
2980         else
2981                 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr,
2982                                                le_addr_type(cp->addr.type));
2983
2984         if (!conn)
2985                 return 0;
2986
2987         return hci_abort_conn_sync(hdev, conn, HCI_ERROR_REMOTE_USER_TERM);
2988 }
2989
2990 static int unpair_device(struct sock *sk, struct hci_dev *hdev, void *data,
2991                          u16 len)
2992 {
2993         struct mgmt_cp_unpair_device *cp = data;
2994         struct mgmt_rp_unpair_device rp;
2995         struct hci_conn_params *params;
2996         struct mgmt_pending_cmd *cmd;
2997         struct hci_conn *conn;
2998         u8 addr_type;
2999         int err;
3000
3001         memset(&rp, 0, sizeof(rp));
3002         bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr);
3003         rp.addr.type = cp->addr.type;
3004
3005         if (!bdaddr_type_is_valid(cp->addr.type))
3006                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE,
3007                                          MGMT_STATUS_INVALID_PARAMS,
3008                                          &rp, sizeof(rp));
3009
3010         if (cp->disconnect != 0x00 && cp->disconnect != 0x01)
3011                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE,
3012                                          MGMT_STATUS_INVALID_PARAMS,
3013                                          &rp, sizeof(rp));
3014
3015         hci_dev_lock(hdev);
3016
3017         if (!hdev_is_powered(hdev)) {
3018                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE,
3019                                         MGMT_STATUS_NOT_POWERED, &rp,
3020                                         sizeof(rp));
3021                 goto unlock;
3022         }
3023
3024         if (cp->addr.type == BDADDR_BREDR) {
3025                 /* If disconnection is requested, then look up the
3026                  * connection. If the remote device is connected, it
3027                  * will be later used to terminate the link.
3028                  *
3029                  * Setting it to NULL explicitly will cause no
3030                  * termination of the link.
3031                  */
3032                 if (cp->disconnect)
3033                         conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK,
3034                                                        &cp->addr.bdaddr);
3035                 else
3036                         conn = NULL;
3037
3038                 err = hci_remove_link_key(hdev, &cp->addr.bdaddr);
3039                 if (err < 0) {
3040                         err = mgmt_cmd_complete(sk, hdev->id,
3041                                                 MGMT_OP_UNPAIR_DEVICE,
3042                                                 MGMT_STATUS_NOT_PAIRED, &rp,
3043                                                 sizeof(rp));
3044                         goto unlock;
3045                 }
3046
3047                 goto done;
3048         }
3049
3050         /* LE address type */
3051         addr_type = le_addr_type(cp->addr.type);
3052
3053         /* Abort any ongoing SMP pairing. Removes ltk and irk if they exist. */
3054         err = smp_cancel_and_remove_pairing(hdev, &cp->addr.bdaddr, addr_type);
3055         if (err < 0) {
3056                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE,
3057                                         MGMT_STATUS_NOT_PAIRED, &rp,
3058                                         sizeof(rp));
3059                 goto unlock;
3060         }
3061
3062         conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr, addr_type);
3063         if (!conn) {
3064                 hci_conn_params_del(hdev, &cp->addr.bdaddr, addr_type);
3065                 goto done;
3066         }
3067
3068
3069         /* Defer clearing up the connection parameters until closing to
3070          * give a chance of keeping them if a repairing happens.
3071          */
3072         set_bit(HCI_CONN_PARAM_REMOVAL_PEND, &conn->flags);
3073
3074         /* Disable auto-connection parameters if present */
3075         params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr, addr_type);
3076         if (params) {
3077                 if (params->explicit_connect)
3078                         params->auto_connect = HCI_AUTO_CONN_EXPLICIT;
3079                 else
3080                         params->auto_connect = HCI_AUTO_CONN_DISABLED;
3081         }
3082
3083         /* If disconnection is not requested, then clear the connection
3084          * variable so that the link is not terminated.
3085          */
3086         if (!cp->disconnect)
3087                 conn = NULL;
3088
3089 done:
3090         /* If the connection variable is set, then termination of the
3091          * link is requested.
3092          */
3093         if (!conn) {
3094                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNPAIR_DEVICE, 0,
3095                                         &rp, sizeof(rp));
3096                 device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, sk);
3097                 goto unlock;
3098         }
3099
3100         cmd = mgmt_pending_new(sk, MGMT_OP_UNPAIR_DEVICE, hdev, cp,
3101                                sizeof(*cp));
3102         if (!cmd) {
3103                 err = -ENOMEM;
3104                 goto unlock;
3105         }
3106
3107         cmd->cmd_complete = addr_cmd_complete;
3108
3109         err = hci_cmd_sync_queue(hdev, unpair_device_sync, cmd,
3110                                  unpair_device_complete);
3111         if (err < 0)
3112                 mgmt_pending_free(cmd);
3113
3114 unlock:
3115         hci_dev_unlock(hdev);
3116         return err;
3117 }
3118
3119 static int disconnect(struct sock *sk, struct hci_dev *hdev, void *data,
3120                       u16 len)
3121 {
3122         struct mgmt_cp_disconnect *cp = data;
3123         struct mgmt_rp_disconnect rp;
3124         struct mgmt_pending_cmd *cmd;
3125         struct hci_conn *conn;
3126         int err;
3127
3128         bt_dev_dbg(hdev, "sock %p", sk);
3129
3130         memset(&rp, 0, sizeof(rp));
3131         bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr);
3132         rp.addr.type = cp->addr.type;
3133
3134         if (!bdaddr_type_is_valid(cp->addr.type))
3135                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT,
3136                                          MGMT_STATUS_INVALID_PARAMS,
3137                                          &rp, sizeof(rp));
3138
3139         hci_dev_lock(hdev);
3140
3141         if (!test_bit(HCI_UP, &hdev->flags)) {
3142                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT,
3143                                         MGMT_STATUS_NOT_POWERED, &rp,
3144                                         sizeof(rp));
3145                 goto failed;
3146         }
3147
3148         if (pending_find(MGMT_OP_DISCONNECT, hdev)) {
3149                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT,
3150                                         MGMT_STATUS_BUSY, &rp, sizeof(rp));
3151                 goto failed;
3152         }
3153
3154         if (cp->addr.type == BDADDR_BREDR)
3155                 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK,
3156                                                &cp->addr.bdaddr);
3157         else
3158                 conn = hci_conn_hash_lookup_le(hdev, &cp->addr.bdaddr,
3159                                                le_addr_type(cp->addr.type));
3160
3161         if (!conn || conn->state == BT_OPEN || conn->state == BT_CLOSED) {
3162                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_DISCONNECT,
3163                                         MGMT_STATUS_NOT_CONNECTED, &rp,
3164                                         sizeof(rp));
3165                 goto failed;
3166         }
3167
3168         cmd = mgmt_pending_add(sk, MGMT_OP_DISCONNECT, hdev, data, len);
3169         if (!cmd) {
3170                 err = -ENOMEM;
3171                 goto failed;
3172         }
3173
3174         cmd->cmd_complete = generic_cmd_complete;
3175
3176         err = hci_disconnect(conn, HCI_ERROR_REMOTE_USER_TERM);
3177         if (err < 0)
3178                 mgmt_pending_remove(cmd);
3179
3180 failed:
3181         hci_dev_unlock(hdev);
3182         return err;
3183 }
3184
3185 static u8 link_to_bdaddr(u8 link_type, u8 addr_type)
3186 {
3187         switch (link_type) {
3188         case LE_LINK:
3189                 switch (addr_type) {
3190                 case ADDR_LE_DEV_PUBLIC:
3191                         return BDADDR_LE_PUBLIC;
3192
3193                 default:
3194                         /* Fallback to LE Random address type */
3195                         return BDADDR_LE_RANDOM;
3196                 }
3197
3198         default:
3199                 /* Fallback to BR/EDR type */
3200                 return BDADDR_BREDR;
3201         }
3202 }
3203
3204 static int get_connections(struct sock *sk, struct hci_dev *hdev, void *data,
3205                            u16 data_len)
3206 {
3207         struct mgmt_rp_get_connections *rp;
3208         struct hci_conn *c;
3209         int err;
3210         u16 i;
3211
3212         bt_dev_dbg(hdev, "sock %p", sk);
3213
3214         hci_dev_lock(hdev);
3215
3216         if (!hdev_is_powered(hdev)) {
3217                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_CONNECTIONS,
3218                                       MGMT_STATUS_NOT_POWERED);
3219                 goto unlock;
3220         }
3221
3222         i = 0;
3223         list_for_each_entry(c, &hdev->conn_hash.list, list) {
3224                 if (test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags))
3225                         i++;
3226         }
3227
3228         rp = kmalloc(struct_size(rp, addr, i), GFP_KERNEL);
3229         if (!rp) {
3230                 err = -ENOMEM;
3231                 goto unlock;
3232         }
3233
3234         i = 0;
3235         list_for_each_entry(c, &hdev->conn_hash.list, list) {
3236                 if (!test_bit(HCI_CONN_MGMT_CONNECTED, &c->flags))
3237                         continue;
3238                 bacpy(&rp->addr[i].bdaddr, &c->dst);
3239                 rp->addr[i].type = link_to_bdaddr(c->type, c->dst_type);
3240                 if (c->type == SCO_LINK || c->type == ESCO_LINK)
3241                         continue;
3242                 i++;
3243         }
3244
3245         rp->conn_count = cpu_to_le16(i);
3246
3247         /* Recalculate length in case of filtered SCO connections, etc */
3248         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONNECTIONS, 0, rp,
3249                                 struct_size(rp, addr, i));
3250
3251         kfree(rp);
3252
3253 unlock:
3254         hci_dev_unlock(hdev);
3255         return err;
3256 }
3257
3258 static int send_pin_code_neg_reply(struct sock *sk, struct hci_dev *hdev,
3259                                    struct mgmt_cp_pin_code_neg_reply *cp)
3260 {
3261         struct mgmt_pending_cmd *cmd;
3262         int err;
3263
3264         cmd = mgmt_pending_add(sk, MGMT_OP_PIN_CODE_NEG_REPLY, hdev, cp,
3265                                sizeof(*cp));
3266         if (!cmd)
3267                 return -ENOMEM;
3268
3269         cmd->cmd_complete = addr_cmd_complete;
3270
3271         err = hci_send_cmd(hdev, HCI_OP_PIN_CODE_NEG_REPLY,
3272                            sizeof(cp->addr.bdaddr), &cp->addr.bdaddr);
3273         if (err < 0)
3274                 mgmt_pending_remove(cmd);
3275
3276         return err;
3277 }
3278
3279 static int pin_code_reply(struct sock *sk, struct hci_dev *hdev, void *data,
3280                           u16 len)
3281 {
3282         struct hci_conn *conn;
3283         struct mgmt_cp_pin_code_reply *cp = data;
3284         struct hci_cp_pin_code_reply reply;
3285         struct mgmt_pending_cmd *cmd;
3286         int err;
3287
3288         bt_dev_dbg(hdev, "sock %p", sk);
3289
3290         hci_dev_lock(hdev);
3291
3292         if (!hdev_is_powered(hdev)) {
3293                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY,
3294                                       MGMT_STATUS_NOT_POWERED);
3295                 goto failed;
3296         }
3297
3298         conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr);
3299         if (!conn) {
3300                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY,
3301                                       MGMT_STATUS_NOT_CONNECTED);
3302                 goto failed;
3303         }
3304
3305         if (conn->pending_sec_level == BT_SECURITY_HIGH && cp->pin_len != 16) {
3306                 struct mgmt_cp_pin_code_neg_reply ncp;
3307
3308                 memcpy(&ncp.addr, &cp->addr, sizeof(ncp.addr));
3309
3310                 bt_dev_err(hdev, "PIN code is not 16 bytes long");
3311
3312                 err = send_pin_code_neg_reply(sk, hdev, &ncp);
3313                 if (err >= 0)
3314                         err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_PIN_CODE_REPLY,
3315                                               MGMT_STATUS_INVALID_PARAMS);
3316
3317                 goto failed;
3318         }
3319
3320         cmd = mgmt_pending_add(sk, MGMT_OP_PIN_CODE_REPLY, hdev, data, len);
3321         if (!cmd) {
3322                 err = -ENOMEM;
3323                 goto failed;
3324         }
3325
3326         cmd->cmd_complete = addr_cmd_complete;
3327
3328         bacpy(&reply.bdaddr, &cp->addr.bdaddr);
3329         reply.pin_len = cp->pin_len;
3330         memcpy(reply.pin_code, cp->pin_code, sizeof(reply.pin_code));
3331
3332         err = hci_send_cmd(hdev, HCI_OP_PIN_CODE_REPLY, sizeof(reply), &reply);
3333         if (err < 0)
3334                 mgmt_pending_remove(cmd);
3335
3336 failed:
3337         hci_dev_unlock(hdev);
3338         return err;
3339 }
3340
3341 static int set_io_capability(struct sock *sk, struct hci_dev *hdev, void *data,
3342                              u16 len)
3343 {
3344         struct mgmt_cp_set_io_capability *cp = data;
3345
3346         bt_dev_dbg(hdev, "sock %p", sk);
3347
3348         if (cp->io_capability > SMP_IO_KEYBOARD_DISPLAY)
3349                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_IO_CAPABILITY,
3350                                        MGMT_STATUS_INVALID_PARAMS);
3351
3352         hci_dev_lock(hdev);
3353
3354         hdev->io_capability = cp->io_capability;
3355
3356         bt_dev_dbg(hdev, "IO capability set to 0x%02x", hdev->io_capability);
3357
3358         hci_dev_unlock(hdev);
3359
3360         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_IO_CAPABILITY, 0,
3361                                  NULL, 0);
3362 }
3363
3364 static struct mgmt_pending_cmd *find_pairing(struct hci_conn *conn)
3365 {
3366         struct hci_dev *hdev = conn->hdev;
3367         struct mgmt_pending_cmd *cmd;
3368
3369         list_for_each_entry(cmd, &hdev->mgmt_pending, list) {
3370                 if (cmd->opcode != MGMT_OP_PAIR_DEVICE)
3371                         continue;
3372
3373                 if (cmd->user_data != conn)
3374                         continue;
3375
3376                 return cmd;
3377         }
3378
3379         return NULL;
3380 }
3381
3382 static int pairing_complete(struct mgmt_pending_cmd *cmd, u8 status)
3383 {
3384         struct mgmt_rp_pair_device rp;
3385         struct hci_conn *conn = cmd->user_data;
3386         int err;
3387
3388         bacpy(&rp.addr.bdaddr, &conn->dst);
3389         rp.addr.type = link_to_bdaddr(conn->type, conn->dst_type);
3390
3391         err = mgmt_cmd_complete(cmd->sk, cmd->index, MGMT_OP_PAIR_DEVICE,
3392                                 status, &rp, sizeof(rp));
3393
3394         /* So we don't get further callbacks for this connection */
3395         conn->connect_cfm_cb = NULL;
3396         conn->security_cfm_cb = NULL;
3397         conn->disconn_cfm_cb = NULL;
3398
3399         hci_conn_drop(conn);
3400
3401         /* The device is paired so there is no need to remove
3402          * its connection parameters anymore.
3403          */
3404         clear_bit(HCI_CONN_PARAM_REMOVAL_PEND, &conn->flags);
3405
3406         hci_conn_put(conn);
3407
3408         return err;
3409 }
3410
3411 void mgmt_smp_complete(struct hci_conn *conn, bool complete)
3412 {
3413         u8 status = complete ? MGMT_STATUS_SUCCESS : MGMT_STATUS_FAILED;
3414         struct mgmt_pending_cmd *cmd;
3415
3416         cmd = find_pairing(conn);
3417         if (cmd) {
3418                 cmd->cmd_complete(cmd, status);
3419                 mgmt_pending_remove(cmd);
3420         }
3421 }
3422
3423 static void pairing_complete_cb(struct hci_conn *conn, u8 status)
3424 {
3425         struct mgmt_pending_cmd *cmd;
3426
3427         BT_DBG("status %u", status);
3428
3429         cmd = find_pairing(conn);
3430         if (!cmd) {
3431                 BT_DBG("Unable to find a pending command");
3432                 return;
3433         }
3434
3435         cmd->cmd_complete(cmd, mgmt_status(status));
3436         mgmt_pending_remove(cmd);
3437 }
3438
3439 static void le_pairing_complete_cb(struct hci_conn *conn, u8 status)
3440 {
3441         struct mgmt_pending_cmd *cmd;
3442
3443         BT_DBG("status %u", status);
3444
3445         if (!status)
3446                 return;
3447
3448         cmd = find_pairing(conn);
3449         if (!cmd) {
3450                 BT_DBG("Unable to find a pending command");
3451                 return;
3452         }
3453
3454         cmd->cmd_complete(cmd, mgmt_status(status));
3455         mgmt_pending_remove(cmd);
3456 }
3457
3458 static int pair_device(struct sock *sk, struct hci_dev *hdev, void *data,
3459                        u16 len)
3460 {
3461         struct mgmt_cp_pair_device *cp = data;
3462         struct mgmt_rp_pair_device rp;
3463         struct mgmt_pending_cmd *cmd;
3464         u8 sec_level, auth_type;
3465         struct hci_conn *conn;
3466         int err;
3467
3468         bt_dev_dbg(hdev, "sock %p", sk);
3469
3470         memset(&rp, 0, sizeof(rp));
3471         bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr);
3472         rp.addr.type = cp->addr.type;
3473
3474         if (!bdaddr_type_is_valid(cp->addr.type))
3475                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE,
3476                                          MGMT_STATUS_INVALID_PARAMS,
3477                                          &rp, sizeof(rp));
3478
3479         if (cp->io_cap > SMP_IO_KEYBOARD_DISPLAY)
3480                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE,
3481                                          MGMT_STATUS_INVALID_PARAMS,
3482                                          &rp, sizeof(rp));
3483
3484         hci_dev_lock(hdev);
3485
3486         if (!hdev_is_powered(hdev)) {
3487                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE,
3488                                         MGMT_STATUS_NOT_POWERED, &rp,
3489                                         sizeof(rp));
3490                 goto unlock;
3491         }
3492
3493         if (hci_bdaddr_is_paired(hdev, &cp->addr.bdaddr, cp->addr.type)) {
3494                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE,
3495                                         MGMT_STATUS_ALREADY_PAIRED, &rp,
3496                                         sizeof(rp));
3497                 goto unlock;
3498         }
3499
3500         sec_level = BT_SECURITY_MEDIUM;
3501         auth_type = HCI_AT_DEDICATED_BONDING;
3502
3503         if (cp->addr.type == BDADDR_BREDR) {
3504                 conn = hci_connect_acl(hdev, &cp->addr.bdaddr, sec_level,
3505                                        auth_type, CONN_REASON_PAIR_DEVICE);
3506         } else {
3507                 u8 addr_type = le_addr_type(cp->addr.type);
3508                 struct hci_conn_params *p;
3509
3510                 /* When pairing a new device, it is expected to remember
3511                  * this device for future connections. Adding the connection
3512                  * parameter information ahead of time allows tracking
3513                  * of the peripheral preferred values and will speed up any
3514                  * further connection establishment.
3515                  *
3516                  * If connection parameters already exist, then they
3517                  * will be kept and this function does nothing.
3518                  */
3519                 p = hci_conn_params_add(hdev, &cp->addr.bdaddr, addr_type);
3520
3521                 if (p->auto_connect == HCI_AUTO_CONN_EXPLICIT)
3522                         p->auto_connect = HCI_AUTO_CONN_DISABLED;
3523
3524                 conn = hci_connect_le_scan(hdev, &cp->addr.bdaddr, addr_type,
3525                                            sec_level, HCI_LE_CONN_TIMEOUT,
3526                                            CONN_REASON_PAIR_DEVICE);
3527         }
3528
3529         if (IS_ERR(conn)) {
3530                 int status;
3531
3532                 if (PTR_ERR(conn) == -EBUSY)
3533                         status = MGMT_STATUS_BUSY;
3534                 else if (PTR_ERR(conn) == -EOPNOTSUPP)
3535                         status = MGMT_STATUS_NOT_SUPPORTED;
3536                 else if (PTR_ERR(conn) == -ECONNREFUSED)
3537                         status = MGMT_STATUS_REJECTED;
3538                 else
3539                         status = MGMT_STATUS_CONNECT_FAILED;
3540
3541                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE,
3542                                         status, &rp, sizeof(rp));
3543                 goto unlock;
3544         }
3545
3546         if (conn->connect_cfm_cb) {
3547                 hci_conn_drop(conn);
3548                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_PAIR_DEVICE,
3549                                         MGMT_STATUS_BUSY, &rp, sizeof(rp));
3550                 goto unlock;
3551         }
3552
3553         cmd = mgmt_pending_add(sk, MGMT_OP_PAIR_DEVICE, hdev, data, len);
3554         if (!cmd) {
3555                 err = -ENOMEM;
3556                 hci_conn_drop(conn);
3557                 goto unlock;
3558         }
3559
3560         cmd->cmd_complete = pairing_complete;
3561
3562         /* For LE, just connecting isn't a proof that the pairing finished */
3563         if (cp->addr.type == BDADDR_BREDR) {
3564                 conn->connect_cfm_cb = pairing_complete_cb;
3565                 conn->security_cfm_cb = pairing_complete_cb;
3566                 conn->disconn_cfm_cb = pairing_complete_cb;
3567         } else {
3568                 conn->connect_cfm_cb = le_pairing_complete_cb;
3569                 conn->security_cfm_cb = le_pairing_complete_cb;
3570                 conn->disconn_cfm_cb = le_pairing_complete_cb;
3571         }
3572
3573         conn->io_capability = cp->io_cap;
3574         cmd->user_data = hci_conn_get(conn);
3575
3576         if ((conn->state == BT_CONNECTED || conn->state == BT_CONFIG) &&
3577             hci_conn_security(conn, sec_level, auth_type, true)) {
3578                 cmd->cmd_complete(cmd, 0);
3579                 mgmt_pending_remove(cmd);
3580         }
3581
3582         err = 0;
3583
3584 unlock:
3585         hci_dev_unlock(hdev);
3586         return err;
3587 }
3588
3589 static int cancel_pair_device(struct sock *sk, struct hci_dev *hdev, void *data,
3590                               u16 len)
3591 {
3592         struct mgmt_addr_info *addr = data;
3593         struct mgmt_pending_cmd *cmd;
3594         struct hci_conn *conn;
3595         int err;
3596
3597         bt_dev_dbg(hdev, "sock %p", sk);
3598
3599         hci_dev_lock(hdev);
3600
3601         if (!hdev_is_powered(hdev)) {
3602                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE,
3603                                       MGMT_STATUS_NOT_POWERED);
3604                 goto unlock;
3605         }
3606
3607         cmd = pending_find(MGMT_OP_PAIR_DEVICE, hdev);
3608         if (!cmd) {
3609                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE,
3610                                       MGMT_STATUS_INVALID_PARAMS);
3611                 goto unlock;
3612         }
3613
3614         conn = cmd->user_data;
3615
3616         if (bacmp(&addr->bdaddr, &conn->dst) != 0) {
3617                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE,
3618                                       MGMT_STATUS_INVALID_PARAMS);
3619                 goto unlock;
3620         }
3621
3622         cmd->cmd_complete(cmd, MGMT_STATUS_CANCELLED);
3623         mgmt_pending_remove(cmd);
3624
3625         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CANCEL_PAIR_DEVICE, 0,
3626                                 addr, sizeof(*addr));
3627
3628         /* Since user doesn't want to proceed with the connection, abort any
3629          * ongoing pairing and then terminate the link if it was created
3630          * because of the pair device action.
3631          */
3632         if (addr->type == BDADDR_BREDR)
3633                 hci_remove_link_key(hdev, &addr->bdaddr);
3634         else
3635                 smp_cancel_and_remove_pairing(hdev, &addr->bdaddr,
3636                                               le_addr_type(addr->type));
3637
3638         if (conn->conn_reason == CONN_REASON_PAIR_DEVICE)
3639                 hci_abort_conn(conn, HCI_ERROR_REMOTE_USER_TERM);
3640
3641 unlock:
3642         hci_dev_unlock(hdev);
3643         return err;
3644 }
3645
3646 static int user_pairing_resp(struct sock *sk, struct hci_dev *hdev,
3647                              struct mgmt_addr_info *addr, u16 mgmt_op,
3648                              u16 hci_op, __le32 passkey)
3649 {
3650         struct mgmt_pending_cmd *cmd;
3651         struct hci_conn *conn;
3652         int err;
3653
3654         hci_dev_lock(hdev);
3655
3656         if (!hdev_is_powered(hdev)) {
3657                 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op,
3658                                         MGMT_STATUS_NOT_POWERED, addr,
3659                                         sizeof(*addr));
3660                 goto done;
3661         }
3662
3663         if (addr->type == BDADDR_BREDR)
3664                 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &addr->bdaddr);
3665         else
3666                 conn = hci_conn_hash_lookup_le(hdev, &addr->bdaddr,
3667                                                le_addr_type(addr->type));
3668
3669         if (!conn) {
3670                 err = mgmt_cmd_complete(sk, hdev->id, mgmt_op,
3671                                         MGMT_STATUS_NOT_CONNECTED, addr,
3672                                         sizeof(*addr));
3673                 goto done;
3674         }
3675
3676         if (addr->type == BDADDR_LE_PUBLIC || addr->type == BDADDR_LE_RANDOM) {
3677                 err = smp_user_confirm_reply(conn, mgmt_op, passkey);
3678                 if (!err)
3679                         err = mgmt_cmd_complete(sk, hdev->id, mgmt_op,
3680                                                 MGMT_STATUS_SUCCESS, addr,
3681                                                 sizeof(*addr));
3682                 else
3683                         err = mgmt_cmd_complete(sk, hdev->id, mgmt_op,
3684                                                 MGMT_STATUS_FAILED, addr,
3685                                                 sizeof(*addr));
3686
3687                 goto done;
3688         }
3689
3690         cmd = mgmt_pending_add(sk, mgmt_op, hdev, addr, sizeof(*addr));
3691         if (!cmd) {
3692                 err = -ENOMEM;
3693                 goto done;
3694         }
3695
3696         cmd->cmd_complete = addr_cmd_complete;
3697
3698         /* Continue with pairing via HCI */
3699         if (hci_op == HCI_OP_USER_PASSKEY_REPLY) {
3700                 struct hci_cp_user_passkey_reply cp;
3701
3702                 bacpy(&cp.bdaddr, &addr->bdaddr);
3703                 cp.passkey = passkey;
3704                 err = hci_send_cmd(hdev, hci_op, sizeof(cp), &cp);
3705         } else
3706                 err = hci_send_cmd(hdev, hci_op, sizeof(addr->bdaddr),
3707                                    &addr->bdaddr);
3708
3709         if (err < 0)
3710                 mgmt_pending_remove(cmd);
3711
3712 done:
3713         hci_dev_unlock(hdev);
3714         return err;
3715 }
3716
3717 static int pin_code_neg_reply(struct sock *sk, struct hci_dev *hdev,
3718                               void *data, u16 len)
3719 {
3720         struct mgmt_cp_pin_code_neg_reply *cp = data;
3721
3722         bt_dev_dbg(hdev, "sock %p", sk);
3723
3724         return user_pairing_resp(sk, hdev, &cp->addr,
3725                                 MGMT_OP_PIN_CODE_NEG_REPLY,
3726                                 HCI_OP_PIN_CODE_NEG_REPLY, 0);
3727 }
3728
3729 static int user_confirm_reply(struct sock *sk, struct hci_dev *hdev, void *data,
3730                               u16 len)
3731 {
3732         struct mgmt_cp_user_confirm_reply *cp = data;
3733
3734         bt_dev_dbg(hdev, "sock %p", sk);
3735
3736         if (len != sizeof(*cp))
3737                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_USER_CONFIRM_REPLY,
3738                                        MGMT_STATUS_INVALID_PARAMS);
3739
3740         return user_pairing_resp(sk, hdev, &cp->addr,
3741                                  MGMT_OP_USER_CONFIRM_REPLY,
3742                                  HCI_OP_USER_CONFIRM_REPLY, 0);
3743 }
3744
3745 static int user_confirm_neg_reply(struct sock *sk, struct hci_dev *hdev,
3746                                   void *data, u16 len)
3747 {
3748         struct mgmt_cp_user_confirm_neg_reply *cp = data;
3749
3750         bt_dev_dbg(hdev, "sock %p", sk);
3751
3752         return user_pairing_resp(sk, hdev, &cp->addr,
3753                                  MGMT_OP_USER_CONFIRM_NEG_REPLY,
3754                                  HCI_OP_USER_CONFIRM_NEG_REPLY, 0);
3755 }
3756
3757 static int user_passkey_reply(struct sock *sk, struct hci_dev *hdev, void *data,
3758                               u16 len)
3759 {
3760         struct mgmt_cp_user_passkey_reply *cp = data;
3761
3762         bt_dev_dbg(hdev, "sock %p", sk);
3763
3764         return user_pairing_resp(sk, hdev, &cp->addr,
3765                                  MGMT_OP_USER_PASSKEY_REPLY,
3766                                  HCI_OP_USER_PASSKEY_REPLY, cp->passkey);
3767 }
3768
3769 static int user_passkey_neg_reply(struct sock *sk, struct hci_dev *hdev,
3770                                   void *data, u16 len)
3771 {
3772         struct mgmt_cp_user_passkey_neg_reply *cp = data;
3773
3774         bt_dev_dbg(hdev, "sock %p", sk);
3775
3776         return user_pairing_resp(sk, hdev, &cp->addr,
3777                                  MGMT_OP_USER_PASSKEY_NEG_REPLY,
3778                                  HCI_OP_USER_PASSKEY_NEG_REPLY, 0);
3779 }
3780
3781 static int adv_expire_sync(struct hci_dev *hdev, u32 flags)
3782 {
3783         struct adv_info *adv_instance;
3784
3785         adv_instance = hci_find_adv_instance(hdev, hdev->cur_adv_instance);
3786         if (!adv_instance)
3787                 return 0;
3788
3789         /* stop if current instance doesn't need to be changed */
3790         if (!(adv_instance->flags & flags))
3791                 return 0;
3792
3793         cancel_adv_timeout(hdev);
3794
3795         adv_instance = hci_get_next_instance(hdev, adv_instance->instance);
3796         if (!adv_instance)
3797                 return 0;
3798
3799         hci_schedule_adv_instance_sync(hdev, adv_instance->instance, true);
3800
3801         return 0;
3802 }
3803
3804 static int name_changed_sync(struct hci_dev *hdev, void *data)
3805 {
3806         return adv_expire_sync(hdev, MGMT_ADV_FLAG_LOCAL_NAME);
3807 }
3808
3809 static void set_name_complete(struct hci_dev *hdev, void *data, int err)
3810 {
3811         struct mgmt_pending_cmd *cmd = data;
3812         struct mgmt_cp_set_local_name *cp = cmd->param;
3813         u8 status = mgmt_status(err);
3814
3815         bt_dev_dbg(hdev, "err %d", err);
3816
3817         if (cmd != pending_find(MGMT_OP_SET_LOCAL_NAME, hdev))
3818                 return;
3819
3820         if (status) {
3821                 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_LOCAL_NAME,
3822                                 status);
3823         } else {
3824                 mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0,
3825                                   cp, sizeof(*cp));
3826
3827                 if (hci_dev_test_flag(hdev, HCI_LE_ADV))
3828                         hci_cmd_sync_queue(hdev, name_changed_sync, NULL, NULL);
3829         }
3830
3831         mgmt_pending_remove(cmd);
3832 }
3833
3834 static int set_name_sync(struct hci_dev *hdev, void *data)
3835 {
3836         if (lmp_bredr_capable(hdev)) {
3837                 hci_update_name_sync(hdev);
3838                 hci_update_eir_sync(hdev);
3839         }
3840
3841         /* The name is stored in the scan response data and so
3842          * no need to update the advertising data here.
3843          */
3844         if (lmp_le_capable(hdev) && hci_dev_test_flag(hdev, HCI_ADVERTISING))
3845                 hci_update_scan_rsp_data_sync(hdev, hdev->cur_adv_instance);
3846
3847         return 0;
3848 }
3849
3850 static int set_local_name(struct sock *sk, struct hci_dev *hdev, void *data,
3851                           u16 len)
3852 {
3853         struct mgmt_cp_set_local_name *cp = data;
3854         struct mgmt_pending_cmd *cmd;
3855         int err;
3856
3857         bt_dev_dbg(hdev, "sock %p", sk);
3858
3859         hci_dev_lock(hdev);
3860
3861         /* If the old values are the same as the new ones just return a
3862          * direct command complete event.
3863          */
3864         if (!memcmp(hdev->dev_name, cp->name, sizeof(hdev->dev_name)) &&
3865             !memcmp(hdev->short_name, cp->short_name,
3866                     sizeof(hdev->short_name))) {
3867                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0,
3868                                         data, len);
3869                 goto failed;
3870         }
3871
3872         memcpy(hdev->short_name, cp->short_name, sizeof(hdev->short_name));
3873
3874         if (!hdev_is_powered(hdev)) {
3875                 memcpy(hdev->dev_name, cp->name, sizeof(hdev->dev_name));
3876
3877                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME, 0,
3878                                         data, len);
3879                 if (err < 0)
3880                         goto failed;
3881
3882                 err = mgmt_limited_event(MGMT_EV_LOCAL_NAME_CHANGED, hdev, data,
3883                                          len, HCI_MGMT_LOCAL_NAME_EVENTS, sk);
3884                 ext_info_changed(hdev, sk);
3885
3886                 goto failed;
3887         }
3888
3889         cmd = mgmt_pending_add(sk, MGMT_OP_SET_LOCAL_NAME, hdev, data, len);
3890         if (!cmd)
3891                 err = -ENOMEM;
3892         else
3893                 err = hci_cmd_sync_queue(hdev, set_name_sync, cmd,
3894                                          set_name_complete);
3895
3896         if (err < 0) {
3897                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_LOCAL_NAME,
3898                                       MGMT_STATUS_FAILED);
3899
3900                 if (cmd)
3901                         mgmt_pending_remove(cmd);
3902
3903                 goto failed;
3904         }
3905
3906         memcpy(hdev->dev_name, cp->name, sizeof(hdev->dev_name));
3907
3908 failed:
3909         hci_dev_unlock(hdev);
3910         return err;
3911 }
3912
3913 static int appearance_changed_sync(struct hci_dev *hdev, void *data)
3914 {
3915         return adv_expire_sync(hdev, MGMT_ADV_FLAG_APPEARANCE);
3916 }
3917
3918 static int set_appearance(struct sock *sk, struct hci_dev *hdev, void *data,
3919                           u16 len)
3920 {
3921         struct mgmt_cp_set_appearance *cp = data;
3922         u16 appearance;
3923         int err;
3924
3925         bt_dev_dbg(hdev, "sock %p", sk);
3926
3927         if (!lmp_le_capable(hdev))
3928                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_APPEARANCE,
3929                                        MGMT_STATUS_NOT_SUPPORTED);
3930
3931         appearance = le16_to_cpu(cp->appearance);
3932
3933         hci_dev_lock(hdev);
3934
3935         if (hdev->appearance != appearance) {
3936                 hdev->appearance = appearance;
3937
3938                 if (hci_dev_test_flag(hdev, HCI_LE_ADV))
3939                         hci_cmd_sync_queue(hdev, appearance_changed_sync, NULL,
3940                                            NULL);
3941
3942                 ext_info_changed(hdev, sk);
3943         }
3944
3945         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_APPEARANCE, 0, NULL,
3946                                 0);
3947
3948         hci_dev_unlock(hdev);
3949
3950         return err;
3951 }
3952
3953 static int get_phy_configuration(struct sock *sk, struct hci_dev *hdev,
3954                                  void *data, u16 len)
3955 {
3956         struct mgmt_rp_get_phy_configuration rp;
3957
3958         bt_dev_dbg(hdev, "sock %p", sk);
3959
3960         hci_dev_lock(hdev);
3961
3962         memset(&rp, 0, sizeof(rp));
3963
3964         rp.supported_phys = cpu_to_le32(get_supported_phys(hdev));
3965         rp.selected_phys = cpu_to_le32(get_selected_phys(hdev));
3966         rp.configurable_phys = cpu_to_le32(get_configurable_phys(hdev));
3967
3968         hci_dev_unlock(hdev);
3969
3970         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_PHY_CONFIGURATION, 0,
3971                                  &rp, sizeof(rp));
3972 }
3973
3974 int mgmt_phy_configuration_changed(struct hci_dev *hdev, struct sock *skip)
3975 {
3976         struct mgmt_ev_phy_configuration_changed ev;
3977
3978         memset(&ev, 0, sizeof(ev));
3979
3980         ev.selected_phys = cpu_to_le32(get_selected_phys(hdev));
3981
3982         return mgmt_event(MGMT_EV_PHY_CONFIGURATION_CHANGED, hdev, &ev,
3983                           sizeof(ev), skip);
3984 }
3985
3986 static void set_default_phy_complete(struct hci_dev *hdev, void *data, int err)
3987 {
3988         struct mgmt_pending_cmd *cmd = data;
3989         struct sk_buff *skb = cmd->skb;
3990         u8 status = mgmt_status(err);
3991
3992         if (cmd != pending_find(MGMT_OP_SET_PHY_CONFIGURATION, hdev))
3993                 return;
3994
3995         if (!status) {
3996                 if (!skb)
3997                         status = MGMT_STATUS_FAILED;
3998                 else if (IS_ERR(skb))
3999                         status = mgmt_status(PTR_ERR(skb));
4000                 else
4001                         status = mgmt_status(skb->data[0]);
4002         }
4003
4004         bt_dev_dbg(hdev, "status %d", status);
4005
4006         if (status) {
4007                 mgmt_cmd_status(cmd->sk, hdev->id,
4008                                 MGMT_OP_SET_PHY_CONFIGURATION, status);
4009         } else {
4010                 mgmt_cmd_complete(cmd->sk, hdev->id,
4011                                   MGMT_OP_SET_PHY_CONFIGURATION, 0,
4012                                   NULL, 0);
4013
4014                 mgmt_phy_configuration_changed(hdev, cmd->sk);
4015         }
4016
4017         if (skb && !IS_ERR(skb))
4018                 kfree_skb(skb);
4019
4020         mgmt_pending_remove(cmd);
4021 }
4022
4023 static int set_default_phy_sync(struct hci_dev *hdev, void *data)
4024 {
4025         struct mgmt_pending_cmd *cmd = data;
4026         struct mgmt_cp_set_phy_configuration *cp = cmd->param;
4027         struct hci_cp_le_set_default_phy cp_phy;
4028         u32 selected_phys = __le32_to_cpu(cp->selected_phys);
4029
4030         memset(&cp_phy, 0, sizeof(cp_phy));
4031
4032         if (!(selected_phys & MGMT_PHY_LE_TX_MASK))
4033                 cp_phy.all_phys |= 0x01;
4034
4035         if (!(selected_phys & MGMT_PHY_LE_RX_MASK))
4036                 cp_phy.all_phys |= 0x02;
4037
4038         if (selected_phys & MGMT_PHY_LE_1M_TX)
4039                 cp_phy.tx_phys |= HCI_LE_SET_PHY_1M;
4040
4041         if (selected_phys & MGMT_PHY_LE_2M_TX)
4042                 cp_phy.tx_phys |= HCI_LE_SET_PHY_2M;
4043
4044         if (selected_phys & MGMT_PHY_LE_CODED_TX)
4045                 cp_phy.tx_phys |= HCI_LE_SET_PHY_CODED;
4046
4047         if (selected_phys & MGMT_PHY_LE_1M_RX)
4048                 cp_phy.rx_phys |= HCI_LE_SET_PHY_1M;
4049
4050         if (selected_phys & MGMT_PHY_LE_2M_RX)
4051                 cp_phy.rx_phys |= HCI_LE_SET_PHY_2M;
4052
4053         if (selected_phys & MGMT_PHY_LE_CODED_RX)
4054                 cp_phy.rx_phys |= HCI_LE_SET_PHY_CODED;
4055
4056         cmd->skb =  __hci_cmd_sync(hdev, HCI_OP_LE_SET_DEFAULT_PHY,
4057                                    sizeof(cp_phy), &cp_phy, HCI_CMD_TIMEOUT);
4058
4059         return 0;
4060 }
4061
4062 static int set_phy_configuration(struct sock *sk, struct hci_dev *hdev,
4063                                  void *data, u16 len)
4064 {
4065         struct mgmt_cp_set_phy_configuration *cp = data;
4066         struct mgmt_pending_cmd *cmd;
4067         u32 selected_phys, configurable_phys, supported_phys, unconfigure_phys;
4068         u16 pkt_type = (HCI_DH1 | HCI_DM1);
4069         bool changed = false;
4070         int err;
4071
4072         bt_dev_dbg(hdev, "sock %p", sk);
4073
4074         configurable_phys = get_configurable_phys(hdev);
4075         supported_phys = get_supported_phys(hdev);
4076         selected_phys = __le32_to_cpu(cp->selected_phys);
4077
4078         if (selected_phys & ~supported_phys)
4079                 return mgmt_cmd_status(sk, hdev->id,
4080                                        MGMT_OP_SET_PHY_CONFIGURATION,
4081                                        MGMT_STATUS_INVALID_PARAMS);
4082
4083         unconfigure_phys = supported_phys & ~configurable_phys;
4084
4085         if ((selected_phys & unconfigure_phys) != unconfigure_phys)
4086                 return mgmt_cmd_status(sk, hdev->id,
4087                                        MGMT_OP_SET_PHY_CONFIGURATION,
4088                                        MGMT_STATUS_INVALID_PARAMS);
4089
4090         if (selected_phys == get_selected_phys(hdev))
4091                 return mgmt_cmd_complete(sk, hdev->id,
4092                                          MGMT_OP_SET_PHY_CONFIGURATION,
4093                                          0, NULL, 0);
4094
4095         hci_dev_lock(hdev);
4096
4097         if (!hdev_is_powered(hdev)) {
4098                 err = mgmt_cmd_status(sk, hdev->id,
4099                                       MGMT_OP_SET_PHY_CONFIGURATION,
4100                                       MGMT_STATUS_REJECTED);
4101                 goto unlock;
4102         }
4103
4104         if (pending_find(MGMT_OP_SET_PHY_CONFIGURATION, hdev)) {
4105                 err = mgmt_cmd_status(sk, hdev->id,
4106                                       MGMT_OP_SET_PHY_CONFIGURATION,
4107                                       MGMT_STATUS_BUSY);
4108                 goto unlock;
4109         }
4110
4111         if (selected_phys & MGMT_PHY_BR_1M_3SLOT)
4112                 pkt_type |= (HCI_DH3 | HCI_DM3);
4113         else
4114                 pkt_type &= ~(HCI_DH3 | HCI_DM3);
4115
4116         if (selected_phys & MGMT_PHY_BR_1M_5SLOT)
4117                 pkt_type |= (HCI_DH5 | HCI_DM5);
4118         else
4119                 pkt_type &= ~(HCI_DH5 | HCI_DM5);
4120
4121         if (selected_phys & MGMT_PHY_EDR_2M_1SLOT)
4122                 pkt_type &= ~HCI_2DH1;
4123         else
4124                 pkt_type |= HCI_2DH1;
4125
4126         if (selected_phys & MGMT_PHY_EDR_2M_3SLOT)
4127                 pkt_type &= ~HCI_2DH3;
4128         else
4129                 pkt_type |= HCI_2DH3;
4130
4131         if (selected_phys & MGMT_PHY_EDR_2M_5SLOT)
4132                 pkt_type &= ~HCI_2DH5;
4133         else
4134                 pkt_type |= HCI_2DH5;
4135
4136         if (selected_phys & MGMT_PHY_EDR_3M_1SLOT)
4137                 pkt_type &= ~HCI_3DH1;
4138         else
4139                 pkt_type |= HCI_3DH1;
4140
4141         if (selected_phys & MGMT_PHY_EDR_3M_3SLOT)
4142                 pkt_type &= ~HCI_3DH3;
4143         else
4144                 pkt_type |= HCI_3DH3;
4145
4146         if (selected_phys & MGMT_PHY_EDR_3M_5SLOT)
4147                 pkt_type &= ~HCI_3DH5;
4148         else
4149                 pkt_type |= HCI_3DH5;
4150
4151         if (pkt_type != hdev->pkt_type) {
4152                 hdev->pkt_type = pkt_type;
4153                 changed = true;
4154         }
4155
4156         if ((selected_phys & MGMT_PHY_LE_MASK) ==
4157             (get_selected_phys(hdev) & MGMT_PHY_LE_MASK)) {
4158                 if (changed)
4159                         mgmt_phy_configuration_changed(hdev, sk);
4160
4161                 err = mgmt_cmd_complete(sk, hdev->id,
4162                                         MGMT_OP_SET_PHY_CONFIGURATION,
4163                                         0, NULL, 0);
4164
4165                 goto unlock;
4166         }
4167
4168         cmd = mgmt_pending_add(sk, MGMT_OP_SET_PHY_CONFIGURATION, hdev, data,
4169                                len);
4170         if (!cmd)
4171                 err = -ENOMEM;
4172         else
4173                 err = hci_cmd_sync_queue(hdev, set_default_phy_sync, cmd,
4174                                          set_default_phy_complete);
4175
4176         if (err < 0) {
4177                 err = mgmt_cmd_status(sk, hdev->id,
4178                                       MGMT_OP_SET_PHY_CONFIGURATION,
4179                                       MGMT_STATUS_FAILED);
4180
4181                 if (cmd)
4182                         mgmt_pending_remove(cmd);
4183         }
4184
4185 unlock:
4186         hci_dev_unlock(hdev);
4187
4188         return err;
4189 }
4190
4191 static int set_blocked_keys(struct sock *sk, struct hci_dev *hdev, void *data,
4192                             u16 len)
4193 {
4194         int err = MGMT_STATUS_SUCCESS;
4195         struct mgmt_cp_set_blocked_keys *keys = data;
4196         const u16 max_key_count = ((U16_MAX - sizeof(*keys)) /
4197                                    sizeof(struct mgmt_blocked_key_info));
4198         u16 key_count, expected_len;
4199         int i;
4200
4201         bt_dev_dbg(hdev, "sock %p", sk);
4202
4203         key_count = __le16_to_cpu(keys->key_count);
4204         if (key_count > max_key_count) {
4205                 bt_dev_err(hdev, "too big key_count value %u", key_count);
4206                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS,
4207                                        MGMT_STATUS_INVALID_PARAMS);
4208         }
4209
4210         expected_len = struct_size(keys, keys, key_count);
4211         if (expected_len != len) {
4212                 bt_dev_err(hdev, "expected %u bytes, got %u bytes",
4213                            expected_len, len);
4214                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS,
4215                                        MGMT_STATUS_INVALID_PARAMS);
4216         }
4217
4218         hci_dev_lock(hdev);
4219
4220         hci_blocked_keys_clear(hdev);
4221
4222         for (i = 0; i < key_count; ++i) {
4223                 struct blocked_key *b = kzalloc(sizeof(*b), GFP_KERNEL);
4224
4225                 if (!b) {
4226                         err = MGMT_STATUS_NO_RESOURCES;
4227                         break;
4228                 }
4229
4230                 b->type = keys->keys[i].type;
4231                 memcpy(b->val, keys->keys[i].val, sizeof(b->val));
4232                 list_add_rcu(&b->list, &hdev->blocked_keys);
4233         }
4234         hci_dev_unlock(hdev);
4235
4236         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_BLOCKED_KEYS,
4237                                 err, NULL, 0);
4238 }
4239
4240 static int set_wideband_speech(struct sock *sk, struct hci_dev *hdev,
4241                                void *data, u16 len)
4242 {
4243         struct mgmt_mode *cp = data;
4244         int err;
4245         bool changed = false;
4246
4247         bt_dev_dbg(hdev, "sock %p", sk);
4248
4249         if (!test_bit(HCI_QUIRK_WIDEBAND_SPEECH_SUPPORTED, &hdev->quirks))
4250                 return mgmt_cmd_status(sk, hdev->id,
4251                                        MGMT_OP_SET_WIDEBAND_SPEECH,
4252                                        MGMT_STATUS_NOT_SUPPORTED);
4253
4254         if (cp->val != 0x00 && cp->val != 0x01)
4255                 return mgmt_cmd_status(sk, hdev->id,
4256                                        MGMT_OP_SET_WIDEBAND_SPEECH,
4257                                        MGMT_STATUS_INVALID_PARAMS);
4258
4259         hci_dev_lock(hdev);
4260
4261         if (hdev_is_powered(hdev) &&
4262             !!cp->val != hci_dev_test_flag(hdev,
4263                                            HCI_WIDEBAND_SPEECH_ENABLED)) {
4264                 err = mgmt_cmd_status(sk, hdev->id,
4265                                       MGMT_OP_SET_WIDEBAND_SPEECH,
4266                                       MGMT_STATUS_REJECTED);
4267                 goto unlock;
4268         }
4269
4270         if (cp->val)
4271                 changed = !hci_dev_test_and_set_flag(hdev,
4272                                                    HCI_WIDEBAND_SPEECH_ENABLED);
4273         else
4274                 changed = hci_dev_test_and_clear_flag(hdev,
4275                                                    HCI_WIDEBAND_SPEECH_ENABLED);
4276
4277         err = send_settings_rsp(sk, MGMT_OP_SET_WIDEBAND_SPEECH, hdev);
4278         if (err < 0)
4279                 goto unlock;
4280
4281         if (changed)
4282                 err = new_settings(hdev, sk);
4283
4284 unlock:
4285         hci_dev_unlock(hdev);
4286         return err;
4287 }
4288
4289 static int read_controller_cap(struct sock *sk, struct hci_dev *hdev,
4290                                void *data, u16 data_len)
4291 {
4292         char buf[20];
4293         struct mgmt_rp_read_controller_cap *rp = (void *)buf;
4294         u16 cap_len = 0;
4295         u8 flags = 0;
4296         u8 tx_power_range[2];
4297
4298         bt_dev_dbg(hdev, "sock %p", sk);
4299
4300         memset(&buf, 0, sizeof(buf));
4301
4302         hci_dev_lock(hdev);
4303
4304         /* When the Read Simple Pairing Options command is supported, then
4305          * the remote public key validation is supported.
4306          *
4307          * Alternatively, when Microsoft extensions are available, they can
4308          * indicate support for public key validation as well.
4309          */
4310         if ((hdev->commands[41] & 0x08) || msft_curve_validity(hdev))
4311                 flags |= 0x01;  /* Remote public key validation (BR/EDR) */
4312
4313         flags |= 0x02;          /* Remote public key validation (LE) */
4314
4315         /* When the Read Encryption Key Size command is supported, then the
4316          * encryption key size is enforced.
4317          */
4318         if (hdev->commands[20] & 0x10)
4319                 flags |= 0x04;  /* Encryption key size enforcement (BR/EDR) */
4320
4321         flags |= 0x08;          /* Encryption key size enforcement (LE) */
4322
4323         cap_len = eir_append_data(rp->cap, cap_len, MGMT_CAP_SEC_FLAGS,
4324                                   &flags, 1);
4325
4326         /* When the Read Simple Pairing Options command is supported, then
4327          * also max encryption key size information is provided.
4328          */
4329         if (hdev->commands[41] & 0x08)
4330                 cap_len = eir_append_le16(rp->cap, cap_len,
4331                                           MGMT_CAP_MAX_ENC_KEY_SIZE,
4332                                           hdev->max_enc_key_size);
4333
4334         cap_len = eir_append_le16(rp->cap, cap_len,
4335                                   MGMT_CAP_SMP_MAX_ENC_KEY_SIZE,
4336                                   SMP_MAX_ENC_KEY_SIZE);
4337
4338         /* Append the min/max LE tx power parameters if we were able to fetch
4339          * it from the controller
4340          */
4341         if (hdev->commands[38] & 0x80) {
4342                 memcpy(&tx_power_range[0], &hdev->min_le_tx_power, 1);
4343                 memcpy(&tx_power_range[1], &hdev->max_le_tx_power, 1);
4344                 cap_len = eir_append_data(rp->cap, cap_len, MGMT_CAP_LE_TX_PWR,
4345                                           tx_power_range, 2);
4346         }
4347
4348         rp->cap_len = cpu_to_le16(cap_len);
4349
4350         hci_dev_unlock(hdev);
4351
4352         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_CONTROLLER_CAP, 0,
4353                                  rp, sizeof(*rp) + cap_len);
4354 }
4355
4356 #ifdef CONFIG_BT_FEATURE_DEBUG
4357 /* d4992530-b9ec-469f-ab01-6c481c47da1c */
4358 static const u8 debug_uuid[16] = {
4359         0x1c, 0xda, 0x47, 0x1c, 0x48, 0x6c, 0x01, 0xab,
4360         0x9f, 0x46, 0xec, 0xb9, 0x30, 0x25, 0x99, 0xd4,
4361 };
4362 #endif
4363
4364 /* 330859bc-7506-492d-9370-9a6f0614037f */
4365 static const u8 quality_report_uuid[16] = {
4366         0x7f, 0x03, 0x14, 0x06, 0x6f, 0x9a, 0x70, 0x93,
4367         0x2d, 0x49, 0x06, 0x75, 0xbc, 0x59, 0x08, 0x33,
4368 };
4369
4370 /* a6695ace-ee7f-4fb9-881a-5fac66c629af */
4371 static const u8 offload_codecs_uuid[16] = {
4372         0xaf, 0x29, 0xc6, 0x66, 0xac, 0x5f, 0x1a, 0x88,
4373         0xb9, 0x4f, 0x7f, 0xee, 0xce, 0x5a, 0x69, 0xa6,
4374 };
4375
4376 /* 671b10b5-42c0-4696-9227-eb28d1b049d6 */
4377 static const u8 le_simultaneous_roles_uuid[16] = {
4378         0xd6, 0x49, 0xb0, 0xd1, 0x28, 0xeb, 0x27, 0x92,
4379         0x96, 0x46, 0xc0, 0x42, 0xb5, 0x10, 0x1b, 0x67,
4380 };
4381
4382 /* 15c0a148-c273-11ea-b3de-0242ac130004 */
4383 static const u8 rpa_resolution_uuid[16] = {
4384         0x04, 0x00, 0x13, 0xac, 0x42, 0x02, 0xde, 0xb3,
4385         0xea, 0x11, 0x73, 0xc2, 0x48, 0xa1, 0xc0, 0x15,
4386 };
4387
4388 /* 6fbaf188-05e0-496a-9885-d6ddfdb4e03e */
4389 static const u8 iso_socket_uuid[16] = {
4390         0x3e, 0xe0, 0xb4, 0xfd, 0xdd, 0xd6, 0x85, 0x98,
4391         0x6a, 0x49, 0xe0, 0x05, 0x88, 0xf1, 0xba, 0x6f,
4392 };
4393
4394 /* 2ce463d7-7a03-4d8d-bf05-5f24e8f36e76 */
4395 static const u8 mgmt_mesh_uuid[16] = {
4396         0x76, 0x6e, 0xf3, 0xe8, 0x24, 0x5f, 0x05, 0xbf,
4397         0x8d, 0x4d, 0x03, 0x7a, 0xd7, 0x63, 0xe4, 0x2c,
4398 };
4399
4400 static int read_exp_features_info(struct sock *sk, struct hci_dev *hdev,
4401                                   void *data, u16 data_len)
4402 {
4403         struct mgmt_rp_read_exp_features_info *rp;
4404         size_t len;
4405         u16 idx = 0;
4406         u32 flags;
4407         int status;
4408
4409         bt_dev_dbg(hdev, "sock %p", sk);
4410
4411         /* Enough space for 7 features */
4412         len = sizeof(*rp) + (sizeof(rp->features[0]) * 7);
4413         rp = kzalloc(len, GFP_KERNEL);
4414         if (!rp)
4415                 return -ENOMEM;
4416
4417 #ifdef CONFIG_BT_FEATURE_DEBUG
4418         if (!hdev) {
4419                 flags = bt_dbg_get() ? BIT(0) : 0;
4420
4421                 memcpy(rp->features[idx].uuid, debug_uuid, 16);
4422                 rp->features[idx].flags = cpu_to_le32(flags);
4423                 idx++;
4424         }
4425 #endif
4426
4427         if (hdev && hci_dev_le_state_simultaneous(hdev)) {
4428                 if (hci_dev_test_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES))
4429                         flags = BIT(0);
4430                 else
4431                         flags = 0;
4432
4433                 memcpy(rp->features[idx].uuid, le_simultaneous_roles_uuid, 16);
4434                 rp->features[idx].flags = cpu_to_le32(flags);
4435                 idx++;
4436         }
4437
4438         if (hdev && ll_privacy_capable(hdev)) {
4439                 if (hci_dev_test_flag(hdev, HCI_ENABLE_LL_PRIVACY))
4440                         flags = BIT(0) | BIT(1);
4441                 else
4442                         flags = BIT(1);
4443
4444                 memcpy(rp->features[idx].uuid, rpa_resolution_uuid, 16);
4445                 rp->features[idx].flags = cpu_to_le32(flags);
4446                 idx++;
4447         }
4448
4449         if (hdev && (aosp_has_quality_report(hdev) ||
4450                      hdev->set_quality_report)) {
4451                 if (hci_dev_test_flag(hdev, HCI_QUALITY_REPORT))
4452                         flags = BIT(0);
4453                 else
4454                         flags = 0;
4455
4456                 memcpy(rp->features[idx].uuid, quality_report_uuid, 16);
4457                 rp->features[idx].flags = cpu_to_le32(flags);
4458                 idx++;
4459         }
4460
4461         if (hdev && hdev->get_data_path_id) {
4462                 if (hci_dev_test_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED))
4463                         flags = BIT(0);
4464                 else
4465                         flags = 0;
4466
4467                 memcpy(rp->features[idx].uuid, offload_codecs_uuid, 16);
4468                 rp->features[idx].flags = cpu_to_le32(flags);
4469                 idx++;
4470         }
4471
4472         if (IS_ENABLED(CONFIG_BT_LE)) {
4473                 flags = iso_enabled() ? BIT(0) : 0;
4474                 memcpy(rp->features[idx].uuid, iso_socket_uuid, 16);
4475                 rp->features[idx].flags = cpu_to_le32(flags);
4476                 idx++;
4477         }
4478
4479         if (hdev && lmp_le_capable(hdev)) {
4480                 if (hci_dev_test_flag(hdev, HCI_MESH_EXPERIMENTAL))
4481                         flags = BIT(0);
4482                 else
4483                         flags = 0;
4484
4485                 memcpy(rp->features[idx].uuid, mgmt_mesh_uuid, 16);
4486                 rp->features[idx].flags = cpu_to_le32(flags);
4487                 idx++;
4488         }
4489
4490         rp->feature_count = cpu_to_le16(idx);
4491
4492         /* After reading the experimental features information, enable
4493          * the events to update client on any future change.
4494          */
4495         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4496
4497         status = mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE,
4498                                    MGMT_OP_READ_EXP_FEATURES_INFO,
4499                                    0, rp, sizeof(*rp) + (20 * idx));
4500
4501         kfree(rp);
4502         return status;
4503 }
4504
4505 static int exp_ll_privacy_feature_changed(bool enabled, struct hci_dev *hdev,
4506                                           struct sock *skip)
4507 {
4508         struct mgmt_ev_exp_feature_changed ev;
4509
4510         memset(&ev, 0, sizeof(ev));
4511         memcpy(ev.uuid, rpa_resolution_uuid, 16);
4512         ev.flags = cpu_to_le32((enabled ? BIT(0) : 0) | BIT(1));
4513
4514         // Do we need to be atomic with the conn_flags?
4515         if (enabled && privacy_mode_capable(hdev))
4516                 hdev->conn_flags |= HCI_CONN_FLAG_DEVICE_PRIVACY;
4517         else
4518                 hdev->conn_flags &= ~HCI_CONN_FLAG_DEVICE_PRIVACY;
4519
4520         return mgmt_limited_event(MGMT_EV_EXP_FEATURE_CHANGED, hdev,
4521                                   &ev, sizeof(ev),
4522                                   HCI_MGMT_EXP_FEATURE_EVENTS, skip);
4523
4524 }
4525
4526 static int exp_feature_changed(struct hci_dev *hdev, const u8 *uuid,
4527                                bool enabled, struct sock *skip)
4528 {
4529         struct mgmt_ev_exp_feature_changed ev;
4530
4531         memset(&ev, 0, sizeof(ev));
4532         memcpy(ev.uuid, uuid, 16);
4533         ev.flags = cpu_to_le32(enabled ? BIT(0) : 0);
4534
4535         return mgmt_limited_event(MGMT_EV_EXP_FEATURE_CHANGED, hdev,
4536                                   &ev, sizeof(ev),
4537                                   HCI_MGMT_EXP_FEATURE_EVENTS, skip);
4538 }
4539
4540 #define EXP_FEAT(_uuid, _set_func)      \
4541 {                                       \
4542         .uuid = _uuid,                  \
4543         .set_func = _set_func,          \
4544 }
4545
4546 /* The zero key uuid is special. Multiple exp features are set through it. */
4547 static int set_zero_key_func(struct sock *sk, struct hci_dev *hdev,
4548                              struct mgmt_cp_set_exp_feature *cp, u16 data_len)
4549 {
4550         struct mgmt_rp_set_exp_feature rp;
4551
4552         memset(rp.uuid, 0, 16);
4553         rp.flags = cpu_to_le32(0);
4554
4555 #ifdef CONFIG_BT_FEATURE_DEBUG
4556         if (!hdev) {
4557                 bool changed = bt_dbg_get();
4558
4559                 bt_dbg_set(false);
4560
4561                 if (changed)
4562                         exp_feature_changed(NULL, ZERO_KEY, false, sk);
4563         }
4564 #endif
4565
4566         if (hdev && use_ll_privacy(hdev) && !hdev_is_powered(hdev)) {
4567                 bool changed;
4568
4569                 changed = hci_dev_test_and_clear_flag(hdev,
4570                                                       HCI_ENABLE_LL_PRIVACY);
4571                 if (changed)
4572                         exp_feature_changed(hdev, rpa_resolution_uuid, false,
4573                                             sk);
4574         }
4575
4576         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4577
4578         return mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE,
4579                                  MGMT_OP_SET_EXP_FEATURE, 0,
4580                                  &rp, sizeof(rp));
4581 }
4582
4583 #ifdef CONFIG_BT_FEATURE_DEBUG
4584 static int set_debug_func(struct sock *sk, struct hci_dev *hdev,
4585                           struct mgmt_cp_set_exp_feature *cp, u16 data_len)
4586 {
4587         struct mgmt_rp_set_exp_feature rp;
4588
4589         bool val, changed;
4590         int err;
4591
4592         /* Command requires to use the non-controller index */
4593         if (hdev)
4594                 return mgmt_cmd_status(sk, hdev->id,
4595                                        MGMT_OP_SET_EXP_FEATURE,
4596                                        MGMT_STATUS_INVALID_INDEX);
4597
4598         /* Parameters are limited to a single octet */
4599         if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1)
4600                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4601                                        MGMT_OP_SET_EXP_FEATURE,
4602                                        MGMT_STATUS_INVALID_PARAMS);
4603
4604         /* Only boolean on/off is supported */
4605         if (cp->param[0] != 0x00 && cp->param[0] != 0x01)
4606                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4607                                        MGMT_OP_SET_EXP_FEATURE,
4608                                        MGMT_STATUS_INVALID_PARAMS);
4609
4610         val = !!cp->param[0];
4611         changed = val ? !bt_dbg_get() : bt_dbg_get();
4612         bt_dbg_set(val);
4613
4614         memcpy(rp.uuid, debug_uuid, 16);
4615         rp.flags = cpu_to_le32(val ? BIT(0) : 0);
4616
4617         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4618
4619         err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE,
4620                                 MGMT_OP_SET_EXP_FEATURE, 0,
4621                                 &rp, sizeof(rp));
4622
4623         if (changed)
4624                 exp_feature_changed(hdev, debug_uuid, val, sk);
4625
4626         return err;
4627 }
4628 #endif
4629
4630 static int set_mgmt_mesh_func(struct sock *sk, struct hci_dev *hdev,
4631                               struct mgmt_cp_set_exp_feature *cp, u16 data_len)
4632 {
4633         struct mgmt_rp_set_exp_feature rp;
4634         bool val, changed;
4635         int err;
4636
4637         /* Command requires to use the controller index */
4638         if (!hdev)
4639                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4640                                        MGMT_OP_SET_EXP_FEATURE,
4641                                        MGMT_STATUS_INVALID_INDEX);
4642
4643         /* Parameters are limited to a single octet */
4644         if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1)
4645                 return mgmt_cmd_status(sk, hdev->id,
4646                                        MGMT_OP_SET_EXP_FEATURE,
4647                                        MGMT_STATUS_INVALID_PARAMS);
4648
4649         /* Only boolean on/off is supported */
4650         if (cp->param[0] != 0x00 && cp->param[0] != 0x01)
4651                 return mgmt_cmd_status(sk, hdev->id,
4652                                        MGMT_OP_SET_EXP_FEATURE,
4653                                        MGMT_STATUS_INVALID_PARAMS);
4654
4655         val = !!cp->param[0];
4656
4657         if (val) {
4658                 changed = !hci_dev_test_and_set_flag(hdev,
4659                                                      HCI_MESH_EXPERIMENTAL);
4660         } else {
4661                 hci_dev_clear_flag(hdev, HCI_MESH);
4662                 changed = hci_dev_test_and_clear_flag(hdev,
4663                                                       HCI_MESH_EXPERIMENTAL);
4664         }
4665
4666         memcpy(rp.uuid, mgmt_mesh_uuid, 16);
4667         rp.flags = cpu_to_le32(val ? BIT(0) : 0);
4668
4669         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4670
4671         err = mgmt_cmd_complete(sk, hdev->id,
4672                                 MGMT_OP_SET_EXP_FEATURE, 0,
4673                                 &rp, sizeof(rp));
4674
4675         if (changed)
4676                 exp_feature_changed(hdev, mgmt_mesh_uuid, val, sk);
4677
4678         return err;
4679 }
4680
4681 static int set_rpa_resolution_func(struct sock *sk, struct hci_dev *hdev,
4682                                    struct mgmt_cp_set_exp_feature *cp,
4683                                    u16 data_len)
4684 {
4685         struct mgmt_rp_set_exp_feature rp;
4686         bool val, changed;
4687         int err;
4688         u32 flags;
4689
4690         /* Command requires to use the controller index */
4691         if (!hdev)
4692                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4693                                        MGMT_OP_SET_EXP_FEATURE,
4694                                        MGMT_STATUS_INVALID_INDEX);
4695
4696         /* Changes can only be made when controller is powered down */
4697         if (hdev_is_powered(hdev))
4698                 return mgmt_cmd_status(sk, hdev->id,
4699                                        MGMT_OP_SET_EXP_FEATURE,
4700                                        MGMT_STATUS_REJECTED);
4701
4702         /* Parameters are limited to a single octet */
4703         if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1)
4704                 return mgmt_cmd_status(sk, hdev->id,
4705                                        MGMT_OP_SET_EXP_FEATURE,
4706                                        MGMT_STATUS_INVALID_PARAMS);
4707
4708         /* Only boolean on/off is supported */
4709         if (cp->param[0] != 0x00 && cp->param[0] != 0x01)
4710                 return mgmt_cmd_status(sk, hdev->id,
4711                                        MGMT_OP_SET_EXP_FEATURE,
4712                                        MGMT_STATUS_INVALID_PARAMS);
4713
4714         val = !!cp->param[0];
4715
4716         if (val) {
4717                 changed = !hci_dev_test_and_set_flag(hdev,
4718                                                      HCI_ENABLE_LL_PRIVACY);
4719                 hci_dev_clear_flag(hdev, HCI_ADVERTISING);
4720
4721                 /* Enable LL privacy + supported settings changed */
4722                 flags = BIT(0) | BIT(1);
4723         } else {
4724                 changed = hci_dev_test_and_clear_flag(hdev,
4725                                                       HCI_ENABLE_LL_PRIVACY);
4726
4727                 /* Disable LL privacy + supported settings changed */
4728                 flags = BIT(1);
4729         }
4730
4731         memcpy(rp.uuid, rpa_resolution_uuid, 16);
4732         rp.flags = cpu_to_le32(flags);
4733
4734         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4735
4736         err = mgmt_cmd_complete(sk, hdev->id,
4737                                 MGMT_OP_SET_EXP_FEATURE, 0,
4738                                 &rp, sizeof(rp));
4739
4740         if (changed)
4741                 exp_ll_privacy_feature_changed(val, hdev, sk);
4742
4743         return err;
4744 }
4745
4746 static int set_quality_report_func(struct sock *sk, struct hci_dev *hdev,
4747                                    struct mgmt_cp_set_exp_feature *cp,
4748                                    u16 data_len)
4749 {
4750         struct mgmt_rp_set_exp_feature rp;
4751         bool val, changed;
4752         int err;
4753
4754         /* Command requires to use a valid controller index */
4755         if (!hdev)
4756                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4757                                        MGMT_OP_SET_EXP_FEATURE,
4758                                        MGMT_STATUS_INVALID_INDEX);
4759
4760         /* Parameters are limited to a single octet */
4761         if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1)
4762                 return mgmt_cmd_status(sk, hdev->id,
4763                                        MGMT_OP_SET_EXP_FEATURE,
4764                                        MGMT_STATUS_INVALID_PARAMS);
4765
4766         /* Only boolean on/off is supported */
4767         if (cp->param[0] != 0x00 && cp->param[0] != 0x01)
4768                 return mgmt_cmd_status(sk, hdev->id,
4769                                        MGMT_OP_SET_EXP_FEATURE,
4770                                        MGMT_STATUS_INVALID_PARAMS);
4771
4772         hci_req_sync_lock(hdev);
4773
4774         val = !!cp->param[0];
4775         changed = (val != hci_dev_test_flag(hdev, HCI_QUALITY_REPORT));
4776
4777         if (!aosp_has_quality_report(hdev) && !hdev->set_quality_report) {
4778                 err = mgmt_cmd_status(sk, hdev->id,
4779                                       MGMT_OP_SET_EXP_FEATURE,
4780                                       MGMT_STATUS_NOT_SUPPORTED);
4781                 goto unlock_quality_report;
4782         }
4783
4784         if (changed) {
4785                 if (hdev->set_quality_report)
4786                         err = hdev->set_quality_report(hdev, val);
4787                 else
4788                         err = aosp_set_quality_report(hdev, val);
4789
4790                 if (err) {
4791                         err = mgmt_cmd_status(sk, hdev->id,
4792                                               MGMT_OP_SET_EXP_FEATURE,
4793                                               MGMT_STATUS_FAILED);
4794                         goto unlock_quality_report;
4795                 }
4796
4797                 if (val)
4798                         hci_dev_set_flag(hdev, HCI_QUALITY_REPORT);
4799                 else
4800                         hci_dev_clear_flag(hdev, HCI_QUALITY_REPORT);
4801         }
4802
4803         bt_dev_dbg(hdev, "quality report enable %d changed %d", val, changed);
4804
4805         memcpy(rp.uuid, quality_report_uuid, 16);
4806         rp.flags = cpu_to_le32(val ? BIT(0) : 0);
4807         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4808
4809         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_EXP_FEATURE, 0,
4810                                 &rp, sizeof(rp));
4811
4812         if (changed)
4813                 exp_feature_changed(hdev, quality_report_uuid, val, sk);
4814
4815 unlock_quality_report:
4816         hci_req_sync_unlock(hdev);
4817         return err;
4818 }
4819
4820 static int set_offload_codec_func(struct sock *sk, struct hci_dev *hdev,
4821                                   struct mgmt_cp_set_exp_feature *cp,
4822                                   u16 data_len)
4823 {
4824         bool val, changed;
4825         int err;
4826         struct mgmt_rp_set_exp_feature rp;
4827
4828         /* Command requires to use a valid controller index */
4829         if (!hdev)
4830                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4831                                        MGMT_OP_SET_EXP_FEATURE,
4832                                        MGMT_STATUS_INVALID_INDEX);
4833
4834         /* Parameters are limited to a single octet */
4835         if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1)
4836                 return mgmt_cmd_status(sk, hdev->id,
4837                                        MGMT_OP_SET_EXP_FEATURE,
4838                                        MGMT_STATUS_INVALID_PARAMS);
4839
4840         /* Only boolean on/off is supported */
4841         if (cp->param[0] != 0x00 && cp->param[0] != 0x01)
4842                 return mgmt_cmd_status(sk, hdev->id,
4843                                        MGMT_OP_SET_EXP_FEATURE,
4844                                        MGMT_STATUS_INVALID_PARAMS);
4845
4846         val = !!cp->param[0];
4847         changed = (val != hci_dev_test_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED));
4848
4849         if (!hdev->get_data_path_id) {
4850                 return mgmt_cmd_status(sk, hdev->id,
4851                                        MGMT_OP_SET_EXP_FEATURE,
4852                                        MGMT_STATUS_NOT_SUPPORTED);
4853         }
4854
4855         if (changed) {
4856                 if (val)
4857                         hci_dev_set_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED);
4858                 else
4859                         hci_dev_clear_flag(hdev, HCI_OFFLOAD_CODECS_ENABLED);
4860         }
4861
4862         bt_dev_info(hdev, "offload codecs enable %d changed %d",
4863                     val, changed);
4864
4865         memcpy(rp.uuid, offload_codecs_uuid, 16);
4866         rp.flags = cpu_to_le32(val ? BIT(0) : 0);
4867         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4868         err = mgmt_cmd_complete(sk, hdev->id,
4869                                 MGMT_OP_SET_EXP_FEATURE, 0,
4870                                 &rp, sizeof(rp));
4871
4872         if (changed)
4873                 exp_feature_changed(hdev, offload_codecs_uuid, val, sk);
4874
4875         return err;
4876 }
4877
4878 static int set_le_simultaneous_roles_func(struct sock *sk, struct hci_dev *hdev,
4879                                           struct mgmt_cp_set_exp_feature *cp,
4880                                           u16 data_len)
4881 {
4882         bool val, changed;
4883         int err;
4884         struct mgmt_rp_set_exp_feature rp;
4885
4886         /* Command requires to use a valid controller index */
4887         if (!hdev)
4888                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4889                                        MGMT_OP_SET_EXP_FEATURE,
4890                                        MGMT_STATUS_INVALID_INDEX);
4891
4892         /* Parameters are limited to a single octet */
4893         if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1)
4894                 return mgmt_cmd_status(sk, hdev->id,
4895                                        MGMT_OP_SET_EXP_FEATURE,
4896                                        MGMT_STATUS_INVALID_PARAMS);
4897
4898         /* Only boolean on/off is supported */
4899         if (cp->param[0] != 0x00 && cp->param[0] != 0x01)
4900                 return mgmt_cmd_status(sk, hdev->id,
4901                                        MGMT_OP_SET_EXP_FEATURE,
4902                                        MGMT_STATUS_INVALID_PARAMS);
4903
4904         val = !!cp->param[0];
4905         changed = (val != hci_dev_test_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES));
4906
4907         if (!hci_dev_le_state_simultaneous(hdev)) {
4908                 return mgmt_cmd_status(sk, hdev->id,
4909                                        MGMT_OP_SET_EXP_FEATURE,
4910                                        MGMT_STATUS_NOT_SUPPORTED);
4911         }
4912
4913         if (changed) {
4914                 if (val)
4915                         hci_dev_set_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES);
4916                 else
4917                         hci_dev_clear_flag(hdev, HCI_LE_SIMULTANEOUS_ROLES);
4918         }
4919
4920         bt_dev_info(hdev, "LE simultaneous roles enable %d changed %d",
4921                     val, changed);
4922
4923         memcpy(rp.uuid, le_simultaneous_roles_uuid, 16);
4924         rp.flags = cpu_to_le32(val ? BIT(0) : 0);
4925         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4926         err = mgmt_cmd_complete(sk, hdev->id,
4927                                 MGMT_OP_SET_EXP_FEATURE, 0,
4928                                 &rp, sizeof(rp));
4929
4930         if (changed)
4931                 exp_feature_changed(hdev, le_simultaneous_roles_uuid, val, sk);
4932
4933         return err;
4934 }
4935
4936 #ifdef CONFIG_BT_LE
4937 static int set_iso_socket_func(struct sock *sk, struct hci_dev *hdev,
4938                                struct mgmt_cp_set_exp_feature *cp, u16 data_len)
4939 {
4940         struct mgmt_rp_set_exp_feature rp;
4941         bool val, changed = false;
4942         int err;
4943
4944         /* Command requires to use the non-controller index */
4945         if (hdev)
4946                 return mgmt_cmd_status(sk, hdev->id,
4947                                        MGMT_OP_SET_EXP_FEATURE,
4948                                        MGMT_STATUS_INVALID_INDEX);
4949
4950         /* Parameters are limited to a single octet */
4951         if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1)
4952                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4953                                        MGMT_OP_SET_EXP_FEATURE,
4954                                        MGMT_STATUS_INVALID_PARAMS);
4955
4956         /* Only boolean on/off is supported */
4957         if (cp->param[0] != 0x00 && cp->param[0] != 0x01)
4958                 return mgmt_cmd_status(sk, MGMT_INDEX_NONE,
4959                                        MGMT_OP_SET_EXP_FEATURE,
4960                                        MGMT_STATUS_INVALID_PARAMS);
4961
4962         val = cp->param[0] ? true : false;
4963         if (val)
4964                 err = iso_init();
4965         else
4966                 err = iso_exit();
4967
4968         if (!err)
4969                 changed = true;
4970
4971         memcpy(rp.uuid, iso_socket_uuid, 16);
4972         rp.flags = cpu_to_le32(val ? BIT(0) : 0);
4973
4974         hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS);
4975
4976         err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE,
4977                                 MGMT_OP_SET_EXP_FEATURE, 0,
4978                                 &rp, sizeof(rp));
4979
4980         if (changed)
4981                 exp_feature_changed(hdev, iso_socket_uuid, val, sk);
4982
4983         return err;
4984 }
4985 #endif
4986
4987 static const struct mgmt_exp_feature {
4988         const u8 *uuid;
4989         int (*set_func)(struct sock *sk, struct hci_dev *hdev,
4990                         struct mgmt_cp_set_exp_feature *cp, u16 data_len);
4991 } exp_features[] = {
4992         EXP_FEAT(ZERO_KEY, set_zero_key_func),
4993 #ifdef CONFIG_BT_FEATURE_DEBUG
4994         EXP_FEAT(debug_uuid, set_debug_func),
4995 #endif
4996         EXP_FEAT(mgmt_mesh_uuid, set_mgmt_mesh_func),
4997         EXP_FEAT(rpa_resolution_uuid, set_rpa_resolution_func),
4998         EXP_FEAT(quality_report_uuid, set_quality_report_func),
4999         EXP_FEAT(offload_codecs_uuid, set_offload_codec_func),
5000         EXP_FEAT(le_simultaneous_roles_uuid, set_le_simultaneous_roles_func),
5001 #ifdef CONFIG_BT_LE
5002         EXP_FEAT(iso_socket_uuid, set_iso_socket_func),
5003 #endif
5004
5005         /* end with a null feature */
5006         EXP_FEAT(NULL, NULL)
5007 };
5008
5009 static int set_exp_feature(struct sock *sk, struct hci_dev *hdev,
5010                            void *data, u16 data_len)
5011 {
5012         struct mgmt_cp_set_exp_feature *cp = data;
5013         size_t i = 0;
5014
5015         bt_dev_dbg(hdev, "sock %p", sk);
5016
5017         for (i = 0; exp_features[i].uuid; i++) {
5018                 if (!memcmp(cp->uuid, exp_features[i].uuid, 16))
5019                         return exp_features[i].set_func(sk, hdev, cp, data_len);
5020         }
5021
5022         return mgmt_cmd_status(sk, hdev ? hdev->id : MGMT_INDEX_NONE,
5023                                MGMT_OP_SET_EXP_FEATURE,
5024                                MGMT_STATUS_NOT_SUPPORTED);
5025 }
5026
5027 static u32 get_params_flags(struct hci_dev *hdev,
5028                             struct hci_conn_params *params)
5029 {
5030         u32 flags = hdev->conn_flags;
5031
5032         /* Devices using RPAs can only be programmed in the acceptlist if
5033          * LL Privacy has been enable otherwise they cannot mark
5034          * HCI_CONN_FLAG_REMOTE_WAKEUP.
5035          */
5036         if ((flags & HCI_CONN_FLAG_REMOTE_WAKEUP) && !use_ll_privacy(hdev) &&
5037             hci_find_irk_by_addr(hdev, &params->addr, params->addr_type))
5038                 flags &= ~HCI_CONN_FLAG_REMOTE_WAKEUP;
5039
5040         return flags;
5041 }
5042
5043 static int get_device_flags(struct sock *sk, struct hci_dev *hdev, void *data,
5044                             u16 data_len)
5045 {
5046         struct mgmt_cp_get_device_flags *cp = data;
5047         struct mgmt_rp_get_device_flags rp;
5048         struct bdaddr_list_with_flags *br_params;
5049         struct hci_conn_params *params;
5050         u32 supported_flags;
5051         u32 current_flags = 0;
5052         u8 status = MGMT_STATUS_INVALID_PARAMS;
5053
5054         bt_dev_dbg(hdev, "Get device flags %pMR (type 0x%x)\n",
5055                    &cp->addr.bdaddr, cp->addr.type);
5056
5057         hci_dev_lock(hdev);
5058
5059         supported_flags = hdev->conn_flags;
5060
5061         memset(&rp, 0, sizeof(rp));
5062
5063         if (cp->addr.type == BDADDR_BREDR) {
5064                 br_params = hci_bdaddr_list_lookup_with_flags(&hdev->accept_list,
5065                                                               &cp->addr.bdaddr,
5066                                                               cp->addr.type);
5067                 if (!br_params)
5068                         goto done;
5069
5070                 current_flags = br_params->flags;
5071         } else {
5072                 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr,
5073                                                 le_addr_type(cp->addr.type));
5074                 if (!params)
5075                         goto done;
5076
5077                 supported_flags = get_params_flags(hdev, params);
5078                 current_flags = params->flags;
5079         }
5080
5081         bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr);
5082         rp.addr.type = cp->addr.type;
5083         rp.supported_flags = cpu_to_le32(supported_flags);
5084         rp.current_flags = cpu_to_le32(current_flags);
5085
5086         status = MGMT_STATUS_SUCCESS;
5087
5088 done:
5089         hci_dev_unlock(hdev);
5090
5091         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_DEVICE_FLAGS, status,
5092                                 &rp, sizeof(rp));
5093 }
5094
5095 static void device_flags_changed(struct sock *sk, struct hci_dev *hdev,
5096                                  bdaddr_t *bdaddr, u8 bdaddr_type,
5097                                  u32 supported_flags, u32 current_flags)
5098 {
5099         struct mgmt_ev_device_flags_changed ev;
5100
5101         bacpy(&ev.addr.bdaddr, bdaddr);
5102         ev.addr.type = bdaddr_type;
5103         ev.supported_flags = cpu_to_le32(supported_flags);
5104         ev.current_flags = cpu_to_le32(current_flags);
5105
5106         mgmt_event(MGMT_EV_DEVICE_FLAGS_CHANGED, hdev, &ev, sizeof(ev), sk);
5107 }
5108
5109 static int set_device_flags(struct sock *sk, struct hci_dev *hdev, void *data,
5110                             u16 len)
5111 {
5112         struct mgmt_cp_set_device_flags *cp = data;
5113         struct bdaddr_list_with_flags *br_params;
5114         struct hci_conn_params *params;
5115         u8 status = MGMT_STATUS_INVALID_PARAMS;
5116         u32 supported_flags;
5117         u32 current_flags = __le32_to_cpu(cp->current_flags);
5118
5119         bt_dev_dbg(hdev, "Set device flags %pMR (type 0x%x) = 0x%x",
5120                    &cp->addr.bdaddr, cp->addr.type, current_flags);
5121
5122         // We should take hci_dev_lock() early, I think.. conn_flags can change
5123         supported_flags = hdev->conn_flags;
5124
5125         if ((supported_flags | current_flags) != supported_flags) {
5126                 bt_dev_warn(hdev, "Bad flag given (0x%x) vs supported (0x%0x)",
5127                             current_flags, supported_flags);
5128                 goto done;
5129         }
5130
5131         hci_dev_lock(hdev);
5132
5133         if (cp->addr.type == BDADDR_BREDR) {
5134                 br_params = hci_bdaddr_list_lookup_with_flags(&hdev->accept_list,
5135                                                               &cp->addr.bdaddr,
5136                                                               cp->addr.type);
5137
5138                 if (br_params) {
5139                         br_params->flags = current_flags;
5140                         status = MGMT_STATUS_SUCCESS;
5141                 } else {
5142                         bt_dev_warn(hdev, "No such BR/EDR device %pMR (0x%x)",
5143                                     &cp->addr.bdaddr, cp->addr.type);
5144                 }
5145
5146                 goto unlock;
5147         }
5148
5149         params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr,
5150                                         le_addr_type(cp->addr.type));
5151         if (!params) {
5152                 bt_dev_warn(hdev, "No such LE device %pMR (0x%x)",
5153                             &cp->addr.bdaddr, le_addr_type(cp->addr.type));
5154                 goto unlock;
5155         }
5156
5157         supported_flags = get_params_flags(hdev, params);
5158
5159         if ((supported_flags | current_flags) != supported_flags) {
5160                 bt_dev_warn(hdev, "Bad flag given (0x%x) vs supported (0x%0x)",
5161                             current_flags, supported_flags);
5162                 goto unlock;
5163         }
5164
5165         WRITE_ONCE(params->flags, current_flags);
5166         status = MGMT_STATUS_SUCCESS;
5167
5168         /* Update passive scan if HCI_CONN_FLAG_DEVICE_PRIVACY
5169          * has been set.
5170          */
5171         if (params->flags & HCI_CONN_FLAG_DEVICE_PRIVACY)
5172                 hci_update_passive_scan(hdev);
5173
5174 unlock:
5175         hci_dev_unlock(hdev);
5176
5177 done:
5178         if (status == MGMT_STATUS_SUCCESS)
5179                 device_flags_changed(sk, hdev, &cp->addr.bdaddr, cp->addr.type,
5180                                      supported_flags, current_flags);
5181
5182         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEVICE_FLAGS, status,
5183                                  &cp->addr, sizeof(cp->addr));
5184 }
5185
5186 static void mgmt_adv_monitor_added(struct sock *sk, struct hci_dev *hdev,
5187                                    u16 handle)
5188 {
5189         struct mgmt_ev_adv_monitor_added ev;
5190
5191         ev.monitor_handle = cpu_to_le16(handle);
5192
5193         mgmt_event(MGMT_EV_ADV_MONITOR_ADDED, hdev, &ev, sizeof(ev), sk);
5194 }
5195
5196 void mgmt_adv_monitor_removed(struct hci_dev *hdev, u16 handle)
5197 {
5198         struct mgmt_ev_adv_monitor_removed ev;
5199         struct mgmt_pending_cmd *cmd;
5200         struct sock *sk_skip = NULL;
5201         struct mgmt_cp_remove_adv_monitor *cp;
5202
5203         cmd = pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev);
5204         if (cmd) {
5205                 cp = cmd->param;
5206
5207                 if (cp->monitor_handle)
5208                         sk_skip = cmd->sk;
5209         }
5210
5211         ev.monitor_handle = cpu_to_le16(handle);
5212
5213         mgmt_event(MGMT_EV_ADV_MONITOR_REMOVED, hdev, &ev, sizeof(ev), sk_skip);
5214 }
5215
5216 static int read_adv_mon_features(struct sock *sk, struct hci_dev *hdev,
5217                                  void *data, u16 len)
5218 {
5219         struct adv_monitor *monitor = NULL;
5220         struct mgmt_rp_read_adv_monitor_features *rp = NULL;
5221         int handle, err;
5222         size_t rp_size = 0;
5223         __u32 supported = 0;
5224         __u32 enabled = 0;
5225         __u16 num_handles = 0;
5226         __u16 handles[HCI_MAX_ADV_MONITOR_NUM_HANDLES];
5227
5228         BT_DBG("request for %s", hdev->name);
5229
5230         hci_dev_lock(hdev);
5231
5232         if (msft_monitor_supported(hdev))
5233                 supported |= MGMT_ADV_MONITOR_FEATURE_MASK_OR_PATTERNS;
5234
5235         idr_for_each_entry(&hdev->adv_monitors_idr, monitor, handle)
5236                 handles[num_handles++] = monitor->handle;
5237
5238         hci_dev_unlock(hdev);
5239
5240         rp_size = sizeof(*rp) + (num_handles * sizeof(u16));
5241         rp = kmalloc(rp_size, GFP_KERNEL);
5242         if (!rp)
5243                 return -ENOMEM;
5244
5245         /* All supported features are currently enabled */
5246         enabled = supported;
5247
5248         rp->supported_features = cpu_to_le32(supported);
5249         rp->enabled_features = cpu_to_le32(enabled);
5250         rp->max_num_handles = cpu_to_le16(HCI_MAX_ADV_MONITOR_NUM_HANDLES);
5251         rp->max_num_patterns = HCI_MAX_ADV_MONITOR_NUM_PATTERNS;
5252         rp->num_handles = cpu_to_le16(num_handles);
5253         if (num_handles)
5254                 memcpy(&rp->handles, &handles, (num_handles * sizeof(u16)));
5255
5256         err = mgmt_cmd_complete(sk, hdev->id,
5257                                 MGMT_OP_READ_ADV_MONITOR_FEATURES,
5258                                 MGMT_STATUS_SUCCESS, rp, rp_size);
5259
5260         kfree(rp);
5261
5262         return err;
5263 }
5264
5265 static void mgmt_add_adv_patterns_monitor_complete(struct hci_dev *hdev,
5266                                                    void *data, int status)
5267 {
5268         struct mgmt_rp_add_adv_patterns_monitor rp;
5269         struct mgmt_pending_cmd *cmd = data;
5270         struct adv_monitor *monitor = cmd->user_data;
5271
5272         hci_dev_lock(hdev);
5273
5274         rp.monitor_handle = cpu_to_le16(monitor->handle);
5275
5276         if (!status) {
5277                 mgmt_adv_monitor_added(cmd->sk, hdev, monitor->handle);
5278                 hdev->adv_monitors_cnt++;
5279                 if (monitor->state == ADV_MONITOR_STATE_NOT_REGISTERED)
5280                         monitor->state = ADV_MONITOR_STATE_REGISTERED;
5281                 hci_update_passive_scan(hdev);
5282         }
5283
5284         mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode,
5285                           mgmt_status(status), &rp, sizeof(rp));
5286         mgmt_pending_remove(cmd);
5287
5288         hci_dev_unlock(hdev);
5289         bt_dev_dbg(hdev, "add monitor %d complete, status %d",
5290                    rp.monitor_handle, status);
5291 }
5292
5293 static int mgmt_add_adv_patterns_monitor_sync(struct hci_dev *hdev, void *data)
5294 {
5295         struct mgmt_pending_cmd *cmd = data;
5296         struct adv_monitor *monitor = cmd->user_data;
5297
5298         return hci_add_adv_monitor(hdev, monitor);
5299 }
5300
5301 static int __add_adv_patterns_monitor(struct sock *sk, struct hci_dev *hdev,
5302                                       struct adv_monitor *m, u8 status,
5303                                       void *data, u16 len, u16 op)
5304 {
5305         struct mgmt_pending_cmd *cmd;
5306         int err;
5307
5308         hci_dev_lock(hdev);
5309
5310         if (status)
5311                 goto unlock;
5312
5313         if (pending_find(MGMT_OP_SET_LE, hdev) ||
5314             pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR, hdev) ||
5315             pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, hdev) ||
5316             pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev)) {
5317                 status = MGMT_STATUS_BUSY;
5318                 goto unlock;
5319         }
5320
5321         cmd = mgmt_pending_add(sk, op, hdev, data, len);
5322         if (!cmd) {
5323                 status = MGMT_STATUS_NO_RESOURCES;
5324                 goto unlock;
5325         }
5326
5327         cmd->user_data = m;
5328         err = hci_cmd_sync_queue(hdev, mgmt_add_adv_patterns_monitor_sync, cmd,
5329                                  mgmt_add_adv_patterns_monitor_complete);
5330         if (err) {
5331                 if (err == -ENOMEM)
5332                         status = MGMT_STATUS_NO_RESOURCES;
5333                 else
5334                         status = MGMT_STATUS_FAILED;
5335
5336                 goto unlock;
5337         }
5338
5339         hci_dev_unlock(hdev);
5340
5341         return 0;
5342
5343 unlock:
5344         hci_free_adv_monitor(hdev, m);
5345         hci_dev_unlock(hdev);
5346         return mgmt_cmd_status(sk, hdev->id, op, status);
5347 }
5348
5349 static void parse_adv_monitor_rssi(struct adv_monitor *m,
5350                                    struct mgmt_adv_rssi_thresholds *rssi)
5351 {
5352         if (rssi) {
5353                 m->rssi.low_threshold = rssi->low_threshold;
5354                 m->rssi.low_threshold_timeout =
5355                     __le16_to_cpu(rssi->low_threshold_timeout);
5356                 m->rssi.high_threshold = rssi->high_threshold;
5357                 m->rssi.high_threshold_timeout =
5358                     __le16_to_cpu(rssi->high_threshold_timeout);
5359                 m->rssi.sampling_period = rssi->sampling_period;
5360         } else {
5361                 /* Default values. These numbers are the least constricting
5362                  * parameters for MSFT API to work, so it behaves as if there
5363                  * are no rssi parameter to consider. May need to be changed
5364                  * if other API are to be supported.
5365                  */
5366                 m->rssi.low_threshold = -127;
5367                 m->rssi.low_threshold_timeout = 60;
5368                 m->rssi.high_threshold = -127;
5369                 m->rssi.high_threshold_timeout = 0;
5370                 m->rssi.sampling_period = 0;
5371         }
5372 }
5373
5374 static u8 parse_adv_monitor_pattern(struct adv_monitor *m, u8 pattern_count,
5375                                     struct mgmt_adv_pattern *patterns)
5376 {
5377         u8 offset = 0, length = 0;
5378         struct adv_pattern *p = NULL;
5379         int i;
5380
5381         for (i = 0; i < pattern_count; i++) {
5382                 offset = patterns[i].offset;
5383                 length = patterns[i].length;
5384                 if (offset >= HCI_MAX_AD_LENGTH ||
5385                     length > HCI_MAX_AD_LENGTH ||
5386                     (offset + length) > HCI_MAX_AD_LENGTH)
5387                         return MGMT_STATUS_INVALID_PARAMS;
5388
5389                 p = kmalloc(sizeof(*p), GFP_KERNEL);
5390                 if (!p)
5391                         return MGMT_STATUS_NO_RESOURCES;
5392
5393                 p->ad_type = patterns[i].ad_type;
5394                 p->offset = patterns[i].offset;
5395                 p->length = patterns[i].length;
5396                 memcpy(p->value, patterns[i].value, p->length);
5397
5398                 INIT_LIST_HEAD(&p->list);
5399                 list_add(&p->list, &m->patterns);
5400         }
5401
5402         return MGMT_STATUS_SUCCESS;
5403 }
5404
5405 static int add_adv_patterns_monitor(struct sock *sk, struct hci_dev *hdev,
5406                                     void *data, u16 len)
5407 {
5408         struct mgmt_cp_add_adv_patterns_monitor *cp = data;
5409         struct adv_monitor *m = NULL;
5410         u8 status = MGMT_STATUS_SUCCESS;
5411         size_t expected_size = sizeof(*cp);
5412
5413         BT_DBG("request for %s", hdev->name);
5414
5415         if (len <= sizeof(*cp)) {
5416                 status = MGMT_STATUS_INVALID_PARAMS;
5417                 goto done;
5418         }
5419
5420         expected_size += cp->pattern_count * sizeof(struct mgmt_adv_pattern);
5421         if (len != expected_size) {
5422                 status = MGMT_STATUS_INVALID_PARAMS;
5423                 goto done;
5424         }
5425
5426         m = kzalloc(sizeof(*m), GFP_KERNEL);
5427         if (!m) {
5428                 status = MGMT_STATUS_NO_RESOURCES;
5429                 goto done;
5430         }
5431
5432         INIT_LIST_HEAD(&m->patterns);
5433
5434         parse_adv_monitor_rssi(m, NULL);
5435         status = parse_adv_monitor_pattern(m, cp->pattern_count, cp->patterns);
5436
5437 done:
5438         return __add_adv_patterns_monitor(sk, hdev, m, status, data, len,
5439                                           MGMT_OP_ADD_ADV_PATTERNS_MONITOR);
5440 }
5441
5442 static int add_adv_patterns_monitor_rssi(struct sock *sk, struct hci_dev *hdev,
5443                                          void *data, u16 len)
5444 {
5445         struct mgmt_cp_add_adv_patterns_monitor_rssi *cp = data;
5446         struct adv_monitor *m = NULL;
5447         u8 status = MGMT_STATUS_SUCCESS;
5448         size_t expected_size = sizeof(*cp);
5449
5450         BT_DBG("request for %s", hdev->name);
5451
5452         if (len <= sizeof(*cp)) {
5453                 status = MGMT_STATUS_INVALID_PARAMS;
5454                 goto done;
5455         }
5456
5457         expected_size += cp->pattern_count * sizeof(struct mgmt_adv_pattern);
5458         if (len != expected_size) {
5459                 status = MGMT_STATUS_INVALID_PARAMS;
5460                 goto done;
5461         }
5462
5463         m = kzalloc(sizeof(*m), GFP_KERNEL);
5464         if (!m) {
5465                 status = MGMT_STATUS_NO_RESOURCES;
5466                 goto done;
5467         }
5468
5469         INIT_LIST_HEAD(&m->patterns);
5470
5471         parse_adv_monitor_rssi(m, &cp->rssi);
5472         status = parse_adv_monitor_pattern(m, cp->pattern_count, cp->patterns);
5473
5474 done:
5475         return __add_adv_patterns_monitor(sk, hdev, m, status, data, len,
5476                                          MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI);
5477 }
5478
5479 static void mgmt_remove_adv_monitor_complete(struct hci_dev *hdev,
5480                                              void *data, int status)
5481 {
5482         struct mgmt_rp_remove_adv_monitor rp;
5483         struct mgmt_pending_cmd *cmd = data;
5484         struct mgmt_cp_remove_adv_monitor *cp = cmd->param;
5485
5486         hci_dev_lock(hdev);
5487
5488         rp.monitor_handle = cp->monitor_handle;
5489
5490         if (!status)
5491                 hci_update_passive_scan(hdev);
5492
5493         mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode,
5494                           mgmt_status(status), &rp, sizeof(rp));
5495         mgmt_pending_remove(cmd);
5496
5497         hci_dev_unlock(hdev);
5498         bt_dev_dbg(hdev, "remove monitor %d complete, status %d",
5499                    rp.monitor_handle, status);
5500 }
5501
5502 static int mgmt_remove_adv_monitor_sync(struct hci_dev *hdev, void *data)
5503 {
5504         struct mgmt_pending_cmd *cmd = data;
5505         struct mgmt_cp_remove_adv_monitor *cp = cmd->param;
5506         u16 handle = __le16_to_cpu(cp->monitor_handle);
5507
5508         if (!handle)
5509                 return hci_remove_all_adv_monitor(hdev);
5510
5511         return hci_remove_single_adv_monitor(hdev, handle);
5512 }
5513
5514 static int remove_adv_monitor(struct sock *sk, struct hci_dev *hdev,
5515                               void *data, u16 len)
5516 {
5517         struct mgmt_pending_cmd *cmd;
5518         int err, status;
5519
5520         hci_dev_lock(hdev);
5521
5522         if (pending_find(MGMT_OP_SET_LE, hdev) ||
5523             pending_find(MGMT_OP_REMOVE_ADV_MONITOR, hdev) ||
5524             pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR, hdev) ||
5525             pending_find(MGMT_OP_ADD_ADV_PATTERNS_MONITOR_RSSI, hdev)) {
5526                 status = MGMT_STATUS_BUSY;
5527                 goto unlock;
5528         }
5529
5530         cmd = mgmt_pending_add(sk, MGMT_OP_REMOVE_ADV_MONITOR, hdev, data, len);
5531         if (!cmd) {
5532                 status = MGMT_STATUS_NO_RESOURCES;
5533                 goto unlock;
5534         }
5535
5536         err = hci_cmd_sync_queue(hdev, mgmt_remove_adv_monitor_sync, cmd,
5537                                  mgmt_remove_adv_monitor_complete);
5538
5539         if (err) {
5540                 mgmt_pending_remove(cmd);
5541
5542                 if (err == -ENOMEM)
5543                         status = MGMT_STATUS_NO_RESOURCES;
5544                 else
5545                         status = MGMT_STATUS_FAILED;
5546
5547                 goto unlock;
5548         }
5549
5550         hci_dev_unlock(hdev);
5551
5552         return 0;
5553
5554 unlock:
5555         hci_dev_unlock(hdev);
5556         return mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADV_MONITOR,
5557                                status);
5558 }
5559
5560 static void read_local_oob_data_complete(struct hci_dev *hdev, void *data, int err)
5561 {
5562         struct mgmt_rp_read_local_oob_data mgmt_rp;
5563         size_t rp_size = sizeof(mgmt_rp);
5564         struct mgmt_pending_cmd *cmd = data;
5565         struct sk_buff *skb = cmd->skb;
5566         u8 status = mgmt_status(err);
5567
5568         if (!status) {
5569                 if (!skb)
5570                         status = MGMT_STATUS_FAILED;
5571                 else if (IS_ERR(skb))
5572                         status = mgmt_status(PTR_ERR(skb));
5573                 else
5574                         status = mgmt_status(skb->data[0]);
5575         }
5576
5577         bt_dev_dbg(hdev, "status %d", status);
5578
5579         if (status) {
5580                 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA, status);
5581                 goto remove;
5582         }
5583
5584         memset(&mgmt_rp, 0, sizeof(mgmt_rp));
5585
5586         if (!bredr_sc_enabled(hdev)) {
5587                 struct hci_rp_read_local_oob_data *rp = (void *) skb->data;
5588
5589                 if (skb->len < sizeof(*rp)) {
5590                         mgmt_cmd_status(cmd->sk, hdev->id,
5591                                         MGMT_OP_READ_LOCAL_OOB_DATA,
5592                                         MGMT_STATUS_FAILED);
5593                         goto remove;
5594                 }
5595
5596                 memcpy(mgmt_rp.hash192, rp->hash, sizeof(rp->hash));
5597                 memcpy(mgmt_rp.rand192, rp->rand, sizeof(rp->rand));
5598
5599                 rp_size -= sizeof(mgmt_rp.hash256) + sizeof(mgmt_rp.rand256);
5600         } else {
5601                 struct hci_rp_read_local_oob_ext_data *rp = (void *) skb->data;
5602
5603                 if (skb->len < sizeof(*rp)) {
5604                         mgmt_cmd_status(cmd->sk, hdev->id,
5605                                         MGMT_OP_READ_LOCAL_OOB_DATA,
5606                                         MGMT_STATUS_FAILED);
5607                         goto remove;
5608                 }
5609
5610                 memcpy(mgmt_rp.hash192, rp->hash192, sizeof(rp->hash192));
5611                 memcpy(mgmt_rp.rand192, rp->rand192, sizeof(rp->rand192));
5612
5613                 memcpy(mgmt_rp.hash256, rp->hash256, sizeof(rp->hash256));
5614                 memcpy(mgmt_rp.rand256, rp->rand256, sizeof(rp->rand256));
5615         }
5616
5617         mgmt_cmd_complete(cmd->sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA,
5618                           MGMT_STATUS_SUCCESS, &mgmt_rp, rp_size);
5619
5620 remove:
5621         if (skb && !IS_ERR(skb))
5622                 kfree_skb(skb);
5623
5624         mgmt_pending_free(cmd);
5625 }
5626
5627 static int read_local_oob_data_sync(struct hci_dev *hdev, void *data)
5628 {
5629         struct mgmt_pending_cmd *cmd = data;
5630
5631         if (bredr_sc_enabled(hdev))
5632                 cmd->skb = hci_read_local_oob_data_sync(hdev, true, cmd->sk);
5633         else
5634                 cmd->skb = hci_read_local_oob_data_sync(hdev, false, cmd->sk);
5635
5636         if (IS_ERR(cmd->skb))
5637                 return PTR_ERR(cmd->skb);
5638         else
5639                 return 0;
5640 }
5641
5642 static int read_local_oob_data(struct sock *sk, struct hci_dev *hdev,
5643                                void *data, u16 data_len)
5644 {
5645         struct mgmt_pending_cmd *cmd;
5646         int err;
5647
5648         bt_dev_dbg(hdev, "sock %p", sk);
5649
5650         hci_dev_lock(hdev);
5651
5652         if (!hdev_is_powered(hdev)) {
5653                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA,
5654                                       MGMT_STATUS_NOT_POWERED);
5655                 goto unlock;
5656         }
5657
5658         if (!lmp_ssp_capable(hdev)) {
5659                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA,
5660                                       MGMT_STATUS_NOT_SUPPORTED);
5661                 goto unlock;
5662         }
5663
5664         cmd = mgmt_pending_new(sk, MGMT_OP_READ_LOCAL_OOB_DATA, hdev, NULL, 0);
5665         if (!cmd)
5666                 err = -ENOMEM;
5667         else
5668                 err = hci_cmd_sync_queue(hdev, read_local_oob_data_sync, cmd,
5669                                          read_local_oob_data_complete);
5670
5671         if (err < 0) {
5672                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_DATA,
5673                                       MGMT_STATUS_FAILED);
5674
5675                 if (cmd)
5676                         mgmt_pending_free(cmd);
5677         }
5678
5679 unlock:
5680         hci_dev_unlock(hdev);
5681         return err;
5682 }
5683
5684 static int add_remote_oob_data(struct sock *sk, struct hci_dev *hdev,
5685                                void *data, u16 len)
5686 {
5687         struct mgmt_addr_info *addr = data;
5688         int err;
5689
5690         bt_dev_dbg(hdev, "sock %p", sk);
5691
5692         if (!bdaddr_type_is_valid(addr->type))
5693                 return mgmt_cmd_complete(sk, hdev->id,
5694                                          MGMT_OP_ADD_REMOTE_OOB_DATA,
5695                                          MGMT_STATUS_INVALID_PARAMS,
5696                                          addr, sizeof(*addr));
5697
5698         hci_dev_lock(hdev);
5699
5700         if (len == MGMT_ADD_REMOTE_OOB_DATA_SIZE) {
5701                 struct mgmt_cp_add_remote_oob_data *cp = data;
5702                 u8 status;
5703
5704                 if (cp->addr.type != BDADDR_BREDR) {
5705                         err = mgmt_cmd_complete(sk, hdev->id,
5706                                                 MGMT_OP_ADD_REMOTE_OOB_DATA,
5707                                                 MGMT_STATUS_INVALID_PARAMS,
5708                                                 &cp->addr, sizeof(cp->addr));
5709                         goto unlock;
5710                 }
5711
5712                 err = hci_add_remote_oob_data(hdev, &cp->addr.bdaddr,
5713                                               cp->addr.type, cp->hash,
5714                                               cp->rand, NULL, NULL);
5715                 if (err < 0)
5716                         status = MGMT_STATUS_FAILED;
5717                 else
5718                         status = MGMT_STATUS_SUCCESS;
5719
5720                 err = mgmt_cmd_complete(sk, hdev->id,
5721                                         MGMT_OP_ADD_REMOTE_OOB_DATA, status,
5722                                         &cp->addr, sizeof(cp->addr));
5723         } else if (len == MGMT_ADD_REMOTE_OOB_EXT_DATA_SIZE) {
5724                 struct mgmt_cp_add_remote_oob_ext_data *cp = data;
5725                 u8 *rand192, *hash192, *rand256, *hash256;
5726                 u8 status;
5727
5728                 if (bdaddr_type_is_le(cp->addr.type)) {
5729                         /* Enforce zero-valued 192-bit parameters as
5730                          * long as legacy SMP OOB isn't implemented.
5731                          */
5732                         if (memcmp(cp->rand192, ZERO_KEY, 16) ||
5733                             memcmp(cp->hash192, ZERO_KEY, 16)) {
5734                                 err = mgmt_cmd_complete(sk, hdev->id,
5735                                                         MGMT_OP_ADD_REMOTE_OOB_DATA,
5736                                                         MGMT_STATUS_INVALID_PARAMS,
5737                                                         addr, sizeof(*addr));
5738                                 goto unlock;
5739                         }
5740
5741                         rand192 = NULL;
5742                         hash192 = NULL;
5743                 } else {
5744                         /* In case one of the P-192 values is set to zero,
5745                          * then just disable OOB data for P-192.
5746                          */
5747                         if (!memcmp(cp->rand192, ZERO_KEY, 16) ||
5748                             !memcmp(cp->hash192, ZERO_KEY, 16)) {
5749                                 rand192 = NULL;
5750                                 hash192 = NULL;
5751                         } else {
5752                                 rand192 = cp->rand192;
5753                                 hash192 = cp->hash192;
5754                         }
5755                 }
5756
5757                 /* In case one of the P-256 values is set to zero, then just
5758                  * disable OOB data for P-256.
5759                  */
5760                 if (!memcmp(cp->rand256, ZERO_KEY, 16) ||
5761                     !memcmp(cp->hash256, ZERO_KEY, 16)) {
5762                         rand256 = NULL;
5763                         hash256 = NULL;
5764                 } else {
5765                         rand256 = cp->rand256;
5766                         hash256 = cp->hash256;
5767                 }
5768
5769                 err = hci_add_remote_oob_data(hdev, &cp->addr.bdaddr,
5770                                               cp->addr.type, hash192, rand192,
5771                                               hash256, rand256);
5772                 if (err < 0)
5773                         status = MGMT_STATUS_FAILED;
5774                 else
5775                         status = MGMT_STATUS_SUCCESS;
5776
5777                 err = mgmt_cmd_complete(sk, hdev->id,
5778                                         MGMT_OP_ADD_REMOTE_OOB_DATA,
5779                                         status, &cp->addr, sizeof(cp->addr));
5780         } else {
5781                 bt_dev_err(hdev, "add_remote_oob_data: invalid len of %u bytes",
5782                            len);
5783                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_REMOTE_OOB_DATA,
5784                                       MGMT_STATUS_INVALID_PARAMS);
5785         }
5786
5787 unlock:
5788         hci_dev_unlock(hdev);
5789         return err;
5790 }
5791
5792 static int remove_remote_oob_data(struct sock *sk, struct hci_dev *hdev,
5793                                   void *data, u16 len)
5794 {
5795         struct mgmt_cp_remove_remote_oob_data *cp = data;
5796         u8 status;
5797         int err;
5798
5799         bt_dev_dbg(hdev, "sock %p", sk);
5800
5801         if (cp->addr.type != BDADDR_BREDR)
5802                 return mgmt_cmd_complete(sk, hdev->id,
5803                                          MGMT_OP_REMOVE_REMOTE_OOB_DATA,
5804                                          MGMT_STATUS_INVALID_PARAMS,
5805                                          &cp->addr, sizeof(cp->addr));
5806
5807         hci_dev_lock(hdev);
5808
5809         if (!bacmp(&cp->addr.bdaddr, BDADDR_ANY)) {
5810                 hci_remote_oob_data_clear(hdev);
5811                 status = MGMT_STATUS_SUCCESS;
5812                 goto done;
5813         }
5814
5815         err = hci_remove_remote_oob_data(hdev, &cp->addr.bdaddr, cp->addr.type);
5816         if (err < 0)
5817                 status = MGMT_STATUS_INVALID_PARAMS;
5818         else
5819                 status = MGMT_STATUS_SUCCESS;
5820
5821 done:
5822         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_REMOVE_REMOTE_OOB_DATA,
5823                                 status, &cp->addr, sizeof(cp->addr));
5824
5825         hci_dev_unlock(hdev);
5826         return err;
5827 }
5828
5829 void mgmt_start_discovery_complete(struct hci_dev *hdev, u8 status)
5830 {
5831         struct mgmt_pending_cmd *cmd;
5832
5833         bt_dev_dbg(hdev, "status %u", status);
5834
5835         hci_dev_lock(hdev);
5836
5837         cmd = pending_find(MGMT_OP_START_DISCOVERY, hdev);
5838         if (!cmd)
5839                 cmd = pending_find(MGMT_OP_START_SERVICE_DISCOVERY, hdev);
5840
5841         if (!cmd)
5842                 cmd = pending_find(MGMT_OP_START_LIMITED_DISCOVERY, hdev);
5843
5844         if (cmd) {
5845                 cmd->cmd_complete(cmd, mgmt_status(status));
5846                 mgmt_pending_remove(cmd);
5847         }
5848
5849         hci_dev_unlock(hdev);
5850 }
5851
5852 static bool discovery_type_is_valid(struct hci_dev *hdev, uint8_t type,
5853                                     uint8_t *mgmt_status)
5854 {
5855         switch (type) {
5856         case DISCOV_TYPE_LE:
5857                 *mgmt_status = mgmt_le_support(hdev);
5858                 if (*mgmt_status)
5859                         return false;
5860                 break;
5861         case DISCOV_TYPE_INTERLEAVED:
5862                 *mgmt_status = mgmt_le_support(hdev);
5863                 if (*mgmt_status)
5864                         return false;
5865                 fallthrough;
5866         case DISCOV_TYPE_BREDR:
5867                 *mgmt_status = mgmt_bredr_support(hdev);
5868                 if (*mgmt_status)
5869                         return false;
5870                 break;
5871         default:
5872                 *mgmt_status = MGMT_STATUS_INVALID_PARAMS;
5873                 return false;
5874         }
5875
5876         return true;
5877 }
5878
5879 static void start_discovery_complete(struct hci_dev *hdev, void *data, int err)
5880 {
5881         struct mgmt_pending_cmd *cmd = data;
5882
5883         if (cmd != pending_find(MGMT_OP_START_DISCOVERY, hdev) &&
5884             cmd != pending_find(MGMT_OP_START_LIMITED_DISCOVERY, hdev) &&
5885             cmd != pending_find(MGMT_OP_START_SERVICE_DISCOVERY, hdev))
5886                 return;
5887
5888         bt_dev_dbg(hdev, "err %d", err);
5889
5890         mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, mgmt_status(err),
5891                           cmd->param, 1);
5892         mgmt_pending_remove(cmd);
5893
5894         hci_discovery_set_state(hdev, err ? DISCOVERY_STOPPED:
5895                                 DISCOVERY_FINDING);
5896 }
5897
5898 static int start_discovery_sync(struct hci_dev *hdev, void *data)
5899 {
5900         return hci_start_discovery_sync(hdev);
5901 }
5902
5903 static int start_discovery_internal(struct sock *sk, struct hci_dev *hdev,
5904                                     u16 op, void *data, u16 len)
5905 {
5906         struct mgmt_cp_start_discovery *cp = data;
5907         struct mgmt_pending_cmd *cmd;
5908         u8 status;
5909         int err;
5910
5911         bt_dev_dbg(hdev, "sock %p", sk);
5912
5913         hci_dev_lock(hdev);
5914
5915         if (!hdev_is_powered(hdev)) {
5916                 err = mgmt_cmd_complete(sk, hdev->id, op,
5917                                         MGMT_STATUS_NOT_POWERED,
5918                                         &cp->type, sizeof(cp->type));
5919                 goto failed;
5920         }
5921
5922         if (hdev->discovery.state != DISCOVERY_STOPPED ||
5923             hci_dev_test_flag(hdev, HCI_PERIODIC_INQ)) {
5924                 err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY,
5925                                         &cp->type, sizeof(cp->type));
5926                 goto failed;
5927         }
5928
5929         if (!discovery_type_is_valid(hdev, cp->type, &status)) {
5930                 err = mgmt_cmd_complete(sk, hdev->id, op, status,
5931                                         &cp->type, sizeof(cp->type));
5932                 goto failed;
5933         }
5934
5935         /* Can't start discovery when it is paused */
5936         if (hdev->discovery_paused) {
5937                 err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY,
5938                                         &cp->type, sizeof(cp->type));
5939                 goto failed;
5940         }
5941
5942         /* Clear the discovery filter first to free any previously
5943          * allocated memory for the UUID list.
5944          */
5945         hci_discovery_filter_clear(hdev);
5946
5947         hdev->discovery.type = cp->type;
5948         hdev->discovery.report_invalid_rssi = false;
5949         if (op == MGMT_OP_START_LIMITED_DISCOVERY)
5950                 hdev->discovery.limited = true;
5951         else
5952                 hdev->discovery.limited = false;
5953
5954         cmd = mgmt_pending_add(sk, op, hdev, data, len);
5955         if (!cmd) {
5956                 err = -ENOMEM;
5957                 goto failed;
5958         }
5959
5960         err = hci_cmd_sync_queue(hdev, start_discovery_sync, cmd,
5961                                  start_discovery_complete);
5962         if (err < 0) {
5963                 mgmt_pending_remove(cmd);
5964                 goto failed;
5965         }
5966
5967         hci_discovery_set_state(hdev, DISCOVERY_STARTING);
5968
5969 failed:
5970         hci_dev_unlock(hdev);
5971         return err;
5972 }
5973
5974 static int start_discovery(struct sock *sk, struct hci_dev *hdev,
5975                            void *data, u16 len)
5976 {
5977         return start_discovery_internal(sk, hdev, MGMT_OP_START_DISCOVERY,
5978                                         data, len);
5979 }
5980
5981 static int start_limited_discovery(struct sock *sk, struct hci_dev *hdev,
5982                                    void *data, u16 len)
5983 {
5984         return start_discovery_internal(sk, hdev,
5985                                         MGMT_OP_START_LIMITED_DISCOVERY,
5986                                         data, len);
5987 }
5988
5989 static int start_service_discovery(struct sock *sk, struct hci_dev *hdev,
5990                                    void *data, u16 len)
5991 {
5992         struct mgmt_cp_start_service_discovery *cp = data;
5993         struct mgmt_pending_cmd *cmd;
5994         const u16 max_uuid_count = ((U16_MAX - sizeof(*cp)) / 16);
5995         u16 uuid_count, expected_len;
5996         u8 status;
5997         int err;
5998
5999         bt_dev_dbg(hdev, "sock %p", sk);
6000
6001         hci_dev_lock(hdev);
6002
6003         if (!hdev_is_powered(hdev)) {
6004                 err = mgmt_cmd_complete(sk, hdev->id,
6005                                         MGMT_OP_START_SERVICE_DISCOVERY,
6006                                         MGMT_STATUS_NOT_POWERED,
6007                                         &cp->type, sizeof(cp->type));
6008                 goto failed;
6009         }
6010
6011         if (hdev->discovery.state != DISCOVERY_STOPPED ||
6012             hci_dev_test_flag(hdev, HCI_PERIODIC_INQ)) {
6013                 err = mgmt_cmd_complete(sk, hdev->id,
6014                                         MGMT_OP_START_SERVICE_DISCOVERY,
6015                                         MGMT_STATUS_BUSY, &cp->type,
6016                                         sizeof(cp->type));
6017                 goto failed;
6018         }
6019
6020         if (hdev->discovery_paused) {
6021                 err = mgmt_cmd_complete(sk, hdev->id,
6022                                         MGMT_OP_START_SERVICE_DISCOVERY,
6023                                         MGMT_STATUS_BUSY, &cp->type,
6024                                         sizeof(cp->type));
6025                 goto failed;
6026         }
6027
6028         uuid_count = __le16_to_cpu(cp->uuid_count);
6029         if (uuid_count > max_uuid_count) {
6030                 bt_dev_err(hdev, "service_discovery: too big uuid_count value %u",
6031                            uuid_count);
6032                 err = mgmt_cmd_complete(sk, hdev->id,
6033                                         MGMT_OP_START_SERVICE_DISCOVERY,
6034                                         MGMT_STATUS_INVALID_PARAMS, &cp->type,
6035                                         sizeof(cp->type));
6036                 goto failed;
6037         }
6038
6039         expected_len = sizeof(*cp) + uuid_count * 16;
6040         if (expected_len != len) {
6041                 bt_dev_err(hdev, "service_discovery: expected %u bytes, got %u bytes",
6042                            expected_len, len);
6043                 err = mgmt_cmd_complete(sk, hdev->id,
6044                                         MGMT_OP_START_SERVICE_DISCOVERY,
6045                                         MGMT_STATUS_INVALID_PARAMS, &cp->type,
6046                                         sizeof(cp->type));
6047                 goto failed;
6048         }
6049
6050         if (!discovery_type_is_valid(hdev, cp->type, &status)) {
6051                 err = mgmt_cmd_complete(sk, hdev->id,
6052                                         MGMT_OP_START_SERVICE_DISCOVERY,
6053                                         status, &cp->type, sizeof(cp->type));
6054                 goto failed;
6055         }
6056
6057         cmd = mgmt_pending_add(sk, MGMT_OP_START_SERVICE_DISCOVERY,
6058                                hdev, data, len);
6059         if (!cmd) {
6060                 err = -ENOMEM;
6061                 goto failed;
6062         }
6063
6064         /* Clear the discovery filter first to free any previously
6065          * allocated memory for the UUID list.
6066          */
6067         hci_discovery_filter_clear(hdev);
6068
6069         hdev->discovery.result_filtering = true;
6070         hdev->discovery.type = cp->type;
6071         hdev->discovery.rssi = cp->rssi;
6072         hdev->discovery.uuid_count = uuid_count;
6073
6074         if (uuid_count > 0) {
6075                 hdev->discovery.uuids = kmemdup(cp->uuids, uuid_count * 16,
6076                                                 GFP_KERNEL);
6077                 if (!hdev->discovery.uuids) {
6078                         err = mgmt_cmd_complete(sk, hdev->id,
6079                                                 MGMT_OP_START_SERVICE_DISCOVERY,
6080                                                 MGMT_STATUS_FAILED,
6081                                                 &cp->type, sizeof(cp->type));
6082                         mgmt_pending_remove(cmd);
6083                         goto failed;
6084                 }
6085         }
6086
6087         err = hci_cmd_sync_queue(hdev, start_discovery_sync, cmd,
6088                                  start_discovery_complete);
6089         if (err < 0) {
6090                 mgmt_pending_remove(cmd);
6091                 goto failed;
6092         }
6093
6094         hci_discovery_set_state(hdev, DISCOVERY_STARTING);
6095
6096 failed:
6097         hci_dev_unlock(hdev);
6098         return err;
6099 }
6100
6101 void mgmt_stop_discovery_complete(struct hci_dev *hdev, u8 status)
6102 {
6103         struct mgmt_pending_cmd *cmd;
6104
6105         bt_dev_dbg(hdev, "status %u", status);
6106
6107         hci_dev_lock(hdev);
6108
6109         cmd = pending_find(MGMT_OP_STOP_DISCOVERY, hdev);
6110         if (cmd) {
6111                 cmd->cmd_complete(cmd, mgmt_status(status));
6112                 mgmt_pending_remove(cmd);
6113         }
6114
6115         hci_dev_unlock(hdev);
6116 }
6117
6118 static void stop_discovery_complete(struct hci_dev *hdev, void *data, int err)
6119 {
6120         struct mgmt_pending_cmd *cmd = data;
6121
6122         if (cmd != pending_find(MGMT_OP_STOP_DISCOVERY, hdev))
6123                 return;
6124
6125         bt_dev_dbg(hdev, "err %d", err);
6126
6127         mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, mgmt_status(err),
6128                           cmd->param, 1);
6129         mgmt_pending_remove(cmd);
6130
6131         if (!err)
6132                 hci_discovery_set_state(hdev, DISCOVERY_STOPPED);
6133 }
6134
6135 static int stop_discovery_sync(struct hci_dev *hdev, void *data)
6136 {
6137         return hci_stop_discovery_sync(hdev);
6138 }
6139
6140 static int stop_discovery(struct sock *sk, struct hci_dev *hdev, void *data,
6141                           u16 len)
6142 {
6143         struct mgmt_cp_stop_discovery *mgmt_cp = data;
6144         struct mgmt_pending_cmd *cmd;
6145         int err;
6146
6147         bt_dev_dbg(hdev, "sock %p", sk);
6148
6149         hci_dev_lock(hdev);
6150
6151         if (!hci_discovery_active(hdev)) {
6152                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_STOP_DISCOVERY,
6153                                         MGMT_STATUS_REJECTED, &mgmt_cp->type,
6154                                         sizeof(mgmt_cp->type));
6155                 goto unlock;
6156         }
6157
6158         if (hdev->discovery.type != mgmt_cp->type) {
6159                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_STOP_DISCOVERY,
6160                                         MGMT_STATUS_INVALID_PARAMS,
6161                                         &mgmt_cp->type, sizeof(mgmt_cp->type));
6162                 goto unlock;
6163         }
6164
6165         cmd = mgmt_pending_add(sk, MGMT_OP_STOP_DISCOVERY, hdev, data, len);
6166         if (!cmd) {
6167                 err = -ENOMEM;
6168                 goto unlock;
6169         }
6170
6171         err = hci_cmd_sync_queue(hdev, stop_discovery_sync, cmd,
6172                                  stop_discovery_complete);
6173         if (err < 0) {
6174                 mgmt_pending_remove(cmd);
6175                 goto unlock;
6176         }
6177
6178         hci_discovery_set_state(hdev, DISCOVERY_STOPPING);
6179
6180 unlock:
6181         hci_dev_unlock(hdev);
6182         return err;
6183 }
6184
6185 static int confirm_name(struct sock *sk, struct hci_dev *hdev, void *data,
6186                         u16 len)
6187 {
6188         struct mgmt_cp_confirm_name *cp = data;
6189         struct inquiry_entry *e;
6190         int err;
6191
6192         bt_dev_dbg(hdev, "sock %p", sk);
6193
6194         hci_dev_lock(hdev);
6195
6196         if (!hci_discovery_active(hdev)) {
6197                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME,
6198                                         MGMT_STATUS_FAILED, &cp->addr,
6199                                         sizeof(cp->addr));
6200                 goto failed;
6201         }
6202
6203         e = hci_inquiry_cache_lookup_unknown(hdev, &cp->addr.bdaddr);
6204         if (!e) {
6205                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME,
6206                                         MGMT_STATUS_INVALID_PARAMS, &cp->addr,
6207                                         sizeof(cp->addr));
6208                 goto failed;
6209         }
6210
6211         if (cp->name_known) {
6212                 e->name_state = NAME_KNOWN;
6213                 list_del(&e->list);
6214         } else {
6215                 e->name_state = NAME_NEEDED;
6216                 hci_inquiry_cache_update_resolve(hdev, e);
6217         }
6218
6219         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_CONFIRM_NAME, 0,
6220                                 &cp->addr, sizeof(cp->addr));
6221
6222 failed:
6223         hci_dev_unlock(hdev);
6224         return err;
6225 }
6226
6227 static int block_device(struct sock *sk, struct hci_dev *hdev, void *data,
6228                         u16 len)
6229 {
6230         struct mgmt_cp_block_device *cp = data;
6231         u8 status;
6232         int err;
6233
6234         bt_dev_dbg(hdev, "sock %p", sk);
6235
6236         if (!bdaddr_type_is_valid(cp->addr.type))
6237                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_BLOCK_DEVICE,
6238                                          MGMT_STATUS_INVALID_PARAMS,
6239                                          &cp->addr, sizeof(cp->addr));
6240
6241         hci_dev_lock(hdev);
6242
6243         err = hci_bdaddr_list_add(&hdev->reject_list, &cp->addr.bdaddr,
6244                                   cp->addr.type);
6245         if (err < 0) {
6246                 status = MGMT_STATUS_FAILED;
6247                 goto done;
6248         }
6249
6250         mgmt_event(MGMT_EV_DEVICE_BLOCKED, hdev, &cp->addr, sizeof(cp->addr),
6251                    sk);
6252         status = MGMT_STATUS_SUCCESS;
6253
6254 done:
6255         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_BLOCK_DEVICE, status,
6256                                 &cp->addr, sizeof(cp->addr));
6257
6258         hci_dev_unlock(hdev);
6259
6260         return err;
6261 }
6262
6263 static int unblock_device(struct sock *sk, struct hci_dev *hdev, void *data,
6264                           u16 len)
6265 {
6266         struct mgmt_cp_unblock_device *cp = data;
6267         u8 status;
6268         int err;
6269
6270         bt_dev_dbg(hdev, "sock %p", sk);
6271
6272         if (!bdaddr_type_is_valid(cp->addr.type))
6273                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNBLOCK_DEVICE,
6274                                          MGMT_STATUS_INVALID_PARAMS,
6275                                          &cp->addr, sizeof(cp->addr));
6276
6277         hci_dev_lock(hdev);
6278
6279         err = hci_bdaddr_list_del(&hdev->reject_list, &cp->addr.bdaddr,
6280                                   cp->addr.type);
6281         if (err < 0) {
6282                 status = MGMT_STATUS_INVALID_PARAMS;
6283                 goto done;
6284         }
6285
6286         mgmt_event(MGMT_EV_DEVICE_UNBLOCKED, hdev, &cp->addr, sizeof(cp->addr),
6287                    sk);
6288         status = MGMT_STATUS_SUCCESS;
6289
6290 done:
6291         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_UNBLOCK_DEVICE, status,
6292                                 &cp->addr, sizeof(cp->addr));
6293
6294         hci_dev_unlock(hdev);
6295
6296         return err;
6297 }
6298
6299 static int set_device_id_sync(struct hci_dev *hdev, void *data)
6300 {
6301         return hci_update_eir_sync(hdev);
6302 }
6303
6304 static int set_device_id(struct sock *sk, struct hci_dev *hdev, void *data,
6305                          u16 len)
6306 {
6307         struct mgmt_cp_set_device_id *cp = data;
6308         int err;
6309         __u16 source;
6310
6311         bt_dev_dbg(hdev, "sock %p", sk);
6312
6313         source = __le16_to_cpu(cp->source);
6314
6315         if (source > 0x0002)
6316                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEVICE_ID,
6317                                        MGMT_STATUS_INVALID_PARAMS);
6318
6319         hci_dev_lock(hdev);
6320
6321         hdev->devid_source = source;
6322         hdev->devid_vendor = __le16_to_cpu(cp->vendor);
6323         hdev->devid_product = __le16_to_cpu(cp->product);
6324         hdev->devid_version = __le16_to_cpu(cp->version);
6325
6326         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_DEVICE_ID, 0,
6327                                 NULL, 0);
6328
6329         hci_cmd_sync_queue(hdev, set_device_id_sync, NULL, NULL);
6330
6331         hci_dev_unlock(hdev);
6332
6333         return err;
6334 }
6335
6336 static void enable_advertising_instance(struct hci_dev *hdev, int err)
6337 {
6338         if (err)
6339                 bt_dev_err(hdev, "failed to re-configure advertising %d", err);
6340         else
6341                 bt_dev_dbg(hdev, "status %d", err);
6342 }
6343
6344 static void set_advertising_complete(struct hci_dev *hdev, void *data, int err)
6345 {
6346         struct cmd_lookup match = { NULL, hdev };
6347         u8 instance;
6348         struct adv_info *adv_instance;
6349         u8 status = mgmt_status(err);
6350
6351         if (status) {
6352                 mgmt_pending_foreach(MGMT_OP_SET_ADVERTISING, hdev,
6353                                      cmd_status_rsp, &status);
6354                 return;
6355         }
6356
6357         if (hci_dev_test_flag(hdev, HCI_LE_ADV))
6358                 hci_dev_set_flag(hdev, HCI_ADVERTISING);
6359         else
6360                 hci_dev_clear_flag(hdev, HCI_ADVERTISING);
6361
6362         mgmt_pending_foreach(MGMT_OP_SET_ADVERTISING, hdev, settings_rsp,
6363                              &match);
6364
6365         new_settings(hdev, match.sk);
6366
6367         if (match.sk)
6368                 sock_put(match.sk);
6369
6370         /* If "Set Advertising" was just disabled and instance advertising was
6371          * set up earlier, then re-enable multi-instance advertising.
6372          */
6373         if (hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
6374             list_empty(&hdev->adv_instances))
6375                 return;
6376
6377         instance = hdev->cur_adv_instance;
6378         if (!instance) {
6379                 adv_instance = list_first_entry_or_null(&hdev->adv_instances,
6380                                                         struct adv_info, list);
6381                 if (!adv_instance)
6382                         return;
6383
6384                 instance = adv_instance->instance;
6385         }
6386
6387         err = hci_schedule_adv_instance_sync(hdev, instance, true);
6388
6389         enable_advertising_instance(hdev, err);
6390 }
6391
6392 static int set_adv_sync(struct hci_dev *hdev, void *data)
6393 {
6394         struct mgmt_pending_cmd *cmd = data;
6395         struct mgmt_mode *cp = cmd->param;
6396         u8 val = !!cp->val;
6397
6398         if (cp->val == 0x02)
6399                 hci_dev_set_flag(hdev, HCI_ADVERTISING_CONNECTABLE);
6400         else
6401                 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE);
6402
6403         cancel_adv_timeout(hdev);
6404
6405         if (val) {
6406                 /* Switch to instance "0" for the Set Advertising setting.
6407                  * We cannot use update_[adv|scan_rsp]_data() here as the
6408                  * HCI_ADVERTISING flag is not yet set.
6409                  */
6410                 hdev->cur_adv_instance = 0x00;
6411
6412                 if (ext_adv_capable(hdev)) {
6413                         hci_start_ext_adv_sync(hdev, 0x00);
6414                 } else {
6415                         hci_update_adv_data_sync(hdev, 0x00);
6416                         hci_update_scan_rsp_data_sync(hdev, 0x00);
6417                         hci_enable_advertising_sync(hdev);
6418                 }
6419         } else {
6420                 hci_disable_advertising_sync(hdev);
6421         }
6422
6423         return 0;
6424 }
6425
6426 static int set_advertising(struct sock *sk, struct hci_dev *hdev, void *data,
6427                            u16 len)
6428 {
6429         struct mgmt_mode *cp = data;
6430         struct mgmt_pending_cmd *cmd;
6431         u8 val, status;
6432         int err;
6433
6434         bt_dev_dbg(hdev, "sock %p", sk);
6435
6436         status = mgmt_le_support(hdev);
6437         if (status)
6438                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING,
6439                                        status);
6440
6441         if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02)
6442                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING,
6443                                        MGMT_STATUS_INVALID_PARAMS);
6444
6445         if (hdev->advertising_paused)
6446                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING,
6447                                        MGMT_STATUS_BUSY);
6448
6449         hci_dev_lock(hdev);
6450
6451         val = !!cp->val;
6452
6453         /* The following conditions are ones which mean that we should
6454          * not do any HCI communication but directly send a mgmt
6455          * response to user space (after toggling the flag if
6456          * necessary).
6457          */
6458         if (!hdev_is_powered(hdev) ||
6459             (val == hci_dev_test_flag(hdev, HCI_ADVERTISING) &&
6460              (cp->val == 0x02) == hci_dev_test_flag(hdev, HCI_ADVERTISING_CONNECTABLE)) ||
6461             hci_dev_test_flag(hdev, HCI_MESH) ||
6462             hci_conn_num(hdev, LE_LINK) > 0 ||
6463             (hci_dev_test_flag(hdev, HCI_LE_SCAN) &&
6464              hdev->le_scan_type == LE_SCAN_ACTIVE)) {
6465                 bool changed;
6466
6467                 if (cp->val) {
6468                         hdev->cur_adv_instance = 0x00;
6469                         changed = !hci_dev_test_and_set_flag(hdev, HCI_ADVERTISING);
6470                         if (cp->val == 0x02)
6471                                 hci_dev_set_flag(hdev, HCI_ADVERTISING_CONNECTABLE);
6472                         else
6473                                 hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE);
6474                 } else {
6475                         changed = hci_dev_test_and_clear_flag(hdev, HCI_ADVERTISING);
6476                         hci_dev_clear_flag(hdev, HCI_ADVERTISING_CONNECTABLE);
6477                 }
6478
6479                 err = send_settings_rsp(sk, MGMT_OP_SET_ADVERTISING, hdev);
6480                 if (err < 0)
6481                         goto unlock;
6482
6483                 if (changed)
6484                         err = new_settings(hdev, sk);
6485
6486                 goto unlock;
6487         }
6488
6489         if (pending_find(MGMT_OP_SET_ADVERTISING, hdev) ||
6490             pending_find(MGMT_OP_SET_LE, hdev)) {
6491                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_ADVERTISING,
6492                                       MGMT_STATUS_BUSY);
6493                 goto unlock;
6494         }
6495
6496         cmd = mgmt_pending_add(sk, MGMT_OP_SET_ADVERTISING, hdev, data, len);
6497         if (!cmd)
6498                 err = -ENOMEM;
6499         else
6500                 err = hci_cmd_sync_queue(hdev, set_adv_sync, cmd,
6501                                          set_advertising_complete);
6502
6503         if (err < 0 && cmd)
6504                 mgmt_pending_remove(cmd);
6505
6506 unlock:
6507         hci_dev_unlock(hdev);
6508         return err;
6509 }
6510
6511 static int set_static_address(struct sock *sk, struct hci_dev *hdev,
6512                               void *data, u16 len)
6513 {
6514         struct mgmt_cp_set_static_address *cp = data;
6515         int err;
6516
6517         bt_dev_dbg(hdev, "sock %p", sk);
6518
6519         if (!lmp_le_capable(hdev))
6520                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_STATIC_ADDRESS,
6521                                        MGMT_STATUS_NOT_SUPPORTED);
6522
6523         if (hdev_is_powered(hdev))
6524                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_STATIC_ADDRESS,
6525                                        MGMT_STATUS_REJECTED);
6526
6527         if (bacmp(&cp->bdaddr, BDADDR_ANY)) {
6528                 if (!bacmp(&cp->bdaddr, BDADDR_NONE))
6529                         return mgmt_cmd_status(sk, hdev->id,
6530                                                MGMT_OP_SET_STATIC_ADDRESS,
6531                                                MGMT_STATUS_INVALID_PARAMS);
6532
6533                 /* Two most significant bits shall be set */
6534                 if ((cp->bdaddr.b[5] & 0xc0) != 0xc0)
6535                         return mgmt_cmd_status(sk, hdev->id,
6536                                                MGMT_OP_SET_STATIC_ADDRESS,
6537                                                MGMT_STATUS_INVALID_PARAMS);
6538         }
6539
6540         hci_dev_lock(hdev);
6541
6542         bacpy(&hdev->static_addr, &cp->bdaddr);
6543
6544         err = send_settings_rsp(sk, MGMT_OP_SET_STATIC_ADDRESS, hdev);
6545         if (err < 0)
6546                 goto unlock;
6547
6548         err = new_settings(hdev, sk);
6549
6550 unlock:
6551         hci_dev_unlock(hdev);
6552         return err;
6553 }
6554
6555 static int set_scan_params(struct sock *sk, struct hci_dev *hdev,
6556                            void *data, u16 len)
6557 {
6558         struct mgmt_cp_set_scan_params *cp = data;
6559         __u16 interval, window;
6560         int err;
6561
6562         bt_dev_dbg(hdev, "sock %p", sk);
6563
6564         if (!lmp_le_capable(hdev))
6565                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS,
6566                                        MGMT_STATUS_NOT_SUPPORTED);
6567
6568         interval = __le16_to_cpu(cp->interval);
6569
6570         if (interval < 0x0004 || interval > 0x4000)
6571                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS,
6572                                        MGMT_STATUS_INVALID_PARAMS);
6573
6574         window = __le16_to_cpu(cp->window);
6575
6576         if (window < 0x0004 || window > 0x4000)
6577                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS,
6578                                        MGMT_STATUS_INVALID_PARAMS);
6579
6580         if (window > interval)
6581                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS,
6582                                        MGMT_STATUS_INVALID_PARAMS);
6583
6584         hci_dev_lock(hdev);
6585
6586         hdev->le_scan_interval = interval;
6587         hdev->le_scan_window = window;
6588
6589         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_SET_SCAN_PARAMS, 0,
6590                                 NULL, 0);
6591
6592         /* If background scan is running, restart it so new parameters are
6593          * loaded.
6594          */
6595         if (hci_dev_test_flag(hdev, HCI_LE_SCAN) &&
6596             hdev->discovery.state == DISCOVERY_STOPPED)
6597                 hci_update_passive_scan(hdev);
6598
6599         hci_dev_unlock(hdev);
6600
6601         return err;
6602 }
6603
6604 static void fast_connectable_complete(struct hci_dev *hdev, void *data, int err)
6605 {
6606         struct mgmt_pending_cmd *cmd = data;
6607
6608         bt_dev_dbg(hdev, "err %d", err);
6609
6610         if (err) {
6611                 mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_FAST_CONNECTABLE,
6612                                 mgmt_status(err));
6613         } else {
6614                 struct mgmt_mode *cp = cmd->param;
6615
6616                 if (cp->val)
6617                         hci_dev_set_flag(hdev, HCI_FAST_CONNECTABLE);
6618                 else
6619                         hci_dev_clear_flag(hdev, HCI_FAST_CONNECTABLE);
6620
6621                 send_settings_rsp(cmd->sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev);
6622                 new_settings(hdev, cmd->sk);
6623         }
6624
6625         mgmt_pending_free(cmd);
6626 }
6627
6628 static int write_fast_connectable_sync(struct hci_dev *hdev, void *data)
6629 {
6630         struct mgmt_pending_cmd *cmd = data;
6631         struct mgmt_mode *cp = cmd->param;
6632
6633         return hci_write_fast_connectable_sync(hdev, cp->val);
6634 }
6635
6636 static int set_fast_connectable(struct sock *sk, struct hci_dev *hdev,
6637                                 void *data, u16 len)
6638 {
6639         struct mgmt_mode *cp = data;
6640         struct mgmt_pending_cmd *cmd;
6641         int err;
6642
6643         bt_dev_dbg(hdev, "sock %p", sk);
6644
6645         if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) ||
6646             hdev->hci_ver < BLUETOOTH_VER_1_2)
6647                 return mgmt_cmd_status(sk, hdev->id,
6648                                        MGMT_OP_SET_FAST_CONNECTABLE,
6649                                        MGMT_STATUS_NOT_SUPPORTED);
6650
6651         if (cp->val != 0x00 && cp->val != 0x01)
6652                 return mgmt_cmd_status(sk, hdev->id,
6653                                        MGMT_OP_SET_FAST_CONNECTABLE,
6654                                        MGMT_STATUS_INVALID_PARAMS);
6655
6656         hci_dev_lock(hdev);
6657
6658         if (!!cp->val == hci_dev_test_flag(hdev, HCI_FAST_CONNECTABLE)) {
6659                 err = send_settings_rsp(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev);
6660                 goto unlock;
6661         }
6662
6663         if (!hdev_is_powered(hdev)) {
6664                 hci_dev_change_flag(hdev, HCI_FAST_CONNECTABLE);
6665                 err = send_settings_rsp(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev);
6666                 new_settings(hdev, sk);
6667                 goto unlock;
6668         }
6669
6670         cmd = mgmt_pending_new(sk, MGMT_OP_SET_FAST_CONNECTABLE, hdev, data,
6671                                len);
6672         if (!cmd)
6673                 err = -ENOMEM;
6674         else
6675                 err = hci_cmd_sync_queue(hdev, write_fast_connectable_sync, cmd,
6676                                          fast_connectable_complete);
6677
6678         if (err < 0) {
6679                 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_FAST_CONNECTABLE,
6680                                 MGMT_STATUS_FAILED);
6681
6682                 if (cmd)
6683                         mgmt_pending_free(cmd);
6684         }
6685
6686 unlock:
6687         hci_dev_unlock(hdev);
6688
6689         return err;
6690 }
6691
6692 static void set_bredr_complete(struct hci_dev *hdev, void *data, int err)
6693 {
6694         struct mgmt_pending_cmd *cmd = data;
6695
6696         bt_dev_dbg(hdev, "err %d", err);
6697
6698         if (err) {
6699                 u8 mgmt_err = mgmt_status(err);
6700
6701                 /* We need to restore the flag if related HCI commands
6702                  * failed.
6703                  */
6704                 hci_dev_clear_flag(hdev, HCI_BREDR_ENABLED);
6705
6706                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err);
6707         } else {
6708                 send_settings_rsp(cmd->sk, MGMT_OP_SET_BREDR, hdev);
6709                 new_settings(hdev, cmd->sk);
6710         }
6711
6712         mgmt_pending_free(cmd);
6713 }
6714
6715 static int set_bredr_sync(struct hci_dev *hdev, void *data)
6716 {
6717         int status;
6718
6719         status = hci_write_fast_connectable_sync(hdev, false);
6720
6721         if (!status)
6722                 status = hci_update_scan_sync(hdev);
6723
6724         /* Since only the advertising data flags will change, there
6725          * is no need to update the scan response data.
6726          */
6727         if (!status)
6728                 status = hci_update_adv_data_sync(hdev, hdev->cur_adv_instance);
6729
6730         return status;
6731 }
6732
6733 static int set_bredr(struct sock *sk, struct hci_dev *hdev, void *data, u16 len)
6734 {
6735         struct mgmt_mode *cp = data;
6736         struct mgmt_pending_cmd *cmd;
6737         int err;
6738
6739         bt_dev_dbg(hdev, "sock %p", sk);
6740
6741         if (!lmp_bredr_capable(hdev) || !lmp_le_capable(hdev))
6742                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR,
6743                                        MGMT_STATUS_NOT_SUPPORTED);
6744
6745         if (!hci_dev_test_flag(hdev, HCI_LE_ENABLED))
6746                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR,
6747                                        MGMT_STATUS_REJECTED);
6748
6749         if (cp->val != 0x00 && cp->val != 0x01)
6750                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR,
6751                                        MGMT_STATUS_INVALID_PARAMS);
6752
6753         hci_dev_lock(hdev);
6754
6755         if (cp->val == hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) {
6756                 err = send_settings_rsp(sk, MGMT_OP_SET_BREDR, hdev);
6757                 goto unlock;
6758         }
6759
6760         if (!hdev_is_powered(hdev)) {
6761                 if (!cp->val) {
6762                         hci_dev_clear_flag(hdev, HCI_DISCOVERABLE);
6763                         hci_dev_clear_flag(hdev, HCI_SSP_ENABLED);
6764                         hci_dev_clear_flag(hdev, HCI_LINK_SECURITY);
6765                         hci_dev_clear_flag(hdev, HCI_FAST_CONNECTABLE);
6766                         hci_dev_clear_flag(hdev, HCI_HS_ENABLED);
6767                 }
6768
6769                 hci_dev_change_flag(hdev, HCI_BREDR_ENABLED);
6770
6771                 err = send_settings_rsp(sk, MGMT_OP_SET_BREDR, hdev);
6772                 if (err < 0)
6773                         goto unlock;
6774
6775                 err = new_settings(hdev, sk);
6776                 goto unlock;
6777         }
6778
6779         /* Reject disabling when powered on */
6780         if (!cp->val) {
6781                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR,
6782                                       MGMT_STATUS_REJECTED);
6783                 goto unlock;
6784         } else {
6785                 /* When configuring a dual-mode controller to operate
6786                  * with LE only and using a static address, then switching
6787                  * BR/EDR back on is not allowed.
6788                  *
6789                  * Dual-mode controllers shall operate with the public
6790                  * address as its identity address for BR/EDR and LE. So
6791                  * reject the attempt to create an invalid configuration.
6792                  *
6793                  * The same restrictions applies when secure connections
6794                  * has been enabled. For BR/EDR this is a controller feature
6795                  * while for LE it is a host stack feature. This means that
6796                  * switching BR/EDR back on when secure connections has been
6797                  * enabled is not a supported transaction.
6798                  */
6799                 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) &&
6800                     (bacmp(&hdev->static_addr, BDADDR_ANY) ||
6801                      hci_dev_test_flag(hdev, HCI_SC_ENABLED))) {
6802                         err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR,
6803                                               MGMT_STATUS_REJECTED);
6804                         goto unlock;
6805                 }
6806         }
6807
6808         cmd = mgmt_pending_new(sk, MGMT_OP_SET_BREDR, hdev, data, len);
6809         if (!cmd)
6810                 err = -ENOMEM;
6811         else
6812                 err = hci_cmd_sync_queue(hdev, set_bredr_sync, cmd,
6813                                          set_bredr_complete);
6814
6815         if (err < 0) {
6816                 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_BREDR,
6817                                 MGMT_STATUS_FAILED);
6818                 if (cmd)
6819                         mgmt_pending_free(cmd);
6820
6821                 goto unlock;
6822         }
6823
6824         /* We need to flip the bit already here so that
6825          * hci_req_update_adv_data generates the correct flags.
6826          */
6827         hci_dev_set_flag(hdev, HCI_BREDR_ENABLED);
6828
6829 unlock:
6830         hci_dev_unlock(hdev);
6831         return err;
6832 }
6833
6834 static void set_secure_conn_complete(struct hci_dev *hdev, void *data, int err)
6835 {
6836         struct mgmt_pending_cmd *cmd = data;
6837         struct mgmt_mode *cp;
6838
6839         bt_dev_dbg(hdev, "err %d", err);
6840
6841         if (err) {
6842                 u8 mgmt_err = mgmt_status(err);
6843
6844                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode, mgmt_err);
6845                 goto done;
6846         }
6847
6848         cp = cmd->param;
6849
6850         switch (cp->val) {
6851         case 0x00:
6852                 hci_dev_clear_flag(hdev, HCI_SC_ENABLED);
6853                 hci_dev_clear_flag(hdev, HCI_SC_ONLY);
6854                 break;
6855         case 0x01:
6856                 hci_dev_set_flag(hdev, HCI_SC_ENABLED);
6857                 hci_dev_clear_flag(hdev, HCI_SC_ONLY);
6858                 break;
6859         case 0x02:
6860                 hci_dev_set_flag(hdev, HCI_SC_ENABLED);
6861                 hci_dev_set_flag(hdev, HCI_SC_ONLY);
6862                 break;
6863         }
6864
6865         send_settings_rsp(cmd->sk, cmd->opcode, hdev);
6866         new_settings(hdev, cmd->sk);
6867
6868 done:
6869         mgmt_pending_free(cmd);
6870 }
6871
6872 static int set_secure_conn_sync(struct hci_dev *hdev, void *data)
6873 {
6874         struct mgmt_pending_cmd *cmd = data;
6875         struct mgmt_mode *cp = cmd->param;
6876         u8 val = !!cp->val;
6877
6878         /* Force write of val */
6879         hci_dev_set_flag(hdev, HCI_SC_ENABLED);
6880
6881         return hci_write_sc_support_sync(hdev, val);
6882 }
6883
6884 static int set_secure_conn(struct sock *sk, struct hci_dev *hdev,
6885                            void *data, u16 len)
6886 {
6887         struct mgmt_mode *cp = data;
6888         struct mgmt_pending_cmd *cmd;
6889         u8 val;
6890         int err;
6891
6892         bt_dev_dbg(hdev, "sock %p", sk);
6893
6894         if (!lmp_sc_capable(hdev) &&
6895             !hci_dev_test_flag(hdev, HCI_LE_ENABLED))
6896                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN,
6897                                        MGMT_STATUS_NOT_SUPPORTED);
6898
6899         if (hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) &&
6900             lmp_sc_capable(hdev) &&
6901             !hci_dev_test_flag(hdev, HCI_SSP_ENABLED))
6902                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN,
6903                                        MGMT_STATUS_REJECTED);
6904
6905         if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02)
6906                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN,
6907                                        MGMT_STATUS_INVALID_PARAMS);
6908
6909         hci_dev_lock(hdev);
6910
6911         if (!hdev_is_powered(hdev) || !lmp_sc_capable(hdev) ||
6912             !hci_dev_test_flag(hdev, HCI_BREDR_ENABLED)) {
6913                 bool changed;
6914
6915                 if (cp->val) {
6916                         changed = !hci_dev_test_and_set_flag(hdev,
6917                                                              HCI_SC_ENABLED);
6918                         if (cp->val == 0x02)
6919                                 hci_dev_set_flag(hdev, HCI_SC_ONLY);
6920                         else
6921                                 hci_dev_clear_flag(hdev, HCI_SC_ONLY);
6922                 } else {
6923                         changed = hci_dev_test_and_clear_flag(hdev,
6924                                                               HCI_SC_ENABLED);
6925                         hci_dev_clear_flag(hdev, HCI_SC_ONLY);
6926                 }
6927
6928                 err = send_settings_rsp(sk, MGMT_OP_SET_SECURE_CONN, hdev);
6929                 if (err < 0)
6930                         goto failed;
6931
6932                 if (changed)
6933                         err = new_settings(hdev, sk);
6934
6935                 goto failed;
6936         }
6937
6938         val = !!cp->val;
6939
6940         if (val == hci_dev_test_flag(hdev, HCI_SC_ENABLED) &&
6941             (cp->val == 0x02) == hci_dev_test_flag(hdev, HCI_SC_ONLY)) {
6942                 err = send_settings_rsp(sk, MGMT_OP_SET_SECURE_CONN, hdev);
6943                 goto failed;
6944         }
6945
6946         cmd = mgmt_pending_new(sk, MGMT_OP_SET_SECURE_CONN, hdev, data, len);
6947         if (!cmd)
6948                 err = -ENOMEM;
6949         else
6950                 err = hci_cmd_sync_queue(hdev, set_secure_conn_sync, cmd,
6951                                          set_secure_conn_complete);
6952
6953         if (err < 0) {
6954                 mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_SECURE_CONN,
6955                                 MGMT_STATUS_FAILED);
6956                 if (cmd)
6957                         mgmt_pending_free(cmd);
6958         }
6959
6960 failed:
6961         hci_dev_unlock(hdev);
6962         return err;
6963 }
6964
6965 static int set_debug_keys(struct sock *sk, struct hci_dev *hdev,
6966                           void *data, u16 len)
6967 {
6968         struct mgmt_mode *cp = data;
6969         bool changed, use_changed;
6970         int err;
6971
6972         bt_dev_dbg(hdev, "sock %p", sk);
6973
6974         if (cp->val != 0x00 && cp->val != 0x01 && cp->val != 0x02)
6975                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_DEBUG_KEYS,
6976                                        MGMT_STATUS_INVALID_PARAMS);
6977
6978         hci_dev_lock(hdev);
6979
6980         if (cp->val)
6981                 changed = !hci_dev_test_and_set_flag(hdev, HCI_KEEP_DEBUG_KEYS);
6982         else
6983                 changed = hci_dev_test_and_clear_flag(hdev,
6984                                                       HCI_KEEP_DEBUG_KEYS);
6985
6986         if (cp->val == 0x02)
6987                 use_changed = !hci_dev_test_and_set_flag(hdev,
6988                                                          HCI_USE_DEBUG_KEYS);
6989         else
6990                 use_changed = hci_dev_test_and_clear_flag(hdev,
6991                                                           HCI_USE_DEBUG_KEYS);
6992
6993         if (hdev_is_powered(hdev) && use_changed &&
6994             hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) {
6995                 u8 mode = (cp->val == 0x02) ? 0x01 : 0x00;
6996                 hci_send_cmd(hdev, HCI_OP_WRITE_SSP_DEBUG_MODE,
6997                              sizeof(mode), &mode);
6998         }
6999
7000         err = send_settings_rsp(sk, MGMT_OP_SET_DEBUG_KEYS, hdev);
7001         if (err < 0)
7002                 goto unlock;
7003
7004         if (changed)
7005                 err = new_settings(hdev, sk);
7006
7007 unlock:
7008         hci_dev_unlock(hdev);
7009         return err;
7010 }
7011
7012 static int set_privacy(struct sock *sk, struct hci_dev *hdev, void *cp_data,
7013                        u16 len)
7014 {
7015         struct mgmt_cp_set_privacy *cp = cp_data;
7016         bool changed;
7017         int err;
7018
7019         bt_dev_dbg(hdev, "sock %p", sk);
7020
7021         if (!lmp_le_capable(hdev))
7022                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY,
7023                                        MGMT_STATUS_NOT_SUPPORTED);
7024
7025         if (cp->privacy != 0x00 && cp->privacy != 0x01 && cp->privacy != 0x02)
7026                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY,
7027                                        MGMT_STATUS_INVALID_PARAMS);
7028
7029         if (hdev_is_powered(hdev))
7030                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PRIVACY,
7031                                        MGMT_STATUS_REJECTED);
7032
7033         hci_dev_lock(hdev);
7034
7035         /* If user space supports this command it is also expected to
7036          * handle IRKs. Therefore, set the HCI_RPA_RESOLVING flag.
7037          */
7038         hci_dev_set_flag(hdev, HCI_RPA_RESOLVING);
7039
7040         if (cp->privacy) {
7041                 changed = !hci_dev_test_and_set_flag(hdev, HCI_PRIVACY);
7042                 memcpy(hdev->irk, cp->irk, sizeof(hdev->irk));
7043                 hci_dev_set_flag(hdev, HCI_RPA_EXPIRED);
7044                 hci_adv_instances_set_rpa_expired(hdev, true);
7045                 if (cp->privacy == 0x02)
7046                         hci_dev_set_flag(hdev, HCI_LIMITED_PRIVACY);
7047                 else
7048                         hci_dev_clear_flag(hdev, HCI_LIMITED_PRIVACY);
7049         } else {
7050                 changed = hci_dev_test_and_clear_flag(hdev, HCI_PRIVACY);
7051                 memset(hdev->irk, 0, sizeof(hdev->irk));
7052                 hci_dev_clear_flag(hdev, HCI_RPA_EXPIRED);
7053                 hci_adv_instances_set_rpa_expired(hdev, false);
7054                 hci_dev_clear_flag(hdev, HCI_LIMITED_PRIVACY);
7055         }
7056
7057         err = send_settings_rsp(sk, MGMT_OP_SET_PRIVACY, hdev);
7058         if (err < 0)
7059                 goto unlock;
7060
7061         if (changed)
7062                 err = new_settings(hdev, sk);
7063
7064 unlock:
7065         hci_dev_unlock(hdev);
7066         return err;
7067 }
7068
7069 static bool irk_is_valid(struct mgmt_irk_info *irk)
7070 {
7071         switch (irk->addr.type) {
7072         case BDADDR_LE_PUBLIC:
7073                 return true;
7074
7075         case BDADDR_LE_RANDOM:
7076                 /* Two most significant bits shall be set */
7077                 if ((irk->addr.bdaddr.b[5] & 0xc0) != 0xc0)
7078                         return false;
7079                 return true;
7080         }
7081
7082         return false;
7083 }
7084
7085 static int load_irks(struct sock *sk, struct hci_dev *hdev, void *cp_data,
7086                      u16 len)
7087 {
7088         struct mgmt_cp_load_irks *cp = cp_data;
7089         const u16 max_irk_count = ((U16_MAX - sizeof(*cp)) /
7090                                    sizeof(struct mgmt_irk_info));
7091         u16 irk_count, expected_len;
7092         int i, err;
7093
7094         bt_dev_dbg(hdev, "sock %p", sk);
7095
7096         if (!lmp_le_capable(hdev))
7097                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS,
7098                                        MGMT_STATUS_NOT_SUPPORTED);
7099
7100         irk_count = __le16_to_cpu(cp->irk_count);
7101         if (irk_count > max_irk_count) {
7102                 bt_dev_err(hdev, "load_irks: too big irk_count value %u",
7103                            irk_count);
7104                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS,
7105                                        MGMT_STATUS_INVALID_PARAMS);
7106         }
7107
7108         expected_len = struct_size(cp, irks, irk_count);
7109         if (expected_len != len) {
7110                 bt_dev_err(hdev, "load_irks: expected %u bytes, got %u bytes",
7111                            expected_len, len);
7112                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_IRKS,
7113                                        MGMT_STATUS_INVALID_PARAMS);
7114         }
7115
7116         bt_dev_dbg(hdev, "irk_count %u", irk_count);
7117
7118         for (i = 0; i < irk_count; i++) {
7119                 struct mgmt_irk_info *key = &cp->irks[i];
7120
7121                 if (!irk_is_valid(key))
7122                         return mgmt_cmd_status(sk, hdev->id,
7123                                                MGMT_OP_LOAD_IRKS,
7124                                                MGMT_STATUS_INVALID_PARAMS);
7125         }
7126
7127         hci_dev_lock(hdev);
7128
7129         hci_smp_irks_clear(hdev);
7130
7131         for (i = 0; i < irk_count; i++) {
7132                 struct mgmt_irk_info *irk = &cp->irks[i];
7133
7134                 if (hci_is_blocked_key(hdev,
7135                                        HCI_BLOCKED_KEY_TYPE_IRK,
7136                                        irk->val)) {
7137                         bt_dev_warn(hdev, "Skipping blocked IRK for %pMR",
7138                                     &irk->addr.bdaddr);
7139                         continue;
7140                 }
7141
7142                 hci_add_irk(hdev, &irk->addr.bdaddr,
7143                             le_addr_type(irk->addr.type), irk->val,
7144                             BDADDR_ANY);
7145         }
7146
7147         hci_dev_set_flag(hdev, HCI_RPA_RESOLVING);
7148
7149         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_IRKS, 0, NULL, 0);
7150
7151         hci_dev_unlock(hdev);
7152
7153         return err;
7154 }
7155
7156 static bool ltk_is_valid(struct mgmt_ltk_info *key)
7157 {
7158         if (key->initiator != 0x00 && key->initiator != 0x01)
7159                 return false;
7160
7161         switch (key->addr.type) {
7162         case BDADDR_LE_PUBLIC:
7163                 return true;
7164
7165         case BDADDR_LE_RANDOM:
7166                 /* Two most significant bits shall be set */
7167                 if ((key->addr.bdaddr.b[5] & 0xc0) != 0xc0)
7168                         return false;
7169                 return true;
7170         }
7171
7172         return false;
7173 }
7174
7175 static int load_long_term_keys(struct sock *sk, struct hci_dev *hdev,
7176                                void *cp_data, u16 len)
7177 {
7178         struct mgmt_cp_load_long_term_keys *cp = cp_data;
7179         const u16 max_key_count = ((U16_MAX - sizeof(*cp)) /
7180                                    sizeof(struct mgmt_ltk_info));
7181         u16 key_count, expected_len;
7182         int i, err;
7183
7184         bt_dev_dbg(hdev, "sock %p", sk);
7185
7186         if (!lmp_le_capable(hdev))
7187                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS,
7188                                        MGMT_STATUS_NOT_SUPPORTED);
7189
7190         key_count = __le16_to_cpu(cp->key_count);
7191         if (key_count > max_key_count) {
7192                 bt_dev_err(hdev, "load_ltks: too big key_count value %u",
7193                            key_count);
7194                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS,
7195                                        MGMT_STATUS_INVALID_PARAMS);
7196         }
7197
7198         expected_len = struct_size(cp, keys, key_count);
7199         if (expected_len != len) {
7200                 bt_dev_err(hdev, "load_keys: expected %u bytes, got %u bytes",
7201                            expected_len, len);
7202                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS,
7203                                        MGMT_STATUS_INVALID_PARAMS);
7204         }
7205
7206         bt_dev_dbg(hdev, "key_count %u", key_count);
7207
7208         for (i = 0; i < key_count; i++) {
7209                 struct mgmt_ltk_info *key = &cp->keys[i];
7210
7211                 if (!ltk_is_valid(key))
7212                         return mgmt_cmd_status(sk, hdev->id,
7213                                                MGMT_OP_LOAD_LONG_TERM_KEYS,
7214                                                MGMT_STATUS_INVALID_PARAMS);
7215         }
7216
7217         hci_dev_lock(hdev);
7218
7219         hci_smp_ltks_clear(hdev);
7220
7221         for (i = 0; i < key_count; i++) {
7222                 struct mgmt_ltk_info *key = &cp->keys[i];
7223                 u8 type, authenticated;
7224
7225                 if (hci_is_blocked_key(hdev,
7226                                        HCI_BLOCKED_KEY_TYPE_LTK,
7227                                        key->val)) {
7228                         bt_dev_warn(hdev, "Skipping blocked LTK for %pMR",
7229                                     &key->addr.bdaddr);
7230                         continue;
7231                 }
7232
7233                 switch (key->type) {
7234                 case MGMT_LTK_UNAUTHENTICATED:
7235                         authenticated = 0x00;
7236                         type = key->initiator ? SMP_LTK : SMP_LTK_RESPONDER;
7237                         break;
7238                 case MGMT_LTK_AUTHENTICATED:
7239                         authenticated = 0x01;
7240                         type = key->initiator ? SMP_LTK : SMP_LTK_RESPONDER;
7241                         break;
7242                 case MGMT_LTK_P256_UNAUTH:
7243                         authenticated = 0x00;
7244                         type = SMP_LTK_P256;
7245                         break;
7246                 case MGMT_LTK_P256_AUTH:
7247                         authenticated = 0x01;
7248                         type = SMP_LTK_P256;
7249                         break;
7250                 case MGMT_LTK_P256_DEBUG:
7251                         authenticated = 0x00;
7252                         type = SMP_LTK_P256_DEBUG;
7253                         fallthrough;
7254                 default:
7255                         continue;
7256                 }
7257
7258                 hci_add_ltk(hdev, &key->addr.bdaddr,
7259                             le_addr_type(key->addr.type), type, authenticated,
7260                             key->val, key->enc_size, key->ediv, key->rand);
7261         }
7262
7263         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_LONG_TERM_KEYS, 0,
7264                            NULL, 0);
7265
7266         hci_dev_unlock(hdev);
7267
7268         return err;
7269 }
7270
7271 static void get_conn_info_complete(struct hci_dev *hdev, void *data, int err)
7272 {
7273         struct mgmt_pending_cmd *cmd = data;
7274         struct hci_conn *conn = cmd->user_data;
7275         struct mgmt_cp_get_conn_info *cp = cmd->param;
7276         struct mgmt_rp_get_conn_info rp;
7277         u8 status;
7278
7279         bt_dev_dbg(hdev, "err %d", err);
7280
7281         memcpy(&rp.addr, &cp->addr, sizeof(rp.addr));
7282
7283         status = mgmt_status(err);
7284         if (status == MGMT_STATUS_SUCCESS) {
7285                 rp.rssi = conn->rssi;
7286                 rp.tx_power = conn->tx_power;
7287                 rp.max_tx_power = conn->max_tx_power;
7288         } else {
7289                 rp.rssi = HCI_RSSI_INVALID;
7290                 rp.tx_power = HCI_TX_POWER_INVALID;
7291                 rp.max_tx_power = HCI_TX_POWER_INVALID;
7292         }
7293
7294         mgmt_cmd_complete(cmd->sk, cmd->index, MGMT_OP_GET_CONN_INFO, status,
7295                           &rp, sizeof(rp));
7296
7297         mgmt_pending_free(cmd);
7298 }
7299
7300 static int get_conn_info_sync(struct hci_dev *hdev, void *data)
7301 {
7302         struct mgmt_pending_cmd *cmd = data;
7303         struct mgmt_cp_get_conn_info *cp = cmd->param;
7304         struct hci_conn *conn;
7305         int err;
7306         __le16   handle;
7307
7308         /* Make sure we are still connected */
7309         if (cp->addr.type == BDADDR_BREDR)
7310                 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK,
7311                                                &cp->addr.bdaddr);
7312         else
7313                 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr);
7314
7315         if (!conn || conn->state != BT_CONNECTED)
7316                 return MGMT_STATUS_NOT_CONNECTED;
7317
7318         cmd->user_data = conn;
7319         handle = cpu_to_le16(conn->handle);
7320
7321         /* Refresh RSSI each time */
7322         err = hci_read_rssi_sync(hdev, handle);
7323
7324         /* For LE links TX power does not change thus we don't need to
7325          * query for it once value is known.
7326          */
7327         if (!err && (!bdaddr_type_is_le(cp->addr.type) ||
7328                      conn->tx_power == HCI_TX_POWER_INVALID))
7329                 err = hci_read_tx_power_sync(hdev, handle, 0x00);
7330
7331         /* Max TX power needs to be read only once per connection */
7332         if (!err && conn->max_tx_power == HCI_TX_POWER_INVALID)
7333                 err = hci_read_tx_power_sync(hdev, handle, 0x01);
7334
7335         return err;
7336 }
7337
7338 static int get_conn_info(struct sock *sk, struct hci_dev *hdev, void *data,
7339                          u16 len)
7340 {
7341         struct mgmt_cp_get_conn_info *cp = data;
7342         struct mgmt_rp_get_conn_info rp;
7343         struct hci_conn *conn;
7344         unsigned long conn_info_age;
7345         int err = 0;
7346
7347         bt_dev_dbg(hdev, "sock %p", sk);
7348
7349         memset(&rp, 0, sizeof(rp));
7350         bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr);
7351         rp.addr.type = cp->addr.type;
7352
7353         if (!bdaddr_type_is_valid(cp->addr.type))
7354                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO,
7355                                          MGMT_STATUS_INVALID_PARAMS,
7356                                          &rp, sizeof(rp));
7357
7358         hci_dev_lock(hdev);
7359
7360         if (!hdev_is_powered(hdev)) {
7361                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO,
7362                                         MGMT_STATUS_NOT_POWERED, &rp,
7363                                         sizeof(rp));
7364                 goto unlock;
7365         }
7366
7367         if (cp->addr.type == BDADDR_BREDR)
7368                 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK,
7369                                                &cp->addr.bdaddr);
7370         else
7371                 conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, &cp->addr.bdaddr);
7372
7373         if (!conn || conn->state != BT_CONNECTED) {
7374                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO,
7375                                         MGMT_STATUS_NOT_CONNECTED, &rp,
7376                                         sizeof(rp));
7377                 goto unlock;
7378         }
7379
7380         /* To avoid client trying to guess when to poll again for information we
7381          * calculate conn info age as random value between min/max set in hdev.
7382          */
7383         conn_info_age = get_random_u32_inclusive(hdev->conn_info_min_age,
7384                                                  hdev->conn_info_max_age - 1);
7385
7386         /* Query controller to refresh cached values if they are too old or were
7387          * never read.
7388          */
7389         if (time_after(jiffies, conn->conn_info_timestamp +
7390                        msecs_to_jiffies(conn_info_age)) ||
7391             !conn->conn_info_timestamp) {
7392                 struct mgmt_pending_cmd *cmd;
7393
7394                 cmd = mgmt_pending_new(sk, MGMT_OP_GET_CONN_INFO, hdev, data,
7395                                        len);
7396                 if (!cmd) {
7397                         err = -ENOMEM;
7398                 } else {
7399                         err = hci_cmd_sync_queue(hdev, get_conn_info_sync,
7400                                                  cmd, get_conn_info_complete);
7401                 }
7402
7403                 if (err < 0) {
7404                         mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO,
7405                                           MGMT_STATUS_FAILED, &rp, sizeof(rp));
7406
7407                         if (cmd)
7408                                 mgmt_pending_free(cmd);
7409
7410                         goto unlock;
7411                 }
7412
7413                 conn->conn_info_timestamp = jiffies;
7414         } else {
7415                 /* Cache is valid, just reply with values cached in hci_conn */
7416                 rp.rssi = conn->rssi;
7417                 rp.tx_power = conn->tx_power;
7418                 rp.max_tx_power = conn->max_tx_power;
7419
7420                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CONN_INFO,
7421                                         MGMT_STATUS_SUCCESS, &rp, sizeof(rp));
7422         }
7423
7424 unlock:
7425         hci_dev_unlock(hdev);
7426         return err;
7427 }
7428
7429 static void get_clock_info_complete(struct hci_dev *hdev, void *data, int err)
7430 {
7431         struct mgmt_pending_cmd *cmd = data;
7432         struct mgmt_cp_get_clock_info *cp = cmd->param;
7433         struct mgmt_rp_get_clock_info rp;
7434         struct hci_conn *conn = cmd->user_data;
7435         u8 status = mgmt_status(err);
7436
7437         bt_dev_dbg(hdev, "err %d", err);
7438
7439         memset(&rp, 0, sizeof(rp));
7440         bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr);
7441         rp.addr.type = cp->addr.type;
7442
7443         if (err)
7444                 goto complete;
7445
7446         rp.local_clock = cpu_to_le32(hdev->clock);
7447
7448         if (conn) {
7449                 rp.piconet_clock = cpu_to_le32(conn->clock);
7450                 rp.accuracy = cpu_to_le16(conn->clock_accuracy);
7451         }
7452
7453 complete:
7454         mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode, status, &rp,
7455                           sizeof(rp));
7456
7457         mgmt_pending_free(cmd);
7458 }
7459
7460 static int get_clock_info_sync(struct hci_dev *hdev, void *data)
7461 {
7462         struct mgmt_pending_cmd *cmd = data;
7463         struct mgmt_cp_get_clock_info *cp = cmd->param;
7464         struct hci_cp_read_clock hci_cp;
7465         struct hci_conn *conn;
7466
7467         memset(&hci_cp, 0, sizeof(hci_cp));
7468         hci_read_clock_sync(hdev, &hci_cp);
7469
7470         /* Make sure connection still exists */
7471         conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &cp->addr.bdaddr);
7472         if (!conn || conn->state != BT_CONNECTED)
7473                 return MGMT_STATUS_NOT_CONNECTED;
7474
7475         cmd->user_data = conn;
7476         hci_cp.handle = cpu_to_le16(conn->handle);
7477         hci_cp.which = 0x01; /* Piconet clock */
7478
7479         return hci_read_clock_sync(hdev, &hci_cp);
7480 }
7481
7482 static int get_clock_info(struct sock *sk, struct hci_dev *hdev, void *data,
7483                                                                 u16 len)
7484 {
7485         struct mgmt_cp_get_clock_info *cp = data;
7486         struct mgmt_rp_get_clock_info rp;
7487         struct mgmt_pending_cmd *cmd;
7488         struct hci_conn *conn;
7489         int err;
7490
7491         bt_dev_dbg(hdev, "sock %p", sk);
7492
7493         memset(&rp, 0, sizeof(rp));
7494         bacpy(&rp.addr.bdaddr, &cp->addr.bdaddr);
7495         rp.addr.type = cp->addr.type;
7496
7497         if (cp->addr.type != BDADDR_BREDR)
7498                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO,
7499                                          MGMT_STATUS_INVALID_PARAMS,
7500                                          &rp, sizeof(rp));
7501
7502         hci_dev_lock(hdev);
7503
7504         if (!hdev_is_powered(hdev)) {
7505                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO,
7506                                         MGMT_STATUS_NOT_POWERED, &rp,
7507                                         sizeof(rp));
7508                 goto unlock;
7509         }
7510
7511         if (bacmp(&cp->addr.bdaddr, BDADDR_ANY)) {
7512                 conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK,
7513                                                &cp->addr.bdaddr);
7514                 if (!conn || conn->state != BT_CONNECTED) {
7515                         err = mgmt_cmd_complete(sk, hdev->id,
7516                                                 MGMT_OP_GET_CLOCK_INFO,
7517                                                 MGMT_STATUS_NOT_CONNECTED,
7518                                                 &rp, sizeof(rp));
7519                         goto unlock;
7520                 }
7521         } else {
7522                 conn = NULL;
7523         }
7524
7525         cmd = mgmt_pending_new(sk, MGMT_OP_GET_CLOCK_INFO, hdev, data, len);
7526         if (!cmd)
7527                 err = -ENOMEM;
7528         else
7529                 err = hci_cmd_sync_queue(hdev, get_clock_info_sync, cmd,
7530                                          get_clock_info_complete);
7531
7532         if (err < 0) {
7533                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_CLOCK_INFO,
7534                                         MGMT_STATUS_FAILED, &rp, sizeof(rp));
7535
7536                 if (cmd)
7537                         mgmt_pending_free(cmd);
7538         }
7539
7540
7541 unlock:
7542         hci_dev_unlock(hdev);
7543         return err;
7544 }
7545
7546 static bool is_connected(struct hci_dev *hdev, bdaddr_t *addr, u8 type)
7547 {
7548         struct hci_conn *conn;
7549
7550         conn = hci_conn_hash_lookup_ba(hdev, LE_LINK, addr);
7551         if (!conn)
7552                 return false;
7553
7554         if (conn->dst_type != type)
7555                 return false;
7556
7557         if (conn->state != BT_CONNECTED)
7558                 return false;
7559
7560         return true;
7561 }
7562
7563 /* This function requires the caller holds hdev->lock */
7564 static int hci_conn_params_set(struct hci_dev *hdev, bdaddr_t *addr,
7565                                u8 addr_type, u8 auto_connect)
7566 {
7567         struct hci_conn_params *params;
7568
7569         params = hci_conn_params_add(hdev, addr, addr_type);
7570         if (!params)
7571                 return -EIO;
7572
7573         if (params->auto_connect == auto_connect)
7574                 return 0;
7575
7576         hci_pend_le_list_del_init(params);
7577
7578         switch (auto_connect) {
7579         case HCI_AUTO_CONN_DISABLED:
7580         case HCI_AUTO_CONN_LINK_LOSS:
7581                 /* If auto connect is being disabled when we're trying to
7582                  * connect to device, keep connecting.
7583                  */
7584                 if (params->explicit_connect)
7585                         hci_pend_le_list_add(params, &hdev->pend_le_conns);
7586                 break;
7587         case HCI_AUTO_CONN_REPORT:
7588                 if (params->explicit_connect)
7589                         hci_pend_le_list_add(params, &hdev->pend_le_conns);
7590                 else
7591                         hci_pend_le_list_add(params, &hdev->pend_le_reports);
7592                 break;
7593         case HCI_AUTO_CONN_DIRECT:
7594         case HCI_AUTO_CONN_ALWAYS:
7595                 if (!is_connected(hdev, addr, addr_type))
7596                         hci_pend_le_list_add(params, &hdev->pend_le_conns);
7597                 break;
7598         }
7599
7600         params->auto_connect = auto_connect;
7601
7602         bt_dev_dbg(hdev, "addr %pMR (type %u) auto_connect %u",
7603                    addr, addr_type, auto_connect);
7604
7605         return 0;
7606 }
7607
7608 static void device_added(struct sock *sk, struct hci_dev *hdev,
7609                          bdaddr_t *bdaddr, u8 type, u8 action)
7610 {
7611         struct mgmt_ev_device_added ev;
7612
7613         bacpy(&ev.addr.bdaddr, bdaddr);
7614         ev.addr.type = type;
7615         ev.action = action;
7616
7617         mgmt_event(MGMT_EV_DEVICE_ADDED, hdev, &ev, sizeof(ev), sk);
7618 }
7619
7620 static int add_device_sync(struct hci_dev *hdev, void *data)
7621 {
7622         return hci_update_passive_scan_sync(hdev);
7623 }
7624
7625 static int add_device(struct sock *sk, struct hci_dev *hdev,
7626                       void *data, u16 len)
7627 {
7628         struct mgmt_cp_add_device *cp = data;
7629         u8 auto_conn, addr_type;
7630         struct hci_conn_params *params;
7631         int err;
7632         u32 current_flags = 0;
7633         u32 supported_flags;
7634
7635         bt_dev_dbg(hdev, "sock %p", sk);
7636
7637         if (!bdaddr_type_is_valid(cp->addr.type) ||
7638             !bacmp(&cp->addr.bdaddr, BDADDR_ANY))
7639                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE,
7640                                          MGMT_STATUS_INVALID_PARAMS,
7641                                          &cp->addr, sizeof(cp->addr));
7642
7643         if (cp->action != 0x00 && cp->action != 0x01 && cp->action != 0x02)
7644                 return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE,
7645                                          MGMT_STATUS_INVALID_PARAMS,
7646                                          &cp->addr, sizeof(cp->addr));
7647
7648         hci_dev_lock(hdev);
7649
7650         if (cp->addr.type == BDADDR_BREDR) {
7651                 /* Only incoming connections action is supported for now */
7652                 if (cp->action != 0x01) {
7653                         err = mgmt_cmd_complete(sk, hdev->id,
7654                                                 MGMT_OP_ADD_DEVICE,
7655                                                 MGMT_STATUS_INVALID_PARAMS,
7656                                                 &cp->addr, sizeof(cp->addr));
7657                         goto unlock;
7658                 }
7659
7660                 err = hci_bdaddr_list_add_with_flags(&hdev->accept_list,
7661                                                      &cp->addr.bdaddr,
7662                                                      cp->addr.type, 0);
7663                 if (err)
7664                         goto unlock;
7665
7666                 hci_update_scan(hdev);
7667
7668                 goto added;
7669         }
7670
7671         addr_type = le_addr_type(cp->addr.type);
7672
7673         if (cp->action == 0x02)
7674                 auto_conn = HCI_AUTO_CONN_ALWAYS;
7675         else if (cp->action == 0x01)
7676                 auto_conn = HCI_AUTO_CONN_DIRECT;
7677         else
7678                 auto_conn = HCI_AUTO_CONN_REPORT;
7679
7680         /* Kernel internally uses conn_params with resolvable private
7681          * address, but Add Device allows only identity addresses.
7682          * Make sure it is enforced before calling
7683          * hci_conn_params_lookup.
7684          */
7685         if (!hci_is_identity_address(&cp->addr.bdaddr, addr_type)) {
7686                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE,
7687                                         MGMT_STATUS_INVALID_PARAMS,
7688                                         &cp->addr, sizeof(cp->addr));
7689                 goto unlock;
7690         }
7691
7692         /* If the connection parameters don't exist for this device,
7693          * they will be created and configured with defaults.
7694          */
7695         if (hci_conn_params_set(hdev, &cp->addr.bdaddr, addr_type,
7696                                 auto_conn) < 0) {
7697                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE,
7698                                         MGMT_STATUS_FAILED, &cp->addr,
7699                                         sizeof(cp->addr));
7700                 goto unlock;
7701         } else {
7702                 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr,
7703                                                 addr_type);
7704                 if (params)
7705                         current_flags = params->flags;
7706         }
7707
7708         err = hci_cmd_sync_queue(hdev, add_device_sync, NULL, NULL);
7709         if (err < 0)
7710                 goto unlock;
7711
7712 added:
7713         device_added(sk, hdev, &cp->addr.bdaddr, cp->addr.type, cp->action);
7714         supported_flags = hdev->conn_flags;
7715         device_flags_changed(NULL, hdev, &cp->addr.bdaddr, cp->addr.type,
7716                              supported_flags, current_flags);
7717
7718         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_DEVICE,
7719                                 MGMT_STATUS_SUCCESS, &cp->addr,
7720                                 sizeof(cp->addr));
7721
7722 unlock:
7723         hci_dev_unlock(hdev);
7724         return err;
7725 }
7726
7727 static void device_removed(struct sock *sk, struct hci_dev *hdev,
7728                            bdaddr_t *bdaddr, u8 type)
7729 {
7730         struct mgmt_ev_device_removed ev;
7731
7732         bacpy(&ev.addr.bdaddr, bdaddr);
7733         ev.addr.type = type;
7734
7735         mgmt_event(MGMT_EV_DEVICE_REMOVED, hdev, &ev, sizeof(ev), sk);
7736 }
7737
7738 static int remove_device_sync(struct hci_dev *hdev, void *data)
7739 {
7740         return hci_update_passive_scan_sync(hdev);
7741 }
7742
7743 static int remove_device(struct sock *sk, struct hci_dev *hdev,
7744                          void *data, u16 len)
7745 {
7746         struct mgmt_cp_remove_device *cp = data;
7747         int err;
7748
7749         bt_dev_dbg(hdev, "sock %p", sk);
7750
7751         hci_dev_lock(hdev);
7752
7753         if (bacmp(&cp->addr.bdaddr, BDADDR_ANY)) {
7754                 struct hci_conn_params *params;
7755                 u8 addr_type;
7756
7757                 if (!bdaddr_type_is_valid(cp->addr.type)) {
7758                         err = mgmt_cmd_complete(sk, hdev->id,
7759                                                 MGMT_OP_REMOVE_DEVICE,
7760                                                 MGMT_STATUS_INVALID_PARAMS,
7761                                                 &cp->addr, sizeof(cp->addr));
7762                         goto unlock;
7763                 }
7764
7765                 if (cp->addr.type == BDADDR_BREDR) {
7766                         err = hci_bdaddr_list_del(&hdev->accept_list,
7767                                                   &cp->addr.bdaddr,
7768                                                   cp->addr.type);
7769                         if (err) {
7770                                 err = mgmt_cmd_complete(sk, hdev->id,
7771                                                         MGMT_OP_REMOVE_DEVICE,
7772                                                         MGMT_STATUS_INVALID_PARAMS,
7773                                                         &cp->addr,
7774                                                         sizeof(cp->addr));
7775                                 goto unlock;
7776                         }
7777
7778                         hci_update_scan(hdev);
7779
7780                         device_removed(sk, hdev, &cp->addr.bdaddr,
7781                                        cp->addr.type);
7782                         goto complete;
7783                 }
7784
7785                 addr_type = le_addr_type(cp->addr.type);
7786
7787                 /* Kernel internally uses conn_params with resolvable private
7788                  * address, but Remove Device allows only identity addresses.
7789                  * Make sure it is enforced before calling
7790                  * hci_conn_params_lookup.
7791                  */
7792                 if (!hci_is_identity_address(&cp->addr.bdaddr, addr_type)) {
7793                         err = mgmt_cmd_complete(sk, hdev->id,
7794                                                 MGMT_OP_REMOVE_DEVICE,
7795                                                 MGMT_STATUS_INVALID_PARAMS,
7796                                                 &cp->addr, sizeof(cp->addr));
7797                         goto unlock;
7798                 }
7799
7800                 params = hci_conn_params_lookup(hdev, &cp->addr.bdaddr,
7801                                                 addr_type);
7802                 if (!params) {
7803                         err = mgmt_cmd_complete(sk, hdev->id,
7804                                                 MGMT_OP_REMOVE_DEVICE,
7805                                                 MGMT_STATUS_INVALID_PARAMS,
7806                                                 &cp->addr, sizeof(cp->addr));
7807                         goto unlock;
7808                 }
7809
7810                 if (params->auto_connect == HCI_AUTO_CONN_DISABLED ||
7811                     params->auto_connect == HCI_AUTO_CONN_EXPLICIT) {
7812                         err = mgmt_cmd_complete(sk, hdev->id,
7813                                                 MGMT_OP_REMOVE_DEVICE,
7814                                                 MGMT_STATUS_INVALID_PARAMS,
7815                                                 &cp->addr, sizeof(cp->addr));
7816                         goto unlock;
7817                 }
7818
7819                 hci_conn_params_free(params);
7820
7821                 device_removed(sk, hdev, &cp->addr.bdaddr, cp->addr.type);
7822         } else {
7823                 struct hci_conn_params *p, *tmp;
7824                 struct bdaddr_list *b, *btmp;
7825
7826                 if (cp->addr.type) {
7827                         err = mgmt_cmd_complete(sk, hdev->id,
7828                                                 MGMT_OP_REMOVE_DEVICE,
7829                                                 MGMT_STATUS_INVALID_PARAMS,
7830                                                 &cp->addr, sizeof(cp->addr));
7831                         goto unlock;
7832                 }
7833
7834                 list_for_each_entry_safe(b, btmp, &hdev->accept_list, list) {
7835                         device_removed(sk, hdev, &b->bdaddr, b->bdaddr_type);
7836                         list_del(&b->list);
7837                         kfree(b);
7838                 }
7839
7840                 hci_update_scan(hdev);
7841
7842                 list_for_each_entry_safe(p, tmp, &hdev->le_conn_params, list) {
7843                         if (p->auto_connect == HCI_AUTO_CONN_DISABLED)
7844                                 continue;
7845                         device_removed(sk, hdev, &p->addr, p->addr_type);
7846                         if (p->explicit_connect) {
7847                                 p->auto_connect = HCI_AUTO_CONN_EXPLICIT;
7848                                 continue;
7849                         }
7850                         hci_conn_params_free(p);
7851                 }
7852
7853                 bt_dev_dbg(hdev, "All LE connection parameters were removed");
7854         }
7855
7856         hci_cmd_sync_queue(hdev, remove_device_sync, NULL, NULL);
7857
7858 complete:
7859         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_REMOVE_DEVICE,
7860                                 MGMT_STATUS_SUCCESS, &cp->addr,
7861                                 sizeof(cp->addr));
7862 unlock:
7863         hci_dev_unlock(hdev);
7864         return err;
7865 }
7866
7867 static int load_conn_param(struct sock *sk, struct hci_dev *hdev, void *data,
7868                            u16 len)
7869 {
7870         struct mgmt_cp_load_conn_param *cp = data;
7871         const u16 max_param_count = ((U16_MAX - sizeof(*cp)) /
7872                                      sizeof(struct mgmt_conn_param));
7873         u16 param_count, expected_len;
7874         int i;
7875
7876         if (!lmp_le_capable(hdev))
7877                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM,
7878                                        MGMT_STATUS_NOT_SUPPORTED);
7879
7880         param_count = __le16_to_cpu(cp->param_count);
7881         if (param_count > max_param_count) {
7882                 bt_dev_err(hdev, "load_conn_param: too big param_count value %u",
7883                            param_count);
7884                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM,
7885                                        MGMT_STATUS_INVALID_PARAMS);
7886         }
7887
7888         expected_len = struct_size(cp, params, param_count);
7889         if (expected_len != len) {
7890                 bt_dev_err(hdev, "load_conn_param: expected %u bytes, got %u bytes",
7891                            expected_len, len);
7892                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM,
7893                                        MGMT_STATUS_INVALID_PARAMS);
7894         }
7895
7896         bt_dev_dbg(hdev, "param_count %u", param_count);
7897
7898         hci_dev_lock(hdev);
7899
7900         hci_conn_params_clear_disabled(hdev);
7901
7902         for (i = 0; i < param_count; i++) {
7903                 struct mgmt_conn_param *param = &cp->params[i];
7904                 struct hci_conn_params *hci_param;
7905                 u16 min, max, latency, timeout;
7906                 u8 addr_type;
7907
7908                 bt_dev_dbg(hdev, "Adding %pMR (type %u)", &param->addr.bdaddr,
7909                            param->addr.type);
7910
7911                 if (param->addr.type == BDADDR_LE_PUBLIC) {
7912                         addr_type = ADDR_LE_DEV_PUBLIC;
7913                 } else if (param->addr.type == BDADDR_LE_RANDOM) {
7914                         addr_type = ADDR_LE_DEV_RANDOM;
7915                 } else {
7916                         bt_dev_err(hdev, "ignoring invalid connection parameters");
7917                         continue;
7918                 }
7919
7920                 min = le16_to_cpu(param->min_interval);
7921                 max = le16_to_cpu(param->max_interval);
7922                 latency = le16_to_cpu(param->latency);
7923                 timeout = le16_to_cpu(param->timeout);
7924
7925                 bt_dev_dbg(hdev, "min 0x%04x max 0x%04x latency 0x%04x timeout 0x%04x",
7926                            min, max, latency, timeout);
7927
7928                 if (hci_check_conn_params(min, max, latency, timeout) < 0) {
7929                         bt_dev_err(hdev, "ignoring invalid connection parameters");
7930                         continue;
7931                 }
7932
7933                 hci_param = hci_conn_params_add(hdev, &param->addr.bdaddr,
7934                                                 addr_type);
7935                 if (!hci_param) {
7936                         bt_dev_err(hdev, "failed to add connection parameters");
7937                         continue;
7938                 }
7939
7940                 hci_param->conn_min_interval = min;
7941                 hci_param->conn_max_interval = max;
7942                 hci_param->conn_latency = latency;
7943                 hci_param->supervision_timeout = timeout;
7944         }
7945
7946         hci_dev_unlock(hdev);
7947
7948         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_LOAD_CONN_PARAM, 0,
7949                                  NULL, 0);
7950 }
7951
7952 static int set_external_config(struct sock *sk, struct hci_dev *hdev,
7953                                void *data, u16 len)
7954 {
7955         struct mgmt_cp_set_external_config *cp = data;
7956         bool changed;
7957         int err;
7958
7959         bt_dev_dbg(hdev, "sock %p", sk);
7960
7961         if (hdev_is_powered(hdev))
7962                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG,
7963                                        MGMT_STATUS_REJECTED);
7964
7965         if (cp->config != 0x00 && cp->config != 0x01)
7966                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG,
7967                                          MGMT_STATUS_INVALID_PARAMS);
7968
7969         if (!test_bit(HCI_QUIRK_EXTERNAL_CONFIG, &hdev->quirks))
7970                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_EXTERNAL_CONFIG,
7971                                        MGMT_STATUS_NOT_SUPPORTED);
7972
7973         hci_dev_lock(hdev);
7974
7975         if (cp->config)
7976                 changed = !hci_dev_test_and_set_flag(hdev, HCI_EXT_CONFIGURED);
7977         else
7978                 changed = hci_dev_test_and_clear_flag(hdev, HCI_EXT_CONFIGURED);
7979
7980         err = send_options_rsp(sk, MGMT_OP_SET_EXTERNAL_CONFIG, hdev);
7981         if (err < 0)
7982                 goto unlock;
7983
7984         if (!changed)
7985                 goto unlock;
7986
7987         err = new_options(hdev, sk);
7988
7989         if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED) == is_configured(hdev)) {
7990                 mgmt_index_removed(hdev);
7991
7992                 if (hci_dev_test_and_change_flag(hdev, HCI_UNCONFIGURED)) {
7993                         hci_dev_set_flag(hdev, HCI_CONFIG);
7994                         hci_dev_set_flag(hdev, HCI_AUTO_OFF);
7995
7996                         queue_work(hdev->req_workqueue, &hdev->power_on);
7997                 } else {
7998                         set_bit(HCI_RAW, &hdev->flags);
7999                         mgmt_index_added(hdev);
8000                 }
8001         }
8002
8003 unlock:
8004         hci_dev_unlock(hdev);
8005         return err;
8006 }
8007
8008 static int set_public_address(struct sock *sk, struct hci_dev *hdev,
8009                               void *data, u16 len)
8010 {
8011         struct mgmt_cp_set_public_address *cp = data;
8012         bool changed;
8013         int err;
8014
8015         bt_dev_dbg(hdev, "sock %p", sk);
8016
8017         if (hdev_is_powered(hdev))
8018                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS,
8019                                        MGMT_STATUS_REJECTED);
8020
8021         if (!bacmp(&cp->bdaddr, BDADDR_ANY))
8022                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS,
8023                                        MGMT_STATUS_INVALID_PARAMS);
8024
8025         if (!hdev->set_bdaddr)
8026                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_SET_PUBLIC_ADDRESS,
8027                                        MGMT_STATUS_NOT_SUPPORTED);
8028
8029         hci_dev_lock(hdev);
8030
8031         changed = !!bacmp(&hdev->public_addr, &cp->bdaddr);
8032         bacpy(&hdev->public_addr, &cp->bdaddr);
8033
8034         err = send_options_rsp(sk, MGMT_OP_SET_PUBLIC_ADDRESS, hdev);
8035         if (err < 0)
8036                 goto unlock;
8037
8038         if (!changed)
8039                 goto unlock;
8040
8041         if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED))
8042                 err = new_options(hdev, sk);
8043
8044         if (is_configured(hdev)) {
8045                 mgmt_index_removed(hdev);
8046
8047                 hci_dev_clear_flag(hdev, HCI_UNCONFIGURED);
8048
8049                 hci_dev_set_flag(hdev, HCI_CONFIG);
8050                 hci_dev_set_flag(hdev, HCI_AUTO_OFF);
8051
8052                 queue_work(hdev->req_workqueue, &hdev->power_on);
8053         }
8054
8055 unlock:
8056         hci_dev_unlock(hdev);
8057         return err;
8058 }
8059
8060 static void read_local_oob_ext_data_complete(struct hci_dev *hdev, void *data,
8061                                              int err)
8062 {
8063         const struct mgmt_cp_read_local_oob_ext_data *mgmt_cp;
8064         struct mgmt_rp_read_local_oob_ext_data *mgmt_rp;
8065         u8 *h192, *r192, *h256, *r256;
8066         struct mgmt_pending_cmd *cmd = data;
8067         struct sk_buff *skb = cmd->skb;
8068         u8 status = mgmt_status(err);
8069         u16 eir_len;
8070
8071         if (cmd != pending_find(MGMT_OP_READ_LOCAL_OOB_EXT_DATA, hdev))
8072                 return;
8073
8074         if (!status) {
8075                 if (!skb)
8076                         status = MGMT_STATUS_FAILED;
8077                 else if (IS_ERR(skb))
8078                         status = mgmt_status(PTR_ERR(skb));
8079                 else
8080                         status = mgmt_status(skb->data[0]);
8081         }
8082
8083         bt_dev_dbg(hdev, "status %u", status);
8084
8085         mgmt_cp = cmd->param;
8086
8087         if (status) {
8088                 status = mgmt_status(status);
8089                 eir_len = 0;
8090
8091                 h192 = NULL;
8092                 r192 = NULL;
8093                 h256 = NULL;
8094                 r256 = NULL;
8095         } else if (!bredr_sc_enabled(hdev)) {
8096                 struct hci_rp_read_local_oob_data *rp;
8097
8098                 if (skb->len != sizeof(*rp)) {
8099                         status = MGMT_STATUS_FAILED;
8100                         eir_len = 0;
8101                 } else {
8102                         status = MGMT_STATUS_SUCCESS;
8103                         rp = (void *)skb->data;
8104
8105                         eir_len = 5 + 18 + 18;
8106                         h192 = rp->hash;
8107                         r192 = rp->rand;
8108                         h256 = NULL;
8109                         r256 = NULL;
8110                 }
8111         } else {
8112                 struct hci_rp_read_local_oob_ext_data *rp;
8113
8114                 if (skb->len != sizeof(*rp)) {
8115                         status = MGMT_STATUS_FAILED;
8116                         eir_len = 0;
8117                 } else {
8118                         status = MGMT_STATUS_SUCCESS;
8119                         rp = (void *)skb->data;
8120
8121                         if (hci_dev_test_flag(hdev, HCI_SC_ONLY)) {
8122                                 eir_len = 5 + 18 + 18;
8123                                 h192 = NULL;
8124                                 r192 = NULL;
8125                         } else {
8126                                 eir_len = 5 + 18 + 18 + 18 + 18;
8127                                 h192 = rp->hash192;
8128                                 r192 = rp->rand192;
8129                         }
8130
8131                         h256 = rp->hash256;
8132                         r256 = rp->rand256;
8133                 }
8134         }
8135
8136         mgmt_rp = kmalloc(sizeof(*mgmt_rp) + eir_len, GFP_KERNEL);
8137         if (!mgmt_rp)
8138                 goto done;
8139
8140         if (eir_len == 0)
8141                 goto send_rsp;
8142
8143         eir_len = eir_append_data(mgmt_rp->eir, 0, EIR_CLASS_OF_DEV,
8144                                   hdev->dev_class, 3);
8145
8146         if (h192 && r192) {
8147                 eir_len = eir_append_data(mgmt_rp->eir, eir_len,
8148                                           EIR_SSP_HASH_C192, h192, 16);
8149                 eir_len = eir_append_data(mgmt_rp->eir, eir_len,
8150                                           EIR_SSP_RAND_R192, r192, 16);
8151         }
8152
8153         if (h256 && r256) {
8154                 eir_len = eir_append_data(mgmt_rp->eir, eir_len,
8155                                           EIR_SSP_HASH_C256, h256, 16);
8156                 eir_len = eir_append_data(mgmt_rp->eir, eir_len,
8157                                           EIR_SSP_RAND_R256, r256, 16);
8158         }
8159
8160 send_rsp:
8161         mgmt_rp->type = mgmt_cp->type;
8162         mgmt_rp->eir_len = cpu_to_le16(eir_len);
8163
8164         err = mgmt_cmd_complete(cmd->sk, hdev->id,
8165                                 MGMT_OP_READ_LOCAL_OOB_EXT_DATA, status,
8166                                 mgmt_rp, sizeof(*mgmt_rp) + eir_len);
8167         if (err < 0 || status)
8168                 goto done;
8169
8170         hci_sock_set_flag(cmd->sk, HCI_MGMT_OOB_DATA_EVENTS);
8171
8172         err = mgmt_limited_event(MGMT_EV_LOCAL_OOB_DATA_UPDATED, hdev,
8173                                  mgmt_rp, sizeof(*mgmt_rp) + eir_len,
8174                                  HCI_MGMT_OOB_DATA_EVENTS, cmd->sk);
8175 done:
8176         if (skb && !IS_ERR(skb))
8177                 kfree_skb(skb);
8178
8179         kfree(mgmt_rp);
8180         mgmt_pending_remove(cmd);
8181 }
8182
8183 static int read_local_ssp_oob_req(struct hci_dev *hdev, struct sock *sk,
8184                                   struct mgmt_cp_read_local_oob_ext_data *cp)
8185 {
8186         struct mgmt_pending_cmd *cmd;
8187         int err;
8188
8189         cmd = mgmt_pending_add(sk, MGMT_OP_READ_LOCAL_OOB_EXT_DATA, hdev,
8190                                cp, sizeof(*cp));
8191         if (!cmd)
8192                 return -ENOMEM;
8193
8194         err = hci_cmd_sync_queue(hdev, read_local_oob_data_sync, cmd,
8195                                  read_local_oob_ext_data_complete);
8196
8197         if (err < 0) {
8198                 mgmt_pending_remove(cmd);
8199                 return err;
8200         }
8201
8202         return 0;
8203 }
8204
8205 static int read_local_oob_ext_data(struct sock *sk, struct hci_dev *hdev,
8206                                    void *data, u16 data_len)
8207 {
8208         struct mgmt_cp_read_local_oob_ext_data *cp = data;
8209         struct mgmt_rp_read_local_oob_ext_data *rp;
8210         size_t rp_len;
8211         u16 eir_len;
8212         u8 status, flags, role, addr[7], hash[16], rand[16];
8213         int err;
8214
8215         bt_dev_dbg(hdev, "sock %p", sk);
8216
8217         if (hdev_is_powered(hdev)) {
8218                 switch (cp->type) {
8219                 case BIT(BDADDR_BREDR):
8220                         status = mgmt_bredr_support(hdev);
8221                         if (status)
8222                                 eir_len = 0;
8223                         else
8224                                 eir_len = 5;
8225                         break;
8226                 case (BIT(BDADDR_LE_PUBLIC) | BIT(BDADDR_LE_RANDOM)):
8227                         status = mgmt_le_support(hdev);
8228                         if (status)
8229                                 eir_len = 0;
8230                         else
8231                                 eir_len = 9 + 3 + 18 + 18 + 3;
8232                         break;
8233                 default:
8234                         status = MGMT_STATUS_INVALID_PARAMS;
8235                         eir_len = 0;
8236                         break;
8237                 }
8238         } else {
8239                 status = MGMT_STATUS_NOT_POWERED;
8240                 eir_len = 0;
8241         }
8242
8243         rp_len = sizeof(*rp) + eir_len;
8244         rp = kmalloc(rp_len, GFP_ATOMIC);
8245         if (!rp)
8246                 return -ENOMEM;
8247
8248         if (!status && !lmp_ssp_capable(hdev)) {
8249                 status = MGMT_STATUS_NOT_SUPPORTED;
8250                 eir_len = 0;
8251         }
8252
8253         if (status)
8254                 goto complete;
8255
8256         hci_dev_lock(hdev);
8257
8258         eir_len = 0;
8259         switch (cp->type) {
8260         case BIT(BDADDR_BREDR):
8261                 if (hci_dev_test_flag(hdev, HCI_SSP_ENABLED)) {
8262                         err = read_local_ssp_oob_req(hdev, sk, cp);
8263                         hci_dev_unlock(hdev);
8264                         if (!err)
8265                                 goto done;
8266
8267                         status = MGMT_STATUS_FAILED;
8268                         goto complete;
8269                 } else {
8270                         eir_len = eir_append_data(rp->eir, eir_len,
8271                                                   EIR_CLASS_OF_DEV,
8272                                                   hdev->dev_class, 3);
8273                 }
8274                 break;
8275         case (BIT(BDADDR_LE_PUBLIC) | BIT(BDADDR_LE_RANDOM)):
8276                 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED) &&
8277                     smp_generate_oob(hdev, hash, rand) < 0) {
8278                         hci_dev_unlock(hdev);
8279                         status = MGMT_STATUS_FAILED;
8280                         goto complete;
8281                 }
8282
8283                 /* This should return the active RPA, but since the RPA
8284                  * is only programmed on demand, it is really hard to fill
8285                  * this in at the moment. For now disallow retrieving
8286                  * local out-of-band data when privacy is in use.
8287                  *
8288                  * Returning the identity address will not help here since
8289                  * pairing happens before the identity resolving key is
8290                  * known and thus the connection establishment happens
8291                  * based on the RPA and not the identity address.
8292                  */
8293                 if (hci_dev_test_flag(hdev, HCI_PRIVACY)) {
8294                         hci_dev_unlock(hdev);
8295                         status = MGMT_STATUS_REJECTED;
8296                         goto complete;
8297                 }
8298
8299                 if (hci_dev_test_flag(hdev, HCI_FORCE_STATIC_ADDR) ||
8300                    !bacmp(&hdev->bdaddr, BDADDR_ANY) ||
8301                    (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED) &&
8302                     bacmp(&hdev->static_addr, BDADDR_ANY))) {
8303                         memcpy(addr, &hdev->static_addr, 6);
8304                         addr[6] = 0x01;
8305                 } else {
8306                         memcpy(addr, &hdev->bdaddr, 6);
8307                         addr[6] = 0x00;
8308                 }
8309
8310                 eir_len = eir_append_data(rp->eir, eir_len, EIR_LE_BDADDR,
8311                                           addr, sizeof(addr));
8312
8313                 if (hci_dev_test_flag(hdev, HCI_ADVERTISING))
8314                         role = 0x02;
8315                 else
8316                         role = 0x01;
8317
8318                 eir_len = eir_append_data(rp->eir, eir_len, EIR_LE_ROLE,
8319                                           &role, sizeof(role));
8320
8321                 if (hci_dev_test_flag(hdev, HCI_SC_ENABLED)) {
8322                         eir_len = eir_append_data(rp->eir, eir_len,
8323                                                   EIR_LE_SC_CONFIRM,
8324                                                   hash, sizeof(hash));
8325
8326                         eir_len = eir_append_data(rp->eir, eir_len,
8327                                                   EIR_LE_SC_RANDOM,
8328                                                   rand, sizeof(rand));
8329                 }
8330
8331                 flags = mgmt_get_adv_discov_flags(hdev);
8332
8333                 if (!hci_dev_test_flag(hdev, HCI_BREDR_ENABLED))
8334                         flags |= LE_AD_NO_BREDR;
8335
8336                 eir_len = eir_append_data(rp->eir, eir_len, EIR_FLAGS,
8337                                           &flags, sizeof(flags));
8338                 break;
8339         }
8340
8341         hci_dev_unlock(hdev);
8342
8343         hci_sock_set_flag(sk, HCI_MGMT_OOB_DATA_EVENTS);
8344
8345         status = MGMT_STATUS_SUCCESS;
8346
8347 complete:
8348         rp->type = cp->type;
8349         rp->eir_len = cpu_to_le16(eir_len);
8350
8351         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_LOCAL_OOB_EXT_DATA,
8352                                 status, rp, sizeof(*rp) + eir_len);
8353         if (err < 0 || status)
8354                 goto done;
8355
8356         err = mgmt_limited_event(MGMT_EV_LOCAL_OOB_DATA_UPDATED, hdev,
8357                                  rp, sizeof(*rp) + eir_len,
8358                                  HCI_MGMT_OOB_DATA_EVENTS, sk);
8359
8360 done:
8361         kfree(rp);
8362
8363         return err;
8364 }
8365
8366 static u32 get_supported_adv_flags(struct hci_dev *hdev)
8367 {
8368         u32 flags = 0;
8369
8370         flags |= MGMT_ADV_FLAG_CONNECTABLE;
8371         flags |= MGMT_ADV_FLAG_DISCOV;
8372         flags |= MGMT_ADV_FLAG_LIMITED_DISCOV;
8373         flags |= MGMT_ADV_FLAG_MANAGED_FLAGS;
8374         flags |= MGMT_ADV_FLAG_APPEARANCE;
8375         flags |= MGMT_ADV_FLAG_LOCAL_NAME;
8376         flags |= MGMT_ADV_PARAM_DURATION;
8377         flags |= MGMT_ADV_PARAM_TIMEOUT;
8378         flags |= MGMT_ADV_PARAM_INTERVALS;
8379         flags |= MGMT_ADV_PARAM_TX_POWER;
8380         flags |= MGMT_ADV_PARAM_SCAN_RSP;
8381
8382         /* In extended adv TX_POWER returned from Set Adv Param
8383          * will be always valid.
8384          */
8385         if (hdev->adv_tx_power != HCI_TX_POWER_INVALID || ext_adv_capable(hdev))
8386                 flags |= MGMT_ADV_FLAG_TX_POWER;
8387
8388         if (ext_adv_capable(hdev)) {
8389                 flags |= MGMT_ADV_FLAG_SEC_1M;
8390                 flags |= MGMT_ADV_FLAG_HW_OFFLOAD;
8391                 flags |= MGMT_ADV_FLAG_CAN_SET_TX_POWER;
8392
8393                 if (le_2m_capable(hdev))
8394                         flags |= MGMT_ADV_FLAG_SEC_2M;
8395
8396                 if (le_coded_capable(hdev))
8397                         flags |= MGMT_ADV_FLAG_SEC_CODED;
8398         }
8399
8400         return flags;
8401 }
8402
8403 static int read_adv_features(struct sock *sk, struct hci_dev *hdev,
8404                              void *data, u16 data_len)
8405 {
8406         struct mgmt_rp_read_adv_features *rp;
8407         size_t rp_len;
8408         int err;
8409         struct adv_info *adv_instance;
8410         u32 supported_flags;
8411         u8 *instance;
8412
8413         bt_dev_dbg(hdev, "sock %p", sk);
8414
8415         if (!lmp_le_capable(hdev))
8416                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_READ_ADV_FEATURES,
8417                                        MGMT_STATUS_REJECTED);
8418
8419         hci_dev_lock(hdev);
8420
8421         rp_len = sizeof(*rp) + hdev->adv_instance_cnt;
8422         rp = kmalloc(rp_len, GFP_ATOMIC);
8423         if (!rp) {
8424                 hci_dev_unlock(hdev);
8425                 return -ENOMEM;
8426         }
8427
8428         supported_flags = get_supported_adv_flags(hdev);
8429
8430         rp->supported_flags = cpu_to_le32(supported_flags);
8431         rp->max_adv_data_len = max_adv_len(hdev);
8432         rp->max_scan_rsp_len = max_adv_len(hdev);
8433         rp->max_instances = hdev->le_num_of_adv_sets;
8434         rp->num_instances = hdev->adv_instance_cnt;
8435
8436         instance = rp->instance;
8437         list_for_each_entry(adv_instance, &hdev->adv_instances, list) {
8438                 /* Only instances 1-le_num_of_adv_sets are externally visible */
8439                 if (adv_instance->instance <= hdev->adv_instance_cnt) {
8440                         *instance = adv_instance->instance;
8441                         instance++;
8442                 } else {
8443                         rp->num_instances--;
8444                         rp_len--;
8445                 }
8446         }
8447
8448         hci_dev_unlock(hdev);
8449
8450         err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_READ_ADV_FEATURES,
8451                                 MGMT_STATUS_SUCCESS, rp, rp_len);
8452
8453         kfree(rp);
8454
8455         return err;
8456 }
8457
8458 static u8 calculate_name_len(struct hci_dev *hdev)
8459 {
8460         u8 buf[HCI_MAX_SHORT_NAME_LENGTH + 3];
8461
8462         return eir_append_local_name(hdev, buf, 0);
8463 }
8464
8465 static u8 tlv_data_max_len(struct hci_dev *hdev, u32 adv_flags,
8466                            bool is_adv_data)
8467 {
8468         u8 max_len = max_adv_len(hdev);
8469
8470         if (is_adv_data) {
8471                 if (adv_flags & (MGMT_ADV_FLAG_DISCOV |
8472                                  MGMT_ADV_FLAG_LIMITED_DISCOV |
8473                                  MGMT_ADV_FLAG_MANAGED_FLAGS))
8474                         max_len -= 3;
8475
8476                 if (adv_flags & MGMT_ADV_FLAG_TX_POWER)
8477                         max_len -= 3;
8478         } else {
8479                 if (adv_flags & MGMT_ADV_FLAG_LOCAL_NAME)
8480                         max_len -= calculate_name_len(hdev);
8481
8482                 if (adv_flags & (MGMT_ADV_FLAG_APPEARANCE))
8483                         max_len -= 4;
8484         }
8485
8486         return max_len;
8487 }
8488
8489 static bool flags_managed(u32 adv_flags)
8490 {
8491         return adv_flags & (MGMT_ADV_FLAG_DISCOV |
8492                             MGMT_ADV_FLAG_LIMITED_DISCOV |
8493                             MGMT_ADV_FLAG_MANAGED_FLAGS);
8494 }
8495
8496 static bool tx_power_managed(u32 adv_flags)
8497 {
8498         return adv_flags & MGMT_ADV_FLAG_TX_POWER;
8499 }
8500
8501 static bool name_managed(u32 adv_flags)
8502 {
8503         return adv_flags & MGMT_ADV_FLAG_LOCAL_NAME;
8504 }
8505
8506 static bool appearance_managed(u32 adv_flags)
8507 {
8508         return adv_flags & MGMT_ADV_FLAG_APPEARANCE;
8509 }
8510
8511 static bool tlv_data_is_valid(struct hci_dev *hdev, u32 adv_flags, u8 *data,
8512                               u8 len, bool is_adv_data)
8513 {
8514         int i, cur_len;
8515         u8 max_len;
8516
8517         max_len = tlv_data_max_len(hdev, adv_flags, is_adv_data);
8518
8519         if (len > max_len)
8520                 return false;
8521
8522         /* Make sure that the data is correctly formatted. */
8523         for (i = 0; i < len; i += (cur_len + 1)) {
8524                 cur_len = data[i];
8525
8526                 if (!cur_len)
8527                         continue;
8528
8529                 if (data[i + 1] == EIR_FLAGS &&
8530                     (!is_adv_data || flags_managed(adv_flags)))
8531                         return false;
8532
8533                 if (data[i + 1] == EIR_TX_POWER && tx_power_managed(adv_flags))
8534                         return false;
8535
8536                 if (data[i + 1] == EIR_NAME_COMPLETE && name_managed(adv_flags))
8537                         return false;
8538
8539                 if (data[i + 1] == EIR_NAME_SHORT && name_managed(adv_flags))
8540                         return false;
8541
8542                 if (data[i + 1] == EIR_APPEARANCE &&
8543                     appearance_managed(adv_flags))
8544                         return false;
8545
8546                 /* If the current field length would exceed the total data
8547                  * length, then it's invalid.
8548                  */
8549                 if (i + cur_len >= len)
8550                         return false;
8551         }
8552
8553         return true;
8554 }
8555
8556 static bool requested_adv_flags_are_valid(struct hci_dev *hdev, u32 adv_flags)
8557 {
8558         u32 supported_flags, phy_flags;
8559
8560         /* The current implementation only supports a subset of the specified
8561          * flags. Also need to check mutual exclusiveness of sec flags.
8562          */
8563         supported_flags = get_supported_adv_flags(hdev);
8564         phy_flags = adv_flags & MGMT_ADV_FLAG_SEC_MASK;
8565         if (adv_flags & ~supported_flags ||
8566             ((phy_flags && (phy_flags ^ (phy_flags & -phy_flags)))))
8567                 return false;
8568
8569         return true;
8570 }
8571
8572 static bool adv_busy(struct hci_dev *hdev)
8573 {
8574         return pending_find(MGMT_OP_SET_LE, hdev);
8575 }
8576
8577 static void add_adv_complete(struct hci_dev *hdev, struct sock *sk, u8 instance,
8578                              int err)
8579 {
8580         struct adv_info *adv, *n;
8581
8582         bt_dev_dbg(hdev, "err %d", err);
8583
8584         hci_dev_lock(hdev);
8585
8586         list_for_each_entry_safe(adv, n, &hdev->adv_instances, list) {
8587                 u8 instance;
8588
8589                 if (!adv->pending)
8590                         continue;
8591
8592                 if (!err) {
8593                         adv->pending = false;
8594                         continue;
8595                 }
8596
8597                 instance = adv->instance;
8598
8599                 if (hdev->cur_adv_instance == instance)
8600                         cancel_adv_timeout(hdev);
8601
8602                 hci_remove_adv_instance(hdev, instance);
8603                 mgmt_advertising_removed(sk, hdev, instance);
8604         }
8605
8606         hci_dev_unlock(hdev);
8607 }
8608
8609 static void add_advertising_complete(struct hci_dev *hdev, void *data, int err)
8610 {
8611         struct mgmt_pending_cmd *cmd = data;
8612         struct mgmt_cp_add_advertising *cp = cmd->param;
8613         struct mgmt_rp_add_advertising rp;
8614
8615         memset(&rp, 0, sizeof(rp));
8616
8617         rp.instance = cp->instance;
8618
8619         if (err)
8620                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode,
8621                                 mgmt_status(err));
8622         else
8623                 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode,
8624                                   mgmt_status(err), &rp, sizeof(rp));
8625
8626         add_adv_complete(hdev, cmd->sk, cp->instance, err);
8627
8628         mgmt_pending_free(cmd);
8629 }
8630
8631 static int add_advertising_sync(struct hci_dev *hdev, void *data)
8632 {
8633         struct mgmt_pending_cmd *cmd = data;
8634         struct mgmt_cp_add_advertising *cp = cmd->param;
8635
8636         return hci_schedule_adv_instance_sync(hdev, cp->instance, true);
8637 }
8638
8639 static int add_advertising(struct sock *sk, struct hci_dev *hdev,
8640                            void *data, u16 data_len)
8641 {
8642         struct mgmt_cp_add_advertising *cp = data;
8643         struct mgmt_rp_add_advertising rp;
8644         u32 flags;
8645         u8 status;
8646         u16 timeout, duration;
8647         unsigned int prev_instance_cnt;
8648         u8 schedule_instance = 0;
8649         struct adv_info *adv, *next_instance;
8650         int err;
8651         struct mgmt_pending_cmd *cmd;
8652
8653         bt_dev_dbg(hdev, "sock %p", sk);
8654
8655         status = mgmt_le_support(hdev);
8656         if (status)
8657                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8658                                        status);
8659
8660         if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets)
8661                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8662                                        MGMT_STATUS_INVALID_PARAMS);
8663
8664         if (data_len != sizeof(*cp) + cp->adv_data_len + cp->scan_rsp_len)
8665                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8666                                        MGMT_STATUS_INVALID_PARAMS);
8667
8668         flags = __le32_to_cpu(cp->flags);
8669         timeout = __le16_to_cpu(cp->timeout);
8670         duration = __le16_to_cpu(cp->duration);
8671
8672         if (!requested_adv_flags_are_valid(hdev, flags))
8673                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8674                                        MGMT_STATUS_INVALID_PARAMS);
8675
8676         hci_dev_lock(hdev);
8677
8678         if (timeout && !hdev_is_powered(hdev)) {
8679                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8680                                       MGMT_STATUS_REJECTED);
8681                 goto unlock;
8682         }
8683
8684         if (adv_busy(hdev)) {
8685                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8686                                       MGMT_STATUS_BUSY);
8687                 goto unlock;
8688         }
8689
8690         if (!tlv_data_is_valid(hdev, flags, cp->data, cp->adv_data_len, true) ||
8691             !tlv_data_is_valid(hdev, flags, cp->data + cp->adv_data_len,
8692                                cp->scan_rsp_len, false)) {
8693                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8694                                       MGMT_STATUS_INVALID_PARAMS);
8695                 goto unlock;
8696         }
8697
8698         prev_instance_cnt = hdev->adv_instance_cnt;
8699
8700         adv = hci_add_adv_instance(hdev, cp->instance, flags,
8701                                    cp->adv_data_len, cp->data,
8702                                    cp->scan_rsp_len,
8703                                    cp->data + cp->adv_data_len,
8704                                    timeout, duration,
8705                                    HCI_ADV_TX_POWER_NO_PREFERENCE,
8706                                    hdev->le_adv_min_interval,
8707                                    hdev->le_adv_max_interval, 0);
8708         if (IS_ERR(adv)) {
8709                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8710                                       MGMT_STATUS_FAILED);
8711                 goto unlock;
8712         }
8713
8714         /* Only trigger an advertising added event if a new instance was
8715          * actually added.
8716          */
8717         if (hdev->adv_instance_cnt > prev_instance_cnt)
8718                 mgmt_advertising_added(sk, hdev, cp->instance);
8719
8720         if (hdev->cur_adv_instance == cp->instance) {
8721                 /* If the currently advertised instance is being changed then
8722                  * cancel the current advertising and schedule the next
8723                  * instance. If there is only one instance then the overridden
8724                  * advertising data will be visible right away.
8725                  */
8726                 cancel_adv_timeout(hdev);
8727
8728                 next_instance = hci_get_next_instance(hdev, cp->instance);
8729                 if (next_instance)
8730                         schedule_instance = next_instance->instance;
8731         } else if (!hdev->adv_instance_timeout) {
8732                 /* Immediately advertise the new instance if no other
8733                  * instance is currently being advertised.
8734                  */
8735                 schedule_instance = cp->instance;
8736         }
8737
8738         /* If the HCI_ADVERTISING flag is set or the device isn't powered or
8739          * there is no instance to be advertised then we have no HCI
8740          * communication to make. Simply return.
8741          */
8742         if (!hdev_is_powered(hdev) ||
8743             hci_dev_test_flag(hdev, HCI_ADVERTISING) ||
8744             !schedule_instance) {
8745                 rp.instance = cp->instance;
8746                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_ADVERTISING,
8747                                         MGMT_STATUS_SUCCESS, &rp, sizeof(rp));
8748                 goto unlock;
8749         }
8750
8751         /* We're good to go, update advertising data, parameters, and start
8752          * advertising.
8753          */
8754         cmd = mgmt_pending_new(sk, MGMT_OP_ADD_ADVERTISING, hdev, data,
8755                                data_len);
8756         if (!cmd) {
8757                 err = -ENOMEM;
8758                 goto unlock;
8759         }
8760
8761         cp->instance = schedule_instance;
8762
8763         err = hci_cmd_sync_queue(hdev, add_advertising_sync, cmd,
8764                                  add_advertising_complete);
8765         if (err < 0)
8766                 mgmt_pending_free(cmd);
8767
8768 unlock:
8769         hci_dev_unlock(hdev);
8770
8771         return err;
8772 }
8773
8774 static void add_ext_adv_params_complete(struct hci_dev *hdev, void *data,
8775                                         int err)
8776 {
8777         struct mgmt_pending_cmd *cmd = data;
8778         struct mgmt_cp_add_ext_adv_params *cp = cmd->param;
8779         struct mgmt_rp_add_ext_adv_params rp;
8780         struct adv_info *adv;
8781         u32 flags;
8782
8783         BT_DBG("%s", hdev->name);
8784
8785         hci_dev_lock(hdev);
8786
8787         adv = hci_find_adv_instance(hdev, cp->instance);
8788         if (!adv)
8789                 goto unlock;
8790
8791         rp.instance = cp->instance;
8792         rp.tx_power = adv->tx_power;
8793
8794         /* While we're at it, inform userspace of the available space for this
8795          * advertisement, given the flags that will be used.
8796          */
8797         flags = __le32_to_cpu(cp->flags);
8798         rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true);
8799         rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false);
8800
8801         if (err) {
8802                 /* If this advertisement was previously advertising and we
8803                  * failed to update it, we signal that it has been removed and
8804                  * delete its structure
8805                  */
8806                 if (!adv->pending)
8807                         mgmt_advertising_removed(cmd->sk, hdev, cp->instance);
8808
8809                 hci_remove_adv_instance(hdev, cp->instance);
8810
8811                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode,
8812                                 mgmt_status(err));
8813         } else {
8814                 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode,
8815                                   mgmt_status(err), &rp, sizeof(rp));
8816         }
8817
8818 unlock:
8819         if (cmd)
8820                 mgmt_pending_free(cmd);
8821
8822         hci_dev_unlock(hdev);
8823 }
8824
8825 static int add_ext_adv_params_sync(struct hci_dev *hdev, void *data)
8826 {
8827         struct mgmt_pending_cmd *cmd = data;
8828         struct mgmt_cp_add_ext_adv_params *cp = cmd->param;
8829
8830         return hci_setup_ext_adv_instance_sync(hdev, cp->instance);
8831 }
8832
8833 static int add_ext_adv_params(struct sock *sk, struct hci_dev *hdev,
8834                               void *data, u16 data_len)
8835 {
8836         struct mgmt_cp_add_ext_adv_params *cp = data;
8837         struct mgmt_rp_add_ext_adv_params rp;
8838         struct mgmt_pending_cmd *cmd = NULL;
8839         struct adv_info *adv;
8840         u32 flags, min_interval, max_interval;
8841         u16 timeout, duration;
8842         u8 status;
8843         s8 tx_power;
8844         int err;
8845
8846         BT_DBG("%s", hdev->name);
8847
8848         status = mgmt_le_support(hdev);
8849         if (status)
8850                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS,
8851                                        status);
8852
8853         if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets)
8854                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS,
8855                                        MGMT_STATUS_INVALID_PARAMS);
8856
8857         /* The purpose of breaking add_advertising into two separate MGMT calls
8858          * for params and data is to allow more parameters to be added to this
8859          * structure in the future. For this reason, we verify that we have the
8860          * bare minimum structure we know of when the interface was defined. Any
8861          * extra parameters we don't know about will be ignored in this request.
8862          */
8863         if (data_len < MGMT_ADD_EXT_ADV_PARAMS_MIN_SIZE)
8864                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS,
8865                                        MGMT_STATUS_INVALID_PARAMS);
8866
8867         flags = __le32_to_cpu(cp->flags);
8868
8869         if (!requested_adv_flags_are_valid(hdev, flags))
8870                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS,
8871                                        MGMT_STATUS_INVALID_PARAMS);
8872
8873         hci_dev_lock(hdev);
8874
8875         /* In new interface, we require that we are powered to register */
8876         if (!hdev_is_powered(hdev)) {
8877                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS,
8878                                       MGMT_STATUS_REJECTED);
8879                 goto unlock;
8880         }
8881
8882         if (adv_busy(hdev)) {
8883                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS,
8884                                       MGMT_STATUS_BUSY);
8885                 goto unlock;
8886         }
8887
8888         /* Parse defined parameters from request, use defaults otherwise */
8889         timeout = (flags & MGMT_ADV_PARAM_TIMEOUT) ?
8890                   __le16_to_cpu(cp->timeout) : 0;
8891
8892         duration = (flags & MGMT_ADV_PARAM_DURATION) ?
8893                    __le16_to_cpu(cp->duration) :
8894                    hdev->def_multi_adv_rotation_duration;
8895
8896         min_interval = (flags & MGMT_ADV_PARAM_INTERVALS) ?
8897                        __le32_to_cpu(cp->min_interval) :
8898                        hdev->le_adv_min_interval;
8899
8900         max_interval = (flags & MGMT_ADV_PARAM_INTERVALS) ?
8901                        __le32_to_cpu(cp->max_interval) :
8902                        hdev->le_adv_max_interval;
8903
8904         tx_power = (flags & MGMT_ADV_PARAM_TX_POWER) ?
8905                    cp->tx_power :
8906                    HCI_ADV_TX_POWER_NO_PREFERENCE;
8907
8908         /* Create advertising instance with no advertising or response data */
8909         adv = hci_add_adv_instance(hdev, cp->instance, flags, 0, NULL, 0, NULL,
8910                                    timeout, duration, tx_power, min_interval,
8911                                    max_interval, 0);
8912
8913         if (IS_ERR(adv)) {
8914                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_PARAMS,
8915                                       MGMT_STATUS_FAILED);
8916                 goto unlock;
8917         }
8918
8919         /* Submit request for advertising params if ext adv available */
8920         if (ext_adv_capable(hdev)) {
8921                 cmd = mgmt_pending_new(sk, MGMT_OP_ADD_EXT_ADV_PARAMS, hdev,
8922                                        data, data_len);
8923                 if (!cmd) {
8924                         err = -ENOMEM;
8925                         hci_remove_adv_instance(hdev, cp->instance);
8926                         goto unlock;
8927                 }
8928
8929                 err = hci_cmd_sync_queue(hdev, add_ext_adv_params_sync, cmd,
8930                                          add_ext_adv_params_complete);
8931                 if (err < 0)
8932                         mgmt_pending_free(cmd);
8933         } else {
8934                 rp.instance = cp->instance;
8935                 rp.tx_power = HCI_ADV_TX_POWER_NO_PREFERENCE;
8936                 rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true);
8937                 rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false);
8938                 err = mgmt_cmd_complete(sk, hdev->id,
8939                                         MGMT_OP_ADD_EXT_ADV_PARAMS,
8940                                         MGMT_STATUS_SUCCESS, &rp, sizeof(rp));
8941         }
8942
8943 unlock:
8944         hci_dev_unlock(hdev);
8945
8946         return err;
8947 }
8948
8949 static void add_ext_adv_data_complete(struct hci_dev *hdev, void *data, int err)
8950 {
8951         struct mgmt_pending_cmd *cmd = data;
8952         struct mgmt_cp_add_ext_adv_data *cp = cmd->param;
8953         struct mgmt_rp_add_advertising rp;
8954
8955         add_adv_complete(hdev, cmd->sk, cp->instance, err);
8956
8957         memset(&rp, 0, sizeof(rp));
8958
8959         rp.instance = cp->instance;
8960
8961         if (err)
8962                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode,
8963                                 mgmt_status(err));
8964         else
8965                 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode,
8966                                   mgmt_status(err), &rp, sizeof(rp));
8967
8968         mgmt_pending_free(cmd);
8969 }
8970
8971 static int add_ext_adv_data_sync(struct hci_dev *hdev, void *data)
8972 {
8973         struct mgmt_pending_cmd *cmd = data;
8974         struct mgmt_cp_add_ext_adv_data *cp = cmd->param;
8975         int err;
8976
8977         if (ext_adv_capable(hdev)) {
8978                 err = hci_update_adv_data_sync(hdev, cp->instance);
8979                 if (err)
8980                         return err;
8981
8982                 err = hci_update_scan_rsp_data_sync(hdev, cp->instance);
8983                 if (err)
8984                         return err;
8985
8986                 return hci_enable_ext_advertising_sync(hdev, cp->instance);
8987         }
8988
8989         return hci_schedule_adv_instance_sync(hdev, cp->instance, true);
8990 }
8991
8992 static int add_ext_adv_data(struct sock *sk, struct hci_dev *hdev, void *data,
8993                             u16 data_len)
8994 {
8995         struct mgmt_cp_add_ext_adv_data *cp = data;
8996         struct mgmt_rp_add_ext_adv_data rp;
8997         u8 schedule_instance = 0;
8998         struct adv_info *next_instance;
8999         struct adv_info *adv_instance;
9000         int err = 0;
9001         struct mgmt_pending_cmd *cmd;
9002
9003         BT_DBG("%s", hdev->name);
9004
9005         hci_dev_lock(hdev);
9006
9007         adv_instance = hci_find_adv_instance(hdev, cp->instance);
9008
9009         if (!adv_instance) {
9010                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA,
9011                                       MGMT_STATUS_INVALID_PARAMS);
9012                 goto unlock;
9013         }
9014
9015         /* In new interface, we require that we are powered to register */
9016         if (!hdev_is_powered(hdev)) {
9017                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA,
9018                                       MGMT_STATUS_REJECTED);
9019                 goto clear_new_instance;
9020         }
9021
9022         if (adv_busy(hdev)) {
9023                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA,
9024                                       MGMT_STATUS_BUSY);
9025                 goto clear_new_instance;
9026         }
9027
9028         /* Validate new data */
9029         if (!tlv_data_is_valid(hdev, adv_instance->flags, cp->data,
9030                                cp->adv_data_len, true) ||
9031             !tlv_data_is_valid(hdev, adv_instance->flags, cp->data +
9032                                cp->adv_data_len, cp->scan_rsp_len, false)) {
9033                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA,
9034                                       MGMT_STATUS_INVALID_PARAMS);
9035                 goto clear_new_instance;
9036         }
9037
9038         /* Set the data in the advertising instance */
9039         hci_set_adv_instance_data(hdev, cp->instance, cp->adv_data_len,
9040                                   cp->data, cp->scan_rsp_len,
9041                                   cp->data + cp->adv_data_len);
9042
9043         /* If using software rotation, determine next instance to use */
9044         if (hdev->cur_adv_instance == cp->instance) {
9045                 /* If the currently advertised instance is being changed
9046                  * then cancel the current advertising and schedule the
9047                  * next instance. If there is only one instance then the
9048                  * overridden advertising data will be visible right
9049                  * away
9050                  */
9051                 cancel_adv_timeout(hdev);
9052
9053                 next_instance = hci_get_next_instance(hdev, cp->instance);
9054                 if (next_instance)
9055                         schedule_instance = next_instance->instance;
9056         } else if (!hdev->adv_instance_timeout) {
9057                 /* Immediately advertise the new instance if no other
9058                  * instance is currently being advertised.
9059                  */
9060                 schedule_instance = cp->instance;
9061         }
9062
9063         /* If the HCI_ADVERTISING flag is set or there is no instance to
9064          * be advertised then we have no HCI communication to make.
9065          * Simply return.
9066          */
9067         if (hci_dev_test_flag(hdev, HCI_ADVERTISING) || !schedule_instance) {
9068                 if (adv_instance->pending) {
9069                         mgmt_advertising_added(sk, hdev, cp->instance);
9070                         adv_instance->pending = false;
9071                 }
9072                 rp.instance = cp->instance;
9073                 err = mgmt_cmd_complete(sk, hdev->id, MGMT_OP_ADD_EXT_ADV_DATA,
9074                                         MGMT_STATUS_SUCCESS, &rp, sizeof(rp));
9075                 goto unlock;
9076         }
9077
9078         cmd = mgmt_pending_new(sk, MGMT_OP_ADD_EXT_ADV_DATA, hdev, data,
9079                                data_len);
9080         if (!cmd) {
9081                 err = -ENOMEM;
9082                 goto clear_new_instance;
9083         }
9084
9085         err = hci_cmd_sync_queue(hdev, add_ext_adv_data_sync, cmd,
9086                                  add_ext_adv_data_complete);
9087         if (err < 0) {
9088                 mgmt_pending_free(cmd);
9089                 goto clear_new_instance;
9090         }
9091
9092         /* We were successful in updating data, so trigger advertising_added
9093          * event if this is an instance that wasn't previously advertising. If
9094          * a failure occurs in the requests we initiated, we will remove the
9095          * instance again in add_advertising_complete
9096          */
9097         if (adv_instance->pending)
9098                 mgmt_advertising_added(sk, hdev, cp->instance);
9099
9100         goto unlock;
9101
9102 clear_new_instance:
9103         hci_remove_adv_instance(hdev, cp->instance);
9104
9105 unlock:
9106         hci_dev_unlock(hdev);
9107
9108         return err;
9109 }
9110
9111 static void remove_advertising_complete(struct hci_dev *hdev, void *data,
9112                                         int err)
9113 {
9114         struct mgmt_pending_cmd *cmd = data;
9115         struct mgmt_cp_remove_advertising *cp = cmd->param;
9116         struct mgmt_rp_remove_advertising rp;
9117
9118         bt_dev_dbg(hdev, "err %d", err);
9119
9120         memset(&rp, 0, sizeof(rp));
9121         rp.instance = cp->instance;
9122
9123         if (err)
9124                 mgmt_cmd_status(cmd->sk, cmd->index, cmd->opcode,
9125                                 mgmt_status(err));
9126         else
9127                 mgmt_cmd_complete(cmd->sk, cmd->index, cmd->opcode,
9128                                   MGMT_STATUS_SUCCESS, &rp, sizeof(rp));
9129
9130         mgmt_pending_free(cmd);
9131 }
9132
9133 static int remove_advertising_sync(struct hci_dev *hdev, void *data)
9134 {
9135         struct mgmt_pending_cmd *cmd = data;
9136         struct mgmt_cp_remove_advertising *cp = cmd->param;
9137         int err;
9138
9139         err = hci_remove_advertising_sync(hdev, cmd->sk, cp->instance, true);
9140         if (err)
9141                 return err;
9142
9143         if (list_empty(&hdev->adv_instances))
9144                 err = hci_disable_advertising_sync(hdev);
9145
9146         return err;
9147 }
9148
9149 static int remove_advertising(struct sock *sk, struct hci_dev *hdev,
9150                               void *data, u16 data_len)
9151 {
9152         struct mgmt_cp_remove_advertising *cp = data;
9153         struct mgmt_pending_cmd *cmd;
9154         int err;
9155
9156         bt_dev_dbg(hdev, "sock %p", sk);
9157
9158         hci_dev_lock(hdev);
9159
9160         if (cp->instance && !hci_find_adv_instance(hdev, cp->instance)) {
9161                 err = mgmt_cmd_status(sk, hdev->id,
9162                                       MGMT_OP_REMOVE_ADVERTISING,
9163                                       MGMT_STATUS_INVALID_PARAMS);
9164                 goto unlock;
9165         }
9166
9167         if (pending_find(MGMT_OP_SET_LE, hdev)) {
9168                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADVERTISING,
9169                                       MGMT_STATUS_BUSY);
9170                 goto unlock;
9171         }
9172
9173         if (list_empty(&hdev->adv_instances)) {
9174                 err = mgmt_cmd_status(sk, hdev->id, MGMT_OP_REMOVE_ADVERTISING,
9175                                       MGMT_STATUS_INVALID_PARAMS);
9176                 goto unlock;
9177         }
9178
9179         cmd = mgmt_pending_new(sk, MGMT_OP_REMOVE_ADVERTISING, hdev, data,
9180                                data_len);
9181         if (!cmd) {
9182                 err = -ENOMEM;
9183                 goto unlock;
9184         }
9185
9186         err = hci_cmd_sync_queue(hdev, remove_advertising_sync, cmd,
9187                                  remove_advertising_complete);
9188         if (err < 0)
9189                 mgmt_pending_free(cmd);
9190
9191 unlock:
9192         hci_dev_unlock(hdev);
9193
9194         return err;
9195 }
9196
9197 static int get_adv_size_info(struct sock *sk, struct hci_dev *hdev,
9198                              void *data, u16 data_len)
9199 {
9200         struct mgmt_cp_get_adv_size_info *cp = data;
9201         struct mgmt_rp_get_adv_size_info rp;
9202         u32 flags, supported_flags;
9203
9204         bt_dev_dbg(hdev, "sock %p", sk);
9205
9206         if (!lmp_le_capable(hdev))
9207                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO,
9208                                        MGMT_STATUS_REJECTED);
9209
9210         if (cp->instance < 1 || cp->instance > hdev->le_num_of_adv_sets)
9211                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO,
9212                                        MGMT_STATUS_INVALID_PARAMS);
9213
9214         flags = __le32_to_cpu(cp->flags);
9215
9216         /* The current implementation only supports a subset of the specified
9217          * flags.
9218          */
9219         supported_flags = get_supported_adv_flags(hdev);
9220         if (flags & ~supported_flags)
9221                 return mgmt_cmd_status(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO,
9222                                        MGMT_STATUS_INVALID_PARAMS);
9223
9224         rp.instance = cp->instance;
9225         rp.flags = cp->flags;
9226         rp.max_adv_data_len = tlv_data_max_len(hdev, flags, true);
9227         rp.max_scan_rsp_len = tlv_data_max_len(hdev, flags, false);
9228
9229         return mgmt_cmd_complete(sk, hdev->id, MGMT_OP_GET_ADV_SIZE_INFO,
9230                                  MGMT_STATUS_SUCCESS, &rp, sizeof(rp));
9231 }
9232
9233 static const struct hci_mgmt_handler mgmt_handlers[] = {
9234         { NULL }, /* 0x0000 (no command) */
9235         { read_version,            MGMT_READ_VERSION_SIZE,
9236                                                 HCI_MGMT_NO_HDEV |
9237                                                 HCI_MGMT_UNTRUSTED },
9238         { read_commands,           MGMT_READ_COMMANDS_SIZE,
9239                                                 HCI_MGMT_NO_HDEV |
9240                                                 HCI_MGMT_UNTRUSTED },
9241         { read_index_list,         MGMT_READ_INDEX_LIST_SIZE,
9242                                                 HCI_MGMT_NO_HDEV |
9243                                                 HCI_MGMT_UNTRUSTED },
9244         { read_controller_info,    MGMT_READ_INFO_SIZE,
9245                                                 HCI_MGMT_UNTRUSTED },
9246         { set_powered,             MGMT_SETTING_SIZE },
9247         { set_discoverable,        MGMT_SET_DISCOVERABLE_SIZE },
9248         { set_connectable,         MGMT_SETTING_SIZE },
9249         { set_fast_connectable,    MGMT_SETTING_SIZE },
9250         { set_bondable,            MGMT_SETTING_SIZE },
9251         { set_link_security,       MGMT_SETTING_SIZE },
9252         { set_ssp,                 MGMT_SETTING_SIZE },
9253         { set_hs,                  MGMT_SETTING_SIZE },
9254         { set_le,                  MGMT_SETTING_SIZE },
9255         { set_dev_class,           MGMT_SET_DEV_CLASS_SIZE },
9256         { set_local_name,          MGMT_SET_LOCAL_NAME_SIZE },
9257         { add_uuid,                MGMT_ADD_UUID_SIZE },
9258         { remove_uuid,             MGMT_REMOVE_UUID_SIZE },
9259         { load_link_keys,          MGMT_LOAD_LINK_KEYS_SIZE,
9260                                                 HCI_MGMT_VAR_LEN },
9261         { load_long_term_keys,     MGMT_LOAD_LONG_TERM_KEYS_SIZE,
9262                                                 HCI_MGMT_VAR_LEN },
9263         { disconnect,              MGMT_DISCONNECT_SIZE },
9264         { get_connections,         MGMT_GET_CONNECTIONS_SIZE },
9265         { pin_code_reply,          MGMT_PIN_CODE_REPLY_SIZE },
9266         { pin_code_neg_reply,      MGMT_PIN_CODE_NEG_REPLY_SIZE },
9267         { set_io_capability,       MGMT_SET_IO_CAPABILITY_SIZE },
9268         { pair_device,             MGMT_PAIR_DEVICE_SIZE },
9269         { cancel_pair_device,      MGMT_CANCEL_PAIR_DEVICE_SIZE },
9270         { unpair_device,           MGMT_UNPAIR_DEVICE_SIZE },
9271         { user_confirm_reply,      MGMT_USER_CONFIRM_REPLY_SIZE },
9272         { user_confirm_neg_reply,  MGMT_USER_CONFIRM_NEG_REPLY_SIZE },
9273         { user_passkey_reply,      MGMT_USER_PASSKEY_REPLY_SIZE },
9274         { user_passkey_neg_reply,  MGMT_USER_PASSKEY_NEG_REPLY_SIZE },
9275         { read_local_oob_data,     MGMT_READ_LOCAL_OOB_DATA_SIZE },
9276         { add_remote_oob_data,     MGMT_ADD_REMOTE_OOB_DATA_SIZE,
9277                                                 HCI_MGMT_VAR_LEN },
9278         { remove_remote_oob_data,  MGMT_REMOVE_REMOTE_OOB_DATA_SIZE },
9279         { start_discovery,         MGMT_START_DISCOVERY_SIZE },
9280         { stop_discovery,          MGMT_STOP_DISCOVERY_SIZE },
9281         { confirm_name,            MGMT_CONFIRM_NAME_SIZE },
9282         { block_device,            MGMT_BLOCK_DEVICE_SIZE },
9283         { unblock_device,          MGMT_UNBLOCK_DEVICE_SIZE },
9284         { set_device_id,           MGMT_SET_DEVICE_ID_SIZE },
9285         { set_advertising,         MGMT_SETTING_SIZE },
9286         { set_bredr,               MGMT_SETTING_SIZE },
9287         { set_static_address,      MGMT_SET_STATIC_ADDRESS_SIZE },
9288         { set_scan_params,         MGMT_SET_SCAN_PARAMS_SIZE },
9289         { set_secure_conn,         MGMT_SETTING_SIZE },
9290         { set_debug_keys,          MGMT_SETTING_SIZE },
9291         { set_privacy,             MGMT_SET_PRIVACY_SIZE },
9292         { load_irks,               MGMT_LOAD_IRKS_SIZE,
9293                                                 HCI_MGMT_VAR_LEN },
9294         { get_conn_info,           MGMT_GET_CONN_INFO_SIZE },
9295         { get_clock_info,          MGMT_GET_CLOCK_INFO_SIZE },
9296         { add_device,              MGMT_ADD_DEVICE_SIZE },
9297         { remove_device,           MGMT_REMOVE_DEVICE_SIZE },
9298         { load_conn_param,         MGMT_LOAD_CONN_PARAM_SIZE,
9299                                                 HCI_MGMT_VAR_LEN },
9300         { read_unconf_index_list,  MGMT_READ_UNCONF_INDEX_LIST_SIZE,
9301                                                 HCI_MGMT_NO_HDEV |
9302                                                 HCI_MGMT_UNTRUSTED },
9303         { read_config_info,        MGMT_READ_CONFIG_INFO_SIZE,
9304                                                 HCI_MGMT_UNCONFIGURED |
9305                                                 HCI_MGMT_UNTRUSTED },
9306         { set_external_config,     MGMT_SET_EXTERNAL_CONFIG_SIZE,
9307                                                 HCI_MGMT_UNCONFIGURED },
9308         { set_public_address,      MGMT_SET_PUBLIC_ADDRESS_SIZE,
9309                                                 HCI_MGMT_UNCONFIGURED },
9310         { start_service_discovery, MGMT_START_SERVICE_DISCOVERY_SIZE,
9311                                                 HCI_MGMT_VAR_LEN },
9312         { read_local_oob_ext_data, MGMT_READ_LOCAL_OOB_EXT_DATA_SIZE },
9313         { read_ext_index_list,     MGMT_READ_EXT_INDEX_LIST_SIZE,
9314                                                 HCI_MGMT_NO_HDEV |
9315                                                 HCI_MGMT_UNTRUSTED },
9316         { read_adv_features,       MGMT_READ_ADV_FEATURES_SIZE },
9317         { add_advertising,         MGMT_ADD_ADVERTISING_SIZE,
9318                                                 HCI_MGMT_VAR_LEN },
9319         { remove_advertising,      MGMT_REMOVE_ADVERTISING_SIZE },
9320         { get_adv_size_info,       MGMT_GET_ADV_SIZE_INFO_SIZE },
9321         { start_limited_discovery, MGMT_START_DISCOVERY_SIZE },
9322         { read_ext_controller_info,MGMT_READ_EXT_INFO_SIZE,
9323                                                 HCI_MGMT_UNTRUSTED },
9324         { set_appearance,          MGMT_SET_APPEARANCE_SIZE },
9325         { get_phy_configuration,   MGMT_GET_PHY_CONFIGURATION_SIZE },
9326         { set_phy_configuration,   MGMT_SET_PHY_CONFIGURATION_SIZE },
9327         { set_blocked_keys,        MGMT_OP_SET_BLOCKED_KEYS_SIZE,
9328                                                 HCI_MGMT_VAR_LEN },
9329         { set_wideband_speech,     MGMT_SETTING_SIZE },
9330         { read_controller_cap,     MGMT_READ_CONTROLLER_CAP_SIZE,
9331                                                 HCI_MGMT_UNTRUSTED },
9332         { read_exp_features_info,  MGMT_READ_EXP_FEATURES_INFO_SIZE,
9333                                                 HCI_MGMT_UNTRUSTED |
9334                                                 HCI_MGMT_HDEV_OPTIONAL },
9335         { set_exp_feature,         MGMT_SET_EXP_FEATURE_SIZE,
9336                                                 HCI_MGMT_VAR_LEN |
9337                                                 HCI_MGMT_HDEV_OPTIONAL },
9338         { read_def_system_config,  MGMT_READ_DEF_SYSTEM_CONFIG_SIZE,
9339                                                 HCI_MGMT_UNTRUSTED },
9340         { set_def_system_config,   MGMT_SET_DEF_SYSTEM_CONFIG_SIZE,
9341                                                 HCI_MGMT_VAR_LEN },
9342         { read_def_runtime_config, MGMT_READ_DEF_RUNTIME_CONFIG_SIZE,
9343                                                 HCI_MGMT_UNTRUSTED },
9344         { set_def_runtime_config,  MGMT_SET_DEF_RUNTIME_CONFIG_SIZE,
9345                                                 HCI_MGMT_VAR_LEN },
9346         { get_device_flags,        MGMT_GET_DEVICE_FLAGS_SIZE },
9347         { set_device_flags,        MGMT_SET_DEVICE_FLAGS_SIZE },
9348         { read_adv_mon_features,   MGMT_READ_ADV_MONITOR_FEATURES_SIZE },
9349         { add_adv_patterns_monitor,MGMT_ADD_ADV_PATTERNS_MONITOR_SIZE,
9350                                                 HCI_MGMT_VAR_LEN },
9351         { remove_adv_monitor,      MGMT_REMOVE_ADV_MONITOR_SIZE },
9352         { add_ext_adv_params,      MGMT_ADD_EXT_ADV_PARAMS_MIN_SIZE,
9353                                                 HCI_MGMT_VAR_LEN },
9354         { add_ext_adv_data,        MGMT_ADD_EXT_ADV_DATA_SIZE,
9355                                                 HCI_MGMT_VAR_LEN },
9356         { add_adv_patterns_monitor_rssi,
9357                                    MGMT_ADD_ADV_PATTERNS_MONITOR_RSSI_SIZE,
9358                                                 HCI_MGMT_VAR_LEN },
9359         { set_mesh,                MGMT_SET_MESH_RECEIVER_SIZE,
9360                                                 HCI_MGMT_VAR_LEN },
9361         { mesh_features,           MGMT_MESH_READ_FEATURES_SIZE },
9362         { mesh_send,               MGMT_MESH_SEND_SIZE,
9363                                                 HCI_MGMT_VAR_LEN },
9364         { mesh_send_cancel,        MGMT_MESH_SEND_CANCEL_SIZE },
9365 };
9366
9367 void mgmt_index_added(struct hci_dev *hdev)
9368 {
9369         struct mgmt_ev_ext_index ev;
9370
9371         if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks))
9372                 return;
9373
9374         switch (hdev->dev_type) {
9375         case HCI_PRIMARY:
9376                 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) {
9377                         mgmt_index_event(MGMT_EV_UNCONF_INDEX_ADDED, hdev,
9378                                          NULL, 0, HCI_MGMT_UNCONF_INDEX_EVENTS);
9379                         ev.type = 0x01;
9380                 } else {
9381                         mgmt_index_event(MGMT_EV_INDEX_ADDED, hdev, NULL, 0,
9382                                          HCI_MGMT_INDEX_EVENTS);
9383                         ev.type = 0x00;
9384                 }
9385                 break;
9386         case HCI_AMP:
9387                 ev.type = 0x02;
9388                 break;
9389         default:
9390                 return;
9391         }
9392
9393         ev.bus = hdev->bus;
9394
9395         mgmt_index_event(MGMT_EV_EXT_INDEX_ADDED, hdev, &ev, sizeof(ev),
9396                          HCI_MGMT_EXT_INDEX_EVENTS);
9397 }
9398
9399 void mgmt_index_removed(struct hci_dev *hdev)
9400 {
9401         struct mgmt_ev_ext_index ev;
9402         u8 status = MGMT_STATUS_INVALID_INDEX;
9403
9404         if (test_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks))
9405                 return;
9406
9407         switch (hdev->dev_type) {
9408         case HCI_PRIMARY:
9409                 mgmt_pending_foreach(0, hdev, cmd_complete_rsp, &status);
9410
9411                 if (hci_dev_test_flag(hdev, HCI_UNCONFIGURED)) {
9412                         mgmt_index_event(MGMT_EV_UNCONF_INDEX_REMOVED, hdev,
9413                                          NULL, 0, HCI_MGMT_UNCONF_INDEX_EVENTS);
9414                         ev.type = 0x01;
9415                 } else {
9416                         mgmt_index_event(MGMT_EV_INDEX_REMOVED, hdev, NULL, 0,
9417                                          HCI_MGMT_INDEX_EVENTS);
9418                         ev.type = 0x00;
9419                 }
9420                 break;
9421         case HCI_AMP:
9422                 ev.type = 0x02;
9423                 break;
9424         default:
9425                 return;
9426         }
9427
9428         ev.bus = hdev->bus;
9429
9430         mgmt_index_event(MGMT_EV_EXT_INDEX_REMOVED, hdev, &ev, sizeof(ev),
9431                          HCI_MGMT_EXT_INDEX_EVENTS);
9432
9433         /* Cancel any remaining timed work */
9434         if (!hci_dev_test_flag(hdev, HCI_MGMT))
9435                 return;
9436         cancel_delayed_work_sync(&hdev->discov_off);
9437         cancel_delayed_work_sync(&hdev->service_cache);
9438         cancel_delayed_work_sync(&hdev->rpa_expired);
9439 }
9440
9441 void mgmt_power_on(struct hci_dev *hdev, int err)
9442 {
9443         struct cmd_lookup match = { NULL, hdev };
9444
9445         bt_dev_dbg(hdev, "err %d", err);
9446
9447         hci_dev_lock(hdev);
9448
9449         if (!err) {
9450                 restart_le_actions(hdev);
9451                 hci_update_passive_scan(hdev);
9452         }
9453
9454         mgmt_pending_foreach(MGMT_OP_SET_POWERED, hdev, settings_rsp, &match);
9455
9456         new_settings(hdev, match.sk);
9457
9458         if (match.sk)
9459                 sock_put(match.sk);
9460
9461         hci_dev_unlock(hdev);
9462 }
9463
9464 void __mgmt_power_off(struct hci_dev *hdev)
9465 {
9466         struct cmd_lookup match = { NULL, hdev };
9467         u8 status, zero_cod[] = { 0, 0, 0 };
9468
9469         mgmt_pending_foreach(MGMT_OP_SET_POWERED, hdev, settings_rsp, &match);
9470
9471         /* If the power off is because of hdev unregistration let
9472          * use the appropriate INVALID_INDEX status. Otherwise use
9473          * NOT_POWERED. We cover both scenarios here since later in
9474          * mgmt_index_removed() any hci_conn callbacks will have already
9475          * been triggered, potentially causing misleading DISCONNECTED
9476          * status responses.
9477          */
9478         if (hci_dev_test_flag(hdev, HCI_UNREGISTER))
9479                 status = MGMT_STATUS_INVALID_INDEX;
9480         else
9481                 status = MGMT_STATUS_NOT_POWERED;
9482
9483         mgmt_pending_foreach(0, hdev, cmd_complete_rsp, &status);
9484
9485         if (memcmp(hdev->dev_class, zero_cod, sizeof(zero_cod)) != 0) {
9486                 mgmt_limited_event(MGMT_EV_CLASS_OF_DEV_CHANGED, hdev,
9487                                    zero_cod, sizeof(zero_cod),
9488                                    HCI_MGMT_DEV_CLASS_EVENTS, NULL);
9489                 ext_info_changed(hdev, NULL);
9490         }
9491
9492         new_settings(hdev, match.sk);
9493
9494         if (match.sk)
9495                 sock_put(match.sk);
9496 }
9497
9498 void mgmt_set_powered_failed(struct hci_dev *hdev, int err)
9499 {
9500         struct mgmt_pending_cmd *cmd;
9501         u8 status;
9502
9503         cmd = pending_find(MGMT_OP_SET_POWERED, hdev);
9504         if (!cmd)
9505                 return;
9506
9507         if (err == -ERFKILL)
9508                 status = MGMT_STATUS_RFKILLED;
9509         else
9510                 status = MGMT_STATUS_FAILED;
9511
9512         mgmt_cmd_status(cmd->sk, hdev->id, MGMT_OP_SET_POWERED, status);
9513
9514         mgmt_pending_remove(cmd);
9515 }
9516
9517 void mgmt_new_link_key(struct hci_dev *hdev, struct link_key *key,
9518                        bool persistent)
9519 {
9520         struct mgmt_ev_new_link_key ev;
9521
9522         memset(&ev, 0, sizeof(ev));
9523
9524         ev.store_hint = persistent;
9525         bacpy(&ev.key.addr.bdaddr, &key->bdaddr);
9526         ev.key.addr.type = BDADDR_BREDR;
9527         ev.key.type = key->type;
9528         memcpy(ev.key.val, key->val, HCI_LINK_KEY_SIZE);
9529         ev.key.pin_len = key->pin_len;
9530
9531         mgmt_event(MGMT_EV_NEW_LINK_KEY, hdev, &ev, sizeof(ev), NULL);
9532 }
9533
9534 static u8 mgmt_ltk_type(struct smp_ltk *ltk)
9535 {
9536         switch (ltk->type) {
9537         case SMP_LTK:
9538         case SMP_LTK_RESPONDER:
9539                 if (ltk->authenticated)
9540                         return MGMT_LTK_AUTHENTICATED;
9541                 return MGMT_LTK_UNAUTHENTICATED;
9542         case SMP_LTK_P256:
9543                 if (ltk->authenticated)
9544                         return MGMT_LTK_P256_AUTH;
9545                 return MGMT_LTK_P256_UNAUTH;
9546         case SMP_LTK_P256_DEBUG:
9547                 return MGMT_LTK_P256_DEBUG;
9548         }
9549
9550         return MGMT_LTK_UNAUTHENTICATED;
9551 }
9552
9553 void mgmt_new_ltk(struct hci_dev *hdev, struct smp_ltk *key, bool persistent)
9554 {
9555         struct mgmt_ev_new_long_term_key ev;
9556
9557         memset(&ev, 0, sizeof(ev));
9558
9559         /* Devices using resolvable or non-resolvable random addresses
9560          * without providing an identity resolving key don't require
9561          * to store long term keys. Their addresses will change the
9562          * next time around.
9563          *
9564          * Only when a remote device provides an identity address
9565          * make sure the long term key is stored. If the remote
9566          * identity is known, the long term keys are internally
9567          * mapped to the identity address. So allow static random
9568          * and public addresses here.
9569          */
9570         if (key->bdaddr_type == ADDR_LE_DEV_RANDOM &&
9571             (key->bdaddr.b[5] & 0xc0) != 0xc0)
9572                 ev.store_hint = 0x00;
9573         else
9574                 ev.store_hint = persistent;
9575
9576         bacpy(&ev.key.addr.bdaddr, &key->bdaddr);
9577         ev.key.addr.type = link_to_bdaddr(LE_LINK, key->bdaddr_type);
9578         ev.key.type = mgmt_ltk_type(key);
9579         ev.key.enc_size = key->enc_size;
9580         ev.key.ediv = key->ediv;
9581         ev.key.rand = key->rand;
9582
9583         if (key->type == SMP_LTK)
9584                 ev.key.initiator = 1;
9585
9586         /* Make sure we copy only the significant bytes based on the
9587          * encryption key size, and set the rest of the value to zeroes.
9588          */
9589         memcpy(ev.key.val, key->val, key->enc_size);
9590         memset(ev.key.val + key->enc_size, 0,
9591                sizeof(ev.key.val) - key->enc_size);
9592
9593         mgmt_event(MGMT_EV_NEW_LONG_TERM_KEY, hdev, &ev, sizeof(ev), NULL);
9594 }
9595
9596 void mgmt_new_irk(struct hci_dev *hdev, struct smp_irk *irk, bool persistent)
9597 {
9598         struct mgmt_ev_new_irk ev;
9599
9600         memset(&ev, 0, sizeof(ev));
9601
9602         ev.store_hint = persistent;
9603
9604         bacpy(&ev.rpa, &irk->rpa);
9605         bacpy(&ev.irk.addr.bdaddr, &irk->bdaddr);
9606         ev.irk.addr.type = link_to_bdaddr(LE_LINK, irk->addr_type);
9607         memcpy(ev.irk.val, irk->val, sizeof(irk->val));
9608
9609         mgmt_event(MGMT_EV_NEW_IRK, hdev, &ev, sizeof(ev), NULL);
9610 }
9611
9612 void mgmt_new_csrk(struct hci_dev *hdev, struct smp_csrk *csrk,
9613                    bool persistent)
9614 {
9615         struct mgmt_ev_new_csrk ev;
9616
9617         memset(&ev, 0, sizeof(ev));
9618
9619         /* Devices using resolvable or non-resolvable random addresses
9620          * without providing an identity resolving key don't require
9621          * to store signature resolving keys. Their addresses will change
9622          * the next time around.
9623          *
9624          * Only when a remote device provides an identity address
9625          * make sure the signature resolving key is stored. So allow
9626          * static random and public addresses here.
9627          */
9628         if (csrk->bdaddr_type == ADDR_LE_DEV_RANDOM &&
9629             (csrk->bdaddr.b[5] & 0xc0) != 0xc0)
9630                 ev.store_hint = 0x00;
9631         else
9632                 ev.store_hint = persistent;
9633
9634         bacpy(&ev.key.addr.bdaddr, &csrk->bdaddr);
9635         ev.key.addr.type = link_to_bdaddr(LE_LINK, csrk->bdaddr_type);
9636         ev.key.type = csrk->type;
9637         memcpy(ev.key.val, csrk->val, sizeof(csrk->val));
9638
9639         mgmt_event(MGMT_EV_NEW_CSRK, hdev, &ev, sizeof(ev), NULL);
9640 }
9641
9642 void mgmt_new_conn_param(struct hci_dev *hdev, bdaddr_t *bdaddr,
9643                          u8 bdaddr_type, u8 store_hint, u16 min_interval,
9644                          u16 max_interval, u16 latency, u16 timeout)
9645 {
9646         struct mgmt_ev_new_conn_param ev;
9647
9648         if (!hci_is_identity_address(bdaddr, bdaddr_type))
9649                 return;
9650
9651         memset(&ev, 0, sizeof(ev));
9652         bacpy(&ev.addr.bdaddr, bdaddr);
9653         ev.addr.type = link_to_bdaddr(LE_LINK, bdaddr_type);
9654         ev.store_hint = store_hint;
9655         ev.min_interval = cpu_to_le16(min_interval);
9656         ev.max_interval = cpu_to_le16(max_interval);
9657         ev.latency = cpu_to_le16(latency);
9658         ev.timeout = cpu_to_le16(timeout);
9659
9660         mgmt_event(MGMT_EV_NEW_CONN_PARAM, hdev, &ev, sizeof(ev), NULL);
9661 }
9662
9663 void mgmt_device_connected(struct hci_dev *hdev, struct hci_conn *conn,
9664                            u8 *name, u8 name_len)
9665 {
9666         struct sk_buff *skb;
9667         struct mgmt_ev_device_connected *ev;
9668         u16 eir_len = 0;
9669         u32 flags = 0;
9670
9671         /* allocate buff for LE or BR/EDR adv */
9672         if (conn->le_adv_data_len > 0)
9673                 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_CONNECTED,
9674                                      sizeof(*ev) + conn->le_adv_data_len);
9675         else
9676                 skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_CONNECTED,
9677                                      sizeof(*ev) + (name ? eir_precalc_len(name_len) : 0) +
9678                                      eir_precalc_len(sizeof(conn->dev_class)));
9679
9680         ev = skb_put(skb, sizeof(*ev));
9681         bacpy(&ev->addr.bdaddr, &conn->dst);
9682         ev->addr.type = link_to_bdaddr(conn->type, conn->dst_type);
9683
9684         if (conn->out)
9685                 flags |= MGMT_DEV_FOUND_INITIATED_CONN;
9686
9687         ev->flags = __cpu_to_le32(flags);
9688
9689         /* We must ensure that the EIR Data fields are ordered and
9690          * unique. Keep it simple for now and avoid the problem by not
9691          * adding any BR/EDR data to the LE adv.
9692          */
9693         if (conn->le_adv_data_len > 0) {
9694                 skb_put_data(skb, conn->le_adv_data, conn->le_adv_data_len);
9695                 eir_len = conn->le_adv_data_len;
9696         } else {
9697                 if (name)
9698                         eir_len += eir_skb_put_data(skb, EIR_NAME_COMPLETE, name, name_len);
9699
9700                 if (memcmp(conn->dev_class, "\0\0\0", sizeof(conn->dev_class)))
9701                         eir_len += eir_skb_put_data(skb, EIR_CLASS_OF_DEV,
9702                                                     conn->dev_class, sizeof(conn->dev_class));
9703         }
9704
9705         ev->eir_len = cpu_to_le16(eir_len);
9706
9707         mgmt_event_skb(skb, NULL);
9708 }
9709
9710 static void disconnect_rsp(struct mgmt_pending_cmd *cmd, void *data)
9711 {
9712         struct sock **sk = data;
9713
9714         cmd->cmd_complete(cmd, 0);
9715
9716         *sk = cmd->sk;
9717         sock_hold(*sk);
9718
9719         mgmt_pending_remove(cmd);
9720 }
9721
9722 static void unpair_device_rsp(struct mgmt_pending_cmd *cmd, void *data)
9723 {
9724         struct hci_dev *hdev = data;
9725         struct mgmt_cp_unpair_device *cp = cmd->param;
9726
9727         device_unpaired(hdev, &cp->addr.bdaddr, cp->addr.type, cmd->sk);
9728
9729         cmd->cmd_complete(cmd, 0);
9730         mgmt_pending_remove(cmd);
9731 }
9732
9733 bool mgmt_powering_down(struct hci_dev *hdev)
9734 {
9735         struct mgmt_pending_cmd *cmd;
9736         struct mgmt_mode *cp;
9737
9738         cmd = pending_find(MGMT_OP_SET_POWERED, hdev);
9739         if (!cmd)
9740                 return false;
9741
9742         cp = cmd->param;
9743         if (!cp->val)
9744                 return true;
9745
9746         return false;
9747 }
9748
9749 void mgmt_device_disconnected(struct hci_dev *hdev, bdaddr_t *bdaddr,
9750                               u8 link_type, u8 addr_type, u8 reason,
9751                               bool mgmt_connected)
9752 {
9753         struct mgmt_ev_device_disconnected ev;
9754         struct sock *sk = NULL;
9755
9756         /* The connection is still in hci_conn_hash so test for 1
9757          * instead of 0 to know if this is the last one.
9758          */
9759         if (mgmt_powering_down(hdev) && hci_conn_count(hdev) == 1) {
9760                 cancel_delayed_work(&hdev->power_off);
9761                 queue_work(hdev->req_workqueue, &hdev->power_off.work);
9762         }
9763
9764         if (!mgmt_connected)
9765                 return;
9766
9767         if (link_type != ACL_LINK && link_type != LE_LINK)
9768                 return;
9769
9770         mgmt_pending_foreach(MGMT_OP_DISCONNECT, hdev, disconnect_rsp, &sk);
9771
9772         bacpy(&ev.addr.bdaddr, bdaddr);
9773         ev.addr.type = link_to_bdaddr(link_type, addr_type);
9774         ev.reason = reason;
9775
9776         /* Report disconnects due to suspend */
9777         if (hdev->suspended)
9778                 ev.reason = MGMT_DEV_DISCONN_LOCAL_HOST_SUSPEND;
9779
9780         mgmt_event(MGMT_EV_DEVICE_DISCONNECTED, hdev, &ev, sizeof(ev), sk);
9781
9782         if (sk)
9783                 sock_put(sk);
9784
9785         mgmt_pending_foreach(MGMT_OP_UNPAIR_DEVICE, hdev, unpair_device_rsp,
9786                              hdev);
9787 }
9788
9789 void mgmt_disconnect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr,
9790                             u8 link_type, u8 addr_type, u8 status)
9791 {
9792         u8 bdaddr_type = link_to_bdaddr(link_type, addr_type);
9793         struct mgmt_cp_disconnect *cp;
9794         struct mgmt_pending_cmd *cmd;
9795
9796         mgmt_pending_foreach(MGMT_OP_UNPAIR_DEVICE, hdev, unpair_device_rsp,
9797                              hdev);
9798
9799         cmd = pending_find(MGMT_OP_DISCONNECT, hdev);
9800         if (!cmd)
9801                 return;
9802
9803         cp = cmd->param;
9804
9805         if (bacmp(bdaddr, &cp->addr.bdaddr))
9806                 return;
9807
9808         if (cp->addr.type != bdaddr_type)
9809                 return;
9810
9811         cmd->cmd_complete(cmd, mgmt_status(status));
9812         mgmt_pending_remove(cmd);
9813 }
9814
9815 void mgmt_connect_failed(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type,
9816                          u8 addr_type, u8 status)
9817 {
9818         struct mgmt_ev_connect_failed ev;
9819
9820         /* The connection is still in hci_conn_hash so test for 1
9821          * instead of 0 to know if this is the last one.
9822          */
9823         if (mgmt_powering_down(hdev) && hci_conn_count(hdev) == 1) {
9824                 cancel_delayed_work(&hdev->power_off);
9825                 queue_work(hdev->req_workqueue, &hdev->power_off.work);
9826         }
9827
9828         bacpy(&ev.addr.bdaddr, bdaddr);
9829         ev.addr.type = link_to_bdaddr(link_type, addr_type);
9830         ev.status = mgmt_status(status);
9831
9832         mgmt_event(MGMT_EV_CONNECT_FAILED, hdev, &ev, sizeof(ev), NULL);
9833 }
9834
9835 void mgmt_pin_code_request(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 secure)
9836 {
9837         struct mgmt_ev_pin_code_request ev;
9838
9839         bacpy(&ev.addr.bdaddr, bdaddr);
9840         ev.addr.type = BDADDR_BREDR;
9841         ev.secure = secure;
9842
9843         mgmt_event(MGMT_EV_PIN_CODE_REQUEST, hdev, &ev, sizeof(ev), NULL);
9844 }
9845
9846 void mgmt_pin_code_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
9847                                   u8 status)
9848 {
9849         struct mgmt_pending_cmd *cmd;
9850
9851         cmd = pending_find(MGMT_OP_PIN_CODE_REPLY, hdev);
9852         if (!cmd)
9853                 return;
9854
9855         cmd->cmd_complete(cmd, mgmt_status(status));
9856         mgmt_pending_remove(cmd);
9857 }
9858
9859 void mgmt_pin_code_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
9860                                       u8 status)
9861 {
9862         struct mgmt_pending_cmd *cmd;
9863
9864         cmd = pending_find(MGMT_OP_PIN_CODE_NEG_REPLY, hdev);
9865         if (!cmd)
9866                 return;
9867
9868         cmd->cmd_complete(cmd, mgmt_status(status));
9869         mgmt_pending_remove(cmd);
9870 }
9871
9872 int mgmt_user_confirm_request(struct hci_dev *hdev, bdaddr_t *bdaddr,
9873                               u8 link_type, u8 addr_type, u32 value,
9874                               u8 confirm_hint)
9875 {
9876         struct mgmt_ev_user_confirm_request ev;
9877
9878         bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr);
9879
9880         bacpy(&ev.addr.bdaddr, bdaddr);
9881         ev.addr.type = link_to_bdaddr(link_type, addr_type);
9882         ev.confirm_hint = confirm_hint;
9883         ev.value = cpu_to_le32(value);
9884
9885         return mgmt_event(MGMT_EV_USER_CONFIRM_REQUEST, hdev, &ev, sizeof(ev),
9886                           NULL);
9887 }
9888
9889 int mgmt_user_passkey_request(struct hci_dev *hdev, bdaddr_t *bdaddr,
9890                               u8 link_type, u8 addr_type)
9891 {
9892         struct mgmt_ev_user_passkey_request ev;
9893
9894         bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr);
9895
9896         bacpy(&ev.addr.bdaddr, bdaddr);
9897         ev.addr.type = link_to_bdaddr(link_type, addr_type);
9898
9899         return mgmt_event(MGMT_EV_USER_PASSKEY_REQUEST, hdev, &ev, sizeof(ev),
9900                           NULL);
9901 }
9902
9903 static int user_pairing_resp_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
9904                                       u8 link_type, u8 addr_type, u8 status,
9905                                       u8 opcode)
9906 {
9907         struct mgmt_pending_cmd *cmd;
9908
9909         cmd = pending_find(opcode, hdev);
9910         if (!cmd)
9911                 return -ENOENT;
9912
9913         cmd->cmd_complete(cmd, mgmt_status(status));
9914         mgmt_pending_remove(cmd);
9915
9916         return 0;
9917 }
9918
9919 int mgmt_user_confirm_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
9920                                      u8 link_type, u8 addr_type, u8 status)
9921 {
9922         return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type,
9923                                           status, MGMT_OP_USER_CONFIRM_REPLY);
9924 }
9925
9926 int mgmt_user_confirm_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
9927                                          u8 link_type, u8 addr_type, u8 status)
9928 {
9929         return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type,
9930                                           status,
9931                                           MGMT_OP_USER_CONFIRM_NEG_REPLY);
9932 }
9933
9934 int mgmt_user_passkey_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
9935                                      u8 link_type, u8 addr_type, u8 status)
9936 {
9937         return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type,
9938                                           status, MGMT_OP_USER_PASSKEY_REPLY);
9939 }
9940
9941 int mgmt_user_passkey_neg_reply_complete(struct hci_dev *hdev, bdaddr_t *bdaddr,
9942                                          u8 link_type, u8 addr_type, u8 status)
9943 {
9944         return user_pairing_resp_complete(hdev, bdaddr, link_type, addr_type,
9945                                           status,
9946                                           MGMT_OP_USER_PASSKEY_NEG_REPLY);
9947 }
9948
9949 int mgmt_user_passkey_notify(struct hci_dev *hdev, bdaddr_t *bdaddr,
9950                              u8 link_type, u8 addr_type, u32 passkey,
9951                              u8 entered)
9952 {
9953         struct mgmt_ev_passkey_notify ev;
9954
9955         bt_dev_dbg(hdev, "bdaddr %pMR", bdaddr);
9956
9957         bacpy(&ev.addr.bdaddr, bdaddr);
9958         ev.addr.type = link_to_bdaddr(link_type, addr_type);
9959         ev.passkey = __cpu_to_le32(passkey);
9960         ev.entered = entered;
9961
9962         return mgmt_event(MGMT_EV_PASSKEY_NOTIFY, hdev, &ev, sizeof(ev), NULL);
9963 }
9964
9965 void mgmt_auth_failed(struct hci_conn *conn, u8 hci_status)
9966 {
9967         struct mgmt_ev_auth_failed ev;
9968         struct mgmt_pending_cmd *cmd;
9969         u8 status = mgmt_status(hci_status);
9970
9971         bacpy(&ev.addr.bdaddr, &conn->dst);
9972         ev.addr.type = link_to_bdaddr(conn->type, conn->dst_type);
9973         ev.status = status;
9974
9975         cmd = find_pairing(conn);
9976
9977         mgmt_event(MGMT_EV_AUTH_FAILED, conn->hdev, &ev, sizeof(ev),
9978                     cmd ? cmd->sk : NULL);
9979
9980         if (cmd) {
9981                 cmd->cmd_complete(cmd, status);
9982                 mgmt_pending_remove(cmd);
9983         }
9984 }
9985
9986 void mgmt_auth_enable_complete(struct hci_dev *hdev, u8 status)
9987 {
9988         struct cmd_lookup match = { NULL, hdev };
9989         bool changed;
9990
9991         if (status) {
9992                 u8 mgmt_err = mgmt_status(status);
9993                 mgmt_pending_foreach(MGMT_OP_SET_LINK_SECURITY, hdev,
9994                                      cmd_status_rsp, &mgmt_err);
9995                 return;
9996         }
9997
9998         if (test_bit(HCI_AUTH, &hdev->flags))
9999                 changed = !hci_dev_test_and_set_flag(hdev, HCI_LINK_SECURITY);
10000         else
10001                 changed = hci_dev_test_and_clear_flag(hdev, HCI_LINK_SECURITY);
10002
10003         mgmt_pending_foreach(MGMT_OP_SET_LINK_SECURITY, hdev, settings_rsp,
10004                              &match);
10005
10006         if (changed)
10007                 new_settings(hdev, match.sk);
10008
10009         if (match.sk)
10010                 sock_put(match.sk);
10011 }
10012
10013 static void sk_lookup(struct mgmt_pending_cmd *cmd, void *data)
10014 {
10015         struct cmd_lookup *match = data;
10016
10017         if (match->sk == NULL) {
10018                 match->sk = cmd->sk;
10019                 sock_hold(match->sk);
10020         }
10021 }
10022
10023 void mgmt_set_class_of_dev_complete(struct hci_dev *hdev, u8 *dev_class,
10024                                     u8 status)
10025 {
10026         struct cmd_lookup match = { NULL, hdev, mgmt_status(status) };
10027
10028         mgmt_pending_foreach(MGMT_OP_SET_DEV_CLASS, hdev, sk_lookup, &match);
10029         mgmt_pending_foreach(MGMT_OP_ADD_UUID, hdev, sk_lookup, &match);
10030         mgmt_pending_foreach(MGMT_OP_REMOVE_UUID, hdev, sk_lookup, &match);
10031
10032         if (!status) {
10033                 mgmt_limited_event(MGMT_EV_CLASS_OF_DEV_CHANGED, hdev, dev_class,
10034                                    3, HCI_MGMT_DEV_CLASS_EVENTS, NULL);
10035                 ext_info_changed(hdev, NULL);
10036         }
10037
10038         if (match.sk)
10039                 sock_put(match.sk);
10040 }
10041
10042 void mgmt_set_local_name_complete(struct hci_dev *hdev, u8 *name, u8 status)
10043 {
10044         struct mgmt_cp_set_local_name ev;
10045         struct mgmt_pending_cmd *cmd;
10046
10047         if (status)
10048                 return;
10049
10050         memset(&ev, 0, sizeof(ev));
10051         memcpy(ev.name, name, HCI_MAX_NAME_LENGTH);
10052         memcpy(ev.short_name, hdev->short_name, HCI_MAX_SHORT_NAME_LENGTH);
10053
10054         cmd = pending_find(MGMT_OP_SET_LOCAL_NAME, hdev);
10055         if (!cmd) {
10056                 memcpy(hdev->dev_name, name, sizeof(hdev->dev_name));
10057
10058                 /* If this is a HCI command related to powering on the
10059                  * HCI dev don't send any mgmt signals.
10060                  */
10061                 if (pending_find(MGMT_OP_SET_POWERED, hdev))
10062                         return;
10063         }
10064
10065         mgmt_limited_event(MGMT_EV_LOCAL_NAME_CHANGED, hdev, &ev, sizeof(ev),
10066                            HCI_MGMT_LOCAL_NAME_EVENTS, cmd ? cmd->sk : NULL);
10067         ext_info_changed(hdev, cmd ? cmd->sk : NULL);
10068 }
10069
10070 static inline bool has_uuid(u8 *uuid, u16 uuid_count, u8 (*uuids)[16])
10071 {
10072         int i;
10073
10074         for (i = 0; i < uuid_count; i++) {
10075                 if (!memcmp(uuid, uuids[i], 16))
10076                         return true;
10077         }
10078
10079         return false;
10080 }
10081
10082 static bool eir_has_uuids(u8 *eir, u16 eir_len, u16 uuid_count, u8 (*uuids)[16])
10083 {
10084         u16 parsed = 0;
10085
10086         while (parsed < eir_len) {
10087                 u8 field_len = eir[0];
10088                 u8 uuid[16];
10089                 int i;
10090
10091                 if (field_len == 0)
10092                         break;
10093
10094                 if (eir_len - parsed < field_len + 1)
10095                         break;
10096
10097                 switch (eir[1]) {
10098                 case EIR_UUID16_ALL:
10099                 case EIR_UUID16_SOME:
10100                         for (i = 0; i + 3 <= field_len; i += 2) {
10101                                 memcpy(uuid, bluetooth_base_uuid, 16);
10102                                 uuid[13] = eir[i + 3];
10103                                 uuid[12] = eir[i + 2];
10104                                 if (has_uuid(uuid, uuid_count, uuids))
10105                                         return true;
10106                         }
10107                         break;
10108                 case EIR_UUID32_ALL:
10109                 case EIR_UUID32_SOME:
10110                         for (i = 0; i + 5 <= field_len; i += 4) {
10111                                 memcpy(uuid, bluetooth_base_uuid, 16);
10112                                 uuid[15] = eir[i + 5];
10113                                 uuid[14] = eir[i + 4];
10114                                 uuid[13] = eir[i + 3];
10115                                 uuid[12] = eir[i + 2];
10116                                 if (has_uuid(uuid, uuid_count, uuids))
10117                                         return true;
10118                         }
10119                         break;
10120                 case EIR_UUID128_ALL:
10121                 case EIR_UUID128_SOME:
10122                         for (i = 0; i + 17 <= field_len; i += 16) {
10123                                 memcpy(uuid, eir + i + 2, 16);
10124                                 if (has_uuid(uuid, uuid_count, uuids))
10125                                         return true;
10126                         }
10127                         break;
10128                 }
10129
10130                 parsed += field_len + 1;
10131                 eir += field_len + 1;
10132         }
10133
10134         return false;
10135 }
10136
10137 static void restart_le_scan(struct hci_dev *hdev)
10138 {
10139         /* If controller is not scanning we are done. */
10140         if (!hci_dev_test_flag(hdev, HCI_LE_SCAN))
10141                 return;
10142
10143         if (time_after(jiffies + DISCOV_LE_RESTART_DELAY,
10144                        hdev->discovery.scan_start +
10145                        hdev->discovery.scan_duration))
10146                 return;
10147
10148         queue_delayed_work(hdev->req_workqueue, &hdev->le_scan_restart,
10149                            DISCOV_LE_RESTART_DELAY);
10150 }
10151
10152 static bool is_filter_match(struct hci_dev *hdev, s8 rssi, u8 *eir,
10153                             u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len)
10154 {
10155         /* If a RSSI threshold has been specified, and
10156          * HCI_QUIRK_STRICT_DUPLICATE_FILTER is not set, then all results with
10157          * a RSSI smaller than the RSSI threshold will be dropped. If the quirk
10158          * is set, let it through for further processing, as we might need to
10159          * restart the scan.
10160          *
10161          * For BR/EDR devices (pre 1.2) providing no RSSI during inquiry,
10162          * the results are also dropped.
10163          */
10164         if (hdev->discovery.rssi != HCI_RSSI_INVALID &&
10165             (rssi == HCI_RSSI_INVALID ||
10166             (rssi < hdev->discovery.rssi &&
10167              !test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks))))
10168                 return  false;
10169
10170         if (hdev->discovery.uuid_count != 0) {
10171                 /* If a list of UUIDs is provided in filter, results with no
10172                  * matching UUID should be dropped.
10173                  */
10174                 if (!eir_has_uuids(eir, eir_len, hdev->discovery.uuid_count,
10175                                    hdev->discovery.uuids) &&
10176                     !eir_has_uuids(scan_rsp, scan_rsp_len,
10177                                    hdev->discovery.uuid_count,
10178                                    hdev->discovery.uuids))
10179                         return false;
10180         }
10181
10182         /* If duplicate filtering does not report RSSI changes, then restart
10183          * scanning to ensure updated result with updated RSSI values.
10184          */
10185         if (test_bit(HCI_QUIRK_STRICT_DUPLICATE_FILTER, &hdev->quirks)) {
10186                 restart_le_scan(hdev);
10187
10188                 /* Validate RSSI value against the RSSI threshold once more. */
10189                 if (hdev->discovery.rssi != HCI_RSSI_INVALID &&
10190                     rssi < hdev->discovery.rssi)
10191                         return false;
10192         }
10193
10194         return true;
10195 }
10196
10197 void mgmt_adv_monitor_device_lost(struct hci_dev *hdev, u16 handle,
10198                                   bdaddr_t *bdaddr, u8 addr_type)
10199 {
10200         struct mgmt_ev_adv_monitor_device_lost ev;
10201
10202         ev.monitor_handle = cpu_to_le16(handle);
10203         bacpy(&ev.addr.bdaddr, bdaddr);
10204         ev.addr.type = addr_type;
10205
10206         mgmt_event(MGMT_EV_ADV_MONITOR_DEVICE_LOST, hdev, &ev, sizeof(ev),
10207                    NULL);
10208 }
10209
10210 static void mgmt_send_adv_monitor_device_found(struct hci_dev *hdev,
10211                                                struct sk_buff *skb,
10212                                                struct sock *skip_sk,
10213                                                u16 handle)
10214 {
10215         struct sk_buff *advmon_skb;
10216         size_t advmon_skb_len;
10217         __le16 *monitor_handle;
10218
10219         if (!skb)
10220                 return;
10221
10222         advmon_skb_len = (sizeof(struct mgmt_ev_adv_monitor_device_found) -
10223                           sizeof(struct mgmt_ev_device_found)) + skb->len;
10224         advmon_skb = mgmt_alloc_skb(hdev, MGMT_EV_ADV_MONITOR_DEVICE_FOUND,
10225                                     advmon_skb_len);
10226         if (!advmon_skb)
10227                 return;
10228
10229         /* ADV_MONITOR_DEVICE_FOUND is similar to DEVICE_FOUND event except
10230          * that it also has 'monitor_handle'. Make a copy of DEVICE_FOUND and
10231          * store monitor_handle of the matched monitor.
10232          */
10233         monitor_handle = skb_put(advmon_skb, sizeof(*monitor_handle));
10234         *monitor_handle = cpu_to_le16(handle);
10235         skb_put_data(advmon_skb, skb->data, skb->len);
10236
10237         mgmt_event_skb(advmon_skb, skip_sk);
10238 }
10239
10240 static void mgmt_adv_monitor_device_found(struct hci_dev *hdev,
10241                                           bdaddr_t *bdaddr, bool report_device,
10242                                           struct sk_buff *skb,
10243                                           struct sock *skip_sk)
10244 {
10245         struct monitored_device *dev, *tmp;
10246         bool matched = false;
10247         bool notified = false;
10248
10249         /* We have received the Advertisement Report because:
10250          * 1. the kernel has initiated active discovery
10251          * 2. if not, we have pend_le_reports > 0 in which case we are doing
10252          *    passive scanning
10253          * 3. if none of the above is true, we have one or more active
10254          *    Advertisement Monitor
10255          *
10256          * For case 1 and 2, report all advertisements via MGMT_EV_DEVICE_FOUND
10257          * and report ONLY one advertisement per device for the matched Monitor
10258          * via MGMT_EV_ADV_MONITOR_DEVICE_FOUND event.
10259          *
10260          * For case 3, since we are not active scanning and all advertisements
10261          * received are due to a matched Advertisement Monitor, report all
10262          * advertisements ONLY via MGMT_EV_ADV_MONITOR_DEVICE_FOUND event.
10263          */
10264         if (report_device && !hdev->advmon_pend_notify) {
10265                 mgmt_event_skb(skb, skip_sk);
10266                 return;
10267         }
10268
10269         hdev->advmon_pend_notify = false;
10270
10271         list_for_each_entry_safe(dev, tmp, &hdev->monitored_devices, list) {
10272                 if (!bacmp(&dev->bdaddr, bdaddr)) {
10273                         matched = true;
10274
10275                         if (!dev->notified) {
10276                                 mgmt_send_adv_monitor_device_found(hdev, skb,
10277                                                                    skip_sk,
10278                                                                    dev->handle);
10279                                 notified = true;
10280                                 dev->notified = true;
10281                         }
10282                 }
10283
10284                 if (!dev->notified)
10285                         hdev->advmon_pend_notify = true;
10286         }
10287
10288         if (!report_device &&
10289             ((matched && !notified) || !msft_monitor_supported(hdev))) {
10290                 /* Handle 0 indicates that we are not active scanning and this
10291                  * is a subsequent advertisement report for an already matched
10292                  * Advertisement Monitor or the controller offloading support
10293                  * is not available.
10294                  */
10295                 mgmt_send_adv_monitor_device_found(hdev, skb, skip_sk, 0);
10296         }
10297
10298         if (report_device)
10299                 mgmt_event_skb(skb, skip_sk);
10300         else
10301                 kfree_skb(skb);
10302 }
10303
10304 static void mesh_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr,
10305                               u8 addr_type, s8 rssi, u32 flags, u8 *eir,
10306                               u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len,
10307                               u64 instant)
10308 {
10309         struct sk_buff *skb;
10310         struct mgmt_ev_mesh_device_found *ev;
10311         int i, j;
10312
10313         if (!hdev->mesh_ad_types[0])
10314                 goto accepted;
10315
10316         /* Scan for requested AD types */
10317         if (eir_len > 0) {
10318                 for (i = 0; i + 1 < eir_len; i += eir[i] + 1) {
10319                         for (j = 0; j < sizeof(hdev->mesh_ad_types); j++) {
10320                                 if (!hdev->mesh_ad_types[j])
10321                                         break;
10322
10323                                 if (hdev->mesh_ad_types[j] == eir[i + 1])
10324                                         goto accepted;
10325                         }
10326                 }
10327         }
10328
10329         if (scan_rsp_len > 0) {
10330                 for (i = 0; i + 1 < scan_rsp_len; i += scan_rsp[i] + 1) {
10331                         for (j = 0; j < sizeof(hdev->mesh_ad_types); j++) {
10332                                 if (!hdev->mesh_ad_types[j])
10333                                         break;
10334
10335                                 if (hdev->mesh_ad_types[j] == scan_rsp[i + 1])
10336                                         goto accepted;
10337                         }
10338                 }
10339         }
10340
10341         return;
10342
10343 accepted:
10344         skb = mgmt_alloc_skb(hdev, MGMT_EV_MESH_DEVICE_FOUND,
10345                              sizeof(*ev) + eir_len + scan_rsp_len);
10346         if (!skb)
10347                 return;
10348
10349         ev = skb_put(skb, sizeof(*ev));
10350
10351         bacpy(&ev->addr.bdaddr, bdaddr);
10352         ev->addr.type = link_to_bdaddr(LE_LINK, addr_type);
10353         ev->rssi = rssi;
10354         ev->flags = cpu_to_le32(flags);
10355         ev->instant = cpu_to_le64(instant);
10356
10357         if (eir_len > 0)
10358                 /* Copy EIR or advertising data into event */
10359                 skb_put_data(skb, eir, eir_len);
10360
10361         if (scan_rsp_len > 0)
10362                 /* Append scan response data to event */
10363                 skb_put_data(skb, scan_rsp, scan_rsp_len);
10364
10365         ev->eir_len = cpu_to_le16(eir_len + scan_rsp_len);
10366
10367         mgmt_event_skb(skb, NULL);
10368 }
10369
10370 void mgmt_device_found(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type,
10371                        u8 addr_type, u8 *dev_class, s8 rssi, u32 flags,
10372                        u8 *eir, u16 eir_len, u8 *scan_rsp, u8 scan_rsp_len,
10373                        u64 instant)
10374 {
10375         struct sk_buff *skb;
10376         struct mgmt_ev_device_found *ev;
10377         bool report_device = hci_discovery_active(hdev);
10378
10379         if (hci_dev_test_flag(hdev, HCI_MESH) && link_type == LE_LINK)
10380                 mesh_device_found(hdev, bdaddr, addr_type, rssi, flags,
10381                                   eir, eir_len, scan_rsp, scan_rsp_len,
10382                                   instant);
10383
10384         /* Don't send events for a non-kernel initiated discovery. With
10385          * LE one exception is if we have pend_le_reports > 0 in which
10386          * case we're doing passive scanning and want these events.
10387          */
10388         if (!hci_discovery_active(hdev)) {
10389                 if (link_type == ACL_LINK)
10390                         return;
10391                 if (link_type == LE_LINK && !list_empty(&hdev->pend_le_reports))
10392                         report_device = true;
10393                 else if (!hci_is_adv_monitoring(hdev))
10394                         return;
10395         }
10396
10397         if (hdev->discovery.result_filtering) {
10398                 /* We are using service discovery */
10399                 if (!is_filter_match(hdev, rssi, eir, eir_len, scan_rsp,
10400                                      scan_rsp_len))
10401                         return;
10402         }
10403
10404         if (hdev->discovery.limited) {
10405                 /* Check for limited discoverable bit */
10406                 if (dev_class) {
10407                         if (!(dev_class[1] & 0x20))
10408                                 return;
10409                 } else {
10410                         u8 *flags = eir_get_data(eir, eir_len, EIR_FLAGS, NULL);
10411                         if (!flags || !(flags[0] & LE_AD_LIMITED))
10412                                 return;
10413                 }
10414         }
10415
10416         /* Allocate skb. The 5 extra bytes are for the potential CoD field */
10417         skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_FOUND,
10418                              sizeof(*ev) + eir_len + scan_rsp_len + 5);
10419         if (!skb)
10420                 return;
10421
10422         ev = skb_put(skb, sizeof(*ev));
10423
10424         /* In case of device discovery with BR/EDR devices (pre 1.2), the
10425          * RSSI value was reported as 0 when not available. This behavior
10426          * is kept when using device discovery. This is required for full
10427          * backwards compatibility with the API.
10428          *
10429          * However when using service discovery, the value 127 will be
10430          * returned when the RSSI is not available.
10431          */
10432         if (rssi == HCI_RSSI_INVALID && !hdev->discovery.report_invalid_rssi &&
10433             link_type == ACL_LINK)
10434                 rssi = 0;
10435
10436         bacpy(&ev->addr.bdaddr, bdaddr);
10437         ev->addr.type = link_to_bdaddr(link_type, addr_type);
10438         ev->rssi = rssi;
10439         ev->flags = cpu_to_le32(flags);
10440
10441         if (eir_len > 0)
10442                 /* Copy EIR or advertising data into event */
10443                 skb_put_data(skb, eir, eir_len);
10444
10445         if (dev_class && !eir_get_data(eir, eir_len, EIR_CLASS_OF_DEV, NULL)) {
10446                 u8 eir_cod[5];
10447
10448                 eir_len += eir_append_data(eir_cod, 0, EIR_CLASS_OF_DEV,
10449                                            dev_class, 3);
10450                 skb_put_data(skb, eir_cod, sizeof(eir_cod));
10451         }
10452
10453         if (scan_rsp_len > 0)
10454                 /* Append scan response data to event */
10455                 skb_put_data(skb, scan_rsp, scan_rsp_len);
10456
10457         ev->eir_len = cpu_to_le16(eir_len + scan_rsp_len);
10458
10459         mgmt_adv_monitor_device_found(hdev, bdaddr, report_device, skb, NULL);
10460 }
10461
10462 void mgmt_remote_name(struct hci_dev *hdev, bdaddr_t *bdaddr, u8 link_type,
10463                       u8 addr_type, s8 rssi, u8 *name, u8 name_len)
10464 {
10465         struct sk_buff *skb;
10466         struct mgmt_ev_device_found *ev;
10467         u16 eir_len = 0;
10468         u32 flags = 0;
10469
10470         skb = mgmt_alloc_skb(hdev, MGMT_EV_DEVICE_FOUND,
10471                              sizeof(*ev) + (name ? eir_precalc_len(name_len) : 0));
10472
10473         ev = skb_put(skb, sizeof(*ev));
10474         bacpy(&ev->addr.bdaddr, bdaddr);
10475         ev->addr.type = link_to_bdaddr(link_type, addr_type);
10476         ev->rssi = rssi;
10477
10478         if (name)
10479                 eir_len += eir_skb_put_data(skb, EIR_NAME_COMPLETE, name, name_len);
10480         else
10481                 flags = MGMT_DEV_FOUND_NAME_REQUEST_FAILED;
10482
10483         ev->eir_len = cpu_to_le16(eir_len);
10484         ev->flags = cpu_to_le32(flags);
10485
10486         mgmt_event_skb(skb, NULL);
10487 }
10488
10489 void mgmt_discovering(struct hci_dev *hdev, u8 discovering)
10490 {
10491         struct mgmt_ev_discovering ev;
10492
10493         bt_dev_dbg(hdev, "discovering %u", discovering);
10494
10495         memset(&ev, 0, sizeof(ev));
10496         ev.type = hdev->discovery.type;
10497         ev.discovering = discovering;
10498
10499         mgmt_event(MGMT_EV_DISCOVERING, hdev, &ev, sizeof(ev), NULL);
10500 }
10501
10502 void mgmt_suspending(struct hci_dev *hdev, u8 state)
10503 {
10504         struct mgmt_ev_controller_suspend ev;
10505
10506         ev.suspend_state = state;
10507         mgmt_event(MGMT_EV_CONTROLLER_SUSPEND, hdev, &ev, sizeof(ev), NULL);
10508 }
10509
10510 void mgmt_resuming(struct hci_dev *hdev, u8 reason, bdaddr_t *bdaddr,
10511                    u8 addr_type)
10512 {
10513         struct mgmt_ev_controller_resume ev;
10514
10515         ev.wake_reason = reason;
10516         if (bdaddr) {
10517                 bacpy(&ev.addr.bdaddr, bdaddr);
10518                 ev.addr.type = addr_type;
10519         } else {
10520                 memset(&ev.addr, 0, sizeof(ev.addr));
10521         }
10522
10523         mgmt_event(MGMT_EV_CONTROLLER_RESUME, hdev, &ev, sizeof(ev), NULL);
10524 }
10525
10526 static struct hci_mgmt_chan chan = {
10527         .channel        = HCI_CHANNEL_CONTROL,
10528         .handler_count  = ARRAY_SIZE(mgmt_handlers),
10529         .handlers       = mgmt_handlers,
10530         .hdev_init      = mgmt_init_hdev,
10531 };
10532
10533 int mgmt_init(void)
10534 {
10535         return hci_mgmt_chan_register(&chan);
10536 }
10537
10538 void mgmt_exit(void)
10539 {
10540         hci_mgmt_chan_unregister(&chan);
10541 }
10542
10543 void mgmt_cleanup(struct sock *sk)
10544 {
10545         struct mgmt_mesh_tx *mesh_tx;
10546         struct hci_dev *hdev;
10547
10548         read_lock(&hci_dev_list_lock);
10549
10550         list_for_each_entry(hdev, &hci_dev_list, list) {
10551                 do {
10552                         mesh_tx = mgmt_mesh_next(hdev, sk);
10553
10554                         if (mesh_tx)
10555                                 mesh_send_complete(hdev, mesh_tx, true);
10556                 } while (mesh_tx);
10557         }
10558
10559         read_unlock(&hci_dev_list_lock);
10560 }