Merge tag 'perf-tools-fixes-for-v6.6-2-2023-10-20' of git://git.kernel.org/pub/scm...
[platform/kernel/linux-starfive.git] / net / bluetooth / coredump.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) 2023 Google Corporation
4  */
5
6 #include <linux/devcoredump.h>
7
8 #include <asm/unaligned.h>
9 #include <net/bluetooth/bluetooth.h>
10 #include <net/bluetooth/hci_core.h>
11
12 enum hci_devcoredump_pkt_type {
13         HCI_DEVCOREDUMP_PKT_INIT,
14         HCI_DEVCOREDUMP_PKT_SKB,
15         HCI_DEVCOREDUMP_PKT_PATTERN,
16         HCI_DEVCOREDUMP_PKT_COMPLETE,
17         HCI_DEVCOREDUMP_PKT_ABORT,
18 };
19
20 struct hci_devcoredump_skb_cb {
21         u16 pkt_type;
22 };
23
24 struct hci_devcoredump_skb_pattern {
25         u8 pattern;
26         u32 len;
27 } __packed;
28
29 #define hci_dmp_cb(skb) ((struct hci_devcoredump_skb_cb *)((skb)->cb))
30
31 #define DBG_UNEXPECTED_STATE() \
32         bt_dev_dbg(hdev, \
33                    "Unexpected packet (%d) for state (%d). ", \
34                    hci_dmp_cb(skb)->pkt_type, hdev->dump.state)
35
36 #define MAX_DEVCOREDUMP_HDR_SIZE        512     /* bytes */
37
38 static int hci_devcd_update_hdr_state(char *buf, size_t size, int state)
39 {
40         int len = 0;
41
42         if (!buf)
43                 return 0;
44
45         len = scnprintf(buf, size, "Bluetooth devcoredump\nState: %d\n", state);
46
47         return len + 1; /* scnprintf adds \0 at the end upon state rewrite */
48 }
49
50 /* Call with hci_dev_lock only. */
51 static int hci_devcd_update_state(struct hci_dev *hdev, int state)
52 {
53         bt_dev_dbg(hdev, "Updating devcoredump state from %d to %d.",
54                    hdev->dump.state, state);
55
56         hdev->dump.state = state;
57
58         return hci_devcd_update_hdr_state(hdev->dump.head,
59                                           hdev->dump.alloc_size, state);
60 }
61
62 static int hci_devcd_mkheader(struct hci_dev *hdev, struct sk_buff *skb)
63 {
64         char dump_start[] = "--- Start dump ---\n";
65         char hdr[80];
66         int hdr_len;
67
68         hdr_len = hci_devcd_update_hdr_state(hdr, sizeof(hdr),
69                                              HCI_DEVCOREDUMP_IDLE);
70         skb_put_data(skb, hdr, hdr_len);
71
72         if (hdev->dump.dmp_hdr)
73                 hdev->dump.dmp_hdr(hdev, skb);
74
75         skb_put_data(skb, dump_start, strlen(dump_start));
76
77         return skb->len;
78 }
79
80 /* Do not call with hci_dev_lock since this calls driver code. */
81 static void hci_devcd_notify(struct hci_dev *hdev, int state)
82 {
83         if (hdev->dump.notify_change)
84                 hdev->dump.notify_change(hdev, state);
85 }
86
87 /* Call with hci_dev_lock only. */
88 void hci_devcd_reset(struct hci_dev *hdev)
89 {
90         hdev->dump.head = NULL;
91         hdev->dump.tail = NULL;
92         hdev->dump.alloc_size = 0;
93
94         hci_devcd_update_state(hdev, HCI_DEVCOREDUMP_IDLE);
95
96         cancel_delayed_work(&hdev->dump.dump_timeout);
97         skb_queue_purge(&hdev->dump.dump_q);
98 }
99
100 /* Call with hci_dev_lock only. */
101 static void hci_devcd_free(struct hci_dev *hdev)
102 {
103         vfree(hdev->dump.head);
104
105         hci_devcd_reset(hdev);
106 }
107
108 /* Call with hci_dev_lock only. */
109 static int hci_devcd_alloc(struct hci_dev *hdev, u32 size)
110 {
111         hdev->dump.head = vmalloc(size);
112         if (!hdev->dump.head)
113                 return -ENOMEM;
114
115         hdev->dump.alloc_size = size;
116         hdev->dump.tail = hdev->dump.head;
117         hdev->dump.end = hdev->dump.head + size;
118
119         hci_devcd_update_state(hdev, HCI_DEVCOREDUMP_IDLE);
120
121         return 0;
122 }
123
124 /* Call with hci_dev_lock only. */
125 static bool hci_devcd_copy(struct hci_dev *hdev, char *buf, u32 size)
126 {
127         if (hdev->dump.tail + size > hdev->dump.end)
128                 return false;
129
130         memcpy(hdev->dump.tail, buf, size);
131         hdev->dump.tail += size;
132
133         return true;
134 }
135
136 /* Call with hci_dev_lock only. */
137 static bool hci_devcd_memset(struct hci_dev *hdev, u8 pattern, u32 len)
138 {
139         if (hdev->dump.tail + len > hdev->dump.end)
140                 return false;
141
142         memset(hdev->dump.tail, pattern, len);
143         hdev->dump.tail += len;
144
145         return true;
146 }
147
148 /* Call with hci_dev_lock only. */
149 static int hci_devcd_prepare(struct hci_dev *hdev, u32 dump_size)
150 {
151         struct sk_buff *skb;
152         int dump_hdr_size;
153         int err = 0;
154
155         skb = alloc_skb(MAX_DEVCOREDUMP_HDR_SIZE, GFP_ATOMIC);
156         if (!skb)
157                 return -ENOMEM;
158
159         dump_hdr_size = hci_devcd_mkheader(hdev, skb);
160
161         if (hci_devcd_alloc(hdev, dump_hdr_size + dump_size)) {
162                 err = -ENOMEM;
163                 goto hdr_free;
164         }
165
166         /* Insert the device header */
167         if (!hci_devcd_copy(hdev, skb->data, skb->len)) {
168                 bt_dev_err(hdev, "Failed to insert header");
169                 hci_devcd_free(hdev);
170
171                 err = -ENOMEM;
172                 goto hdr_free;
173         }
174
175 hdr_free:
176         kfree_skb(skb);
177
178         return err;
179 }
180
181 static void hci_devcd_handle_pkt_init(struct hci_dev *hdev, struct sk_buff *skb)
182 {
183         u32 dump_size;
184
185         if (hdev->dump.state != HCI_DEVCOREDUMP_IDLE) {
186                 DBG_UNEXPECTED_STATE();
187                 return;
188         }
189
190         if (skb->len != sizeof(dump_size)) {
191                 bt_dev_dbg(hdev, "Invalid dump init pkt");
192                 return;
193         }
194
195         dump_size = get_unaligned_le32(skb_pull_data(skb, 4));
196         if (!dump_size) {
197                 bt_dev_err(hdev, "Zero size dump init pkt");
198                 return;
199         }
200
201         if (hci_devcd_prepare(hdev, dump_size)) {
202                 bt_dev_err(hdev, "Failed to prepare for dump");
203                 return;
204         }
205
206         hci_devcd_update_state(hdev, HCI_DEVCOREDUMP_ACTIVE);
207         queue_delayed_work(hdev->workqueue, &hdev->dump.dump_timeout,
208                            hdev->dump.timeout);
209 }
210
211 static void hci_devcd_handle_pkt_skb(struct hci_dev *hdev, struct sk_buff *skb)
212 {
213         if (hdev->dump.state != HCI_DEVCOREDUMP_ACTIVE) {
214                 DBG_UNEXPECTED_STATE();
215                 return;
216         }
217
218         if (!hci_devcd_copy(hdev, skb->data, skb->len))
219                 bt_dev_dbg(hdev, "Failed to insert skb");
220 }
221
222 static void hci_devcd_handle_pkt_pattern(struct hci_dev *hdev,
223                                          struct sk_buff *skb)
224 {
225         struct hci_devcoredump_skb_pattern *pattern;
226
227         if (hdev->dump.state != HCI_DEVCOREDUMP_ACTIVE) {
228                 DBG_UNEXPECTED_STATE();
229                 return;
230         }
231
232         if (skb->len != sizeof(*pattern)) {
233                 bt_dev_dbg(hdev, "Invalid pattern skb");
234                 return;
235         }
236
237         pattern = skb_pull_data(skb, sizeof(*pattern));
238
239         if (!hci_devcd_memset(hdev, pattern->pattern, pattern->len))
240                 bt_dev_dbg(hdev, "Failed to set pattern");
241 }
242
243 static void hci_devcd_handle_pkt_complete(struct hci_dev *hdev,
244                                           struct sk_buff *skb)
245 {
246         u32 dump_size;
247
248         if (hdev->dump.state != HCI_DEVCOREDUMP_ACTIVE) {
249                 DBG_UNEXPECTED_STATE();
250                 return;
251         }
252
253         hci_devcd_update_state(hdev, HCI_DEVCOREDUMP_DONE);
254         dump_size = hdev->dump.tail - hdev->dump.head;
255
256         bt_dev_dbg(hdev, "complete with size %u (expect %zu)", dump_size,
257                    hdev->dump.alloc_size);
258
259         dev_coredumpv(&hdev->dev, hdev->dump.head, dump_size, GFP_KERNEL);
260 }
261
262 static void hci_devcd_handle_pkt_abort(struct hci_dev *hdev,
263                                        struct sk_buff *skb)
264 {
265         u32 dump_size;
266
267         if (hdev->dump.state != HCI_DEVCOREDUMP_ACTIVE) {
268                 DBG_UNEXPECTED_STATE();
269                 return;
270         }
271
272         hci_devcd_update_state(hdev, HCI_DEVCOREDUMP_ABORT);
273         dump_size = hdev->dump.tail - hdev->dump.head;
274
275         bt_dev_dbg(hdev, "aborted with size %u (expect %zu)", dump_size,
276                    hdev->dump.alloc_size);
277
278         /* Emit a devcoredump with the available data */
279         dev_coredumpv(&hdev->dev, hdev->dump.head, dump_size, GFP_KERNEL);
280 }
281
282 /* Bluetooth devcoredump state machine.
283  *
284  * Devcoredump states:
285  *
286  *      HCI_DEVCOREDUMP_IDLE: The default state.
287  *
288  *      HCI_DEVCOREDUMP_ACTIVE: A devcoredump will be in this state once it has
289  *              been initialized using hci_devcd_init(). Once active, the driver
290  *              can append data using hci_devcd_append() or insert a pattern
291  *              using hci_devcd_append_pattern().
292  *
293  *      HCI_DEVCOREDUMP_DONE: Once the dump collection is complete, the drive
294  *              can signal the completion using hci_devcd_complete(). A
295  *              devcoredump is generated indicating the completion event and
296  *              then the state machine is reset to the default state.
297  *
298  *      HCI_DEVCOREDUMP_ABORT: The driver can cancel ongoing dump collection in
299  *              case of any error using hci_devcd_abort(). A devcoredump is
300  *              still generated with the available data indicating the abort
301  *              event and then the state machine is reset to the default state.
302  *
303  *      HCI_DEVCOREDUMP_TIMEOUT: A timeout timer for HCI_DEVCOREDUMP_TIMEOUT sec
304  *              is started during devcoredump initialization. Once the timeout
305  *              occurs, the driver is notified, a devcoredump is generated with
306  *              the available data indicating the timeout event and then the
307  *              state machine is reset to the default state.
308  *
309  * The driver must register using hci_devcd_register() before using the hci
310  * devcoredump APIs.
311  */
312 void hci_devcd_rx(struct work_struct *work)
313 {
314         struct hci_dev *hdev = container_of(work, struct hci_dev, dump.dump_rx);
315         struct sk_buff *skb;
316         int start_state;
317
318         while ((skb = skb_dequeue(&hdev->dump.dump_q))) {
319                 /* Return if timeout occurs. The timeout handler function
320                  * hci_devcd_timeout() will report the available dump data.
321                  */
322                 if (hdev->dump.state == HCI_DEVCOREDUMP_TIMEOUT) {
323                         kfree_skb(skb);
324                         return;
325                 }
326
327                 hci_dev_lock(hdev);
328                 start_state = hdev->dump.state;
329
330                 switch (hci_dmp_cb(skb)->pkt_type) {
331                 case HCI_DEVCOREDUMP_PKT_INIT:
332                         hci_devcd_handle_pkt_init(hdev, skb);
333                         break;
334
335                 case HCI_DEVCOREDUMP_PKT_SKB:
336                         hci_devcd_handle_pkt_skb(hdev, skb);
337                         break;
338
339                 case HCI_DEVCOREDUMP_PKT_PATTERN:
340                         hci_devcd_handle_pkt_pattern(hdev, skb);
341                         break;
342
343                 case HCI_DEVCOREDUMP_PKT_COMPLETE:
344                         hci_devcd_handle_pkt_complete(hdev, skb);
345                         break;
346
347                 case HCI_DEVCOREDUMP_PKT_ABORT:
348                         hci_devcd_handle_pkt_abort(hdev, skb);
349                         break;
350
351                 default:
352                         bt_dev_dbg(hdev, "Unknown packet (%d) for state (%d). ",
353                                    hci_dmp_cb(skb)->pkt_type, hdev->dump.state);
354                         break;
355                 }
356
357                 hci_dev_unlock(hdev);
358                 kfree_skb(skb);
359
360                 /* Notify the driver about any state changes before resetting
361                  * the state machine
362                  */
363                 if (start_state != hdev->dump.state)
364                         hci_devcd_notify(hdev, hdev->dump.state);
365
366                 /* Reset the state machine if the devcoredump is complete */
367                 hci_dev_lock(hdev);
368                 if (hdev->dump.state == HCI_DEVCOREDUMP_DONE ||
369                     hdev->dump.state == HCI_DEVCOREDUMP_ABORT)
370                         hci_devcd_reset(hdev);
371                 hci_dev_unlock(hdev);
372         }
373 }
374 EXPORT_SYMBOL(hci_devcd_rx);
375
376 void hci_devcd_timeout(struct work_struct *work)
377 {
378         struct hci_dev *hdev = container_of(work, struct hci_dev,
379                                             dump.dump_timeout.work);
380         u32 dump_size;
381
382         hci_devcd_notify(hdev, HCI_DEVCOREDUMP_TIMEOUT);
383
384         hci_dev_lock(hdev);
385
386         cancel_work(&hdev->dump.dump_rx);
387
388         hci_devcd_update_state(hdev, HCI_DEVCOREDUMP_TIMEOUT);
389
390         dump_size = hdev->dump.tail - hdev->dump.head;
391         bt_dev_dbg(hdev, "timeout with size %u (expect %zu)", dump_size,
392                    hdev->dump.alloc_size);
393
394         /* Emit a devcoredump with the available data */
395         dev_coredumpv(&hdev->dev, hdev->dump.head, dump_size, GFP_KERNEL);
396
397         hci_devcd_reset(hdev);
398
399         hci_dev_unlock(hdev);
400 }
401 EXPORT_SYMBOL(hci_devcd_timeout);
402
403 int hci_devcd_register(struct hci_dev *hdev, coredump_t coredump,
404                        dmp_hdr_t dmp_hdr, notify_change_t notify_change)
405 {
406         /* Driver must implement coredump() and dmp_hdr() functions for
407          * bluetooth devcoredump. The coredump() should trigger a coredump
408          * event on the controller when the device's coredump sysfs entry is
409          * written to. The dmp_hdr() should create a dump header to identify
410          * the controller/fw/driver info.
411          */
412         if (!coredump || !dmp_hdr)
413                 return -EINVAL;
414
415         hci_dev_lock(hdev);
416         hdev->dump.coredump = coredump;
417         hdev->dump.dmp_hdr = dmp_hdr;
418         hdev->dump.notify_change = notify_change;
419         hdev->dump.supported = true;
420         hdev->dump.timeout = DEVCOREDUMP_TIMEOUT;
421         hci_dev_unlock(hdev);
422
423         return 0;
424 }
425 EXPORT_SYMBOL(hci_devcd_register);
426
427 static inline bool hci_devcd_enabled(struct hci_dev *hdev)
428 {
429         return hdev->dump.supported;
430 }
431
432 int hci_devcd_init(struct hci_dev *hdev, u32 dump_size)
433 {
434         struct sk_buff *skb;
435
436         if (!hci_devcd_enabled(hdev))
437                 return -EOPNOTSUPP;
438
439         skb = alloc_skb(sizeof(dump_size), GFP_ATOMIC);
440         if (!skb)
441                 return -ENOMEM;
442
443         hci_dmp_cb(skb)->pkt_type = HCI_DEVCOREDUMP_PKT_INIT;
444         put_unaligned_le32(dump_size, skb_put(skb, 4));
445
446         skb_queue_tail(&hdev->dump.dump_q, skb);
447         queue_work(hdev->workqueue, &hdev->dump.dump_rx);
448
449         return 0;
450 }
451 EXPORT_SYMBOL(hci_devcd_init);
452
453 int hci_devcd_append(struct hci_dev *hdev, struct sk_buff *skb)
454 {
455         if (!skb)
456                 return -ENOMEM;
457
458         if (!hci_devcd_enabled(hdev)) {
459                 kfree_skb(skb);
460                 return -EOPNOTSUPP;
461         }
462
463         hci_dmp_cb(skb)->pkt_type = HCI_DEVCOREDUMP_PKT_SKB;
464
465         skb_queue_tail(&hdev->dump.dump_q, skb);
466         queue_work(hdev->workqueue, &hdev->dump.dump_rx);
467
468         return 0;
469 }
470 EXPORT_SYMBOL(hci_devcd_append);
471
472 int hci_devcd_append_pattern(struct hci_dev *hdev, u8 pattern, u32 len)
473 {
474         struct hci_devcoredump_skb_pattern p;
475         struct sk_buff *skb;
476
477         if (!hci_devcd_enabled(hdev))
478                 return -EOPNOTSUPP;
479
480         skb = alloc_skb(sizeof(p), GFP_ATOMIC);
481         if (!skb)
482                 return -ENOMEM;
483
484         p.pattern = pattern;
485         p.len = len;
486
487         hci_dmp_cb(skb)->pkt_type = HCI_DEVCOREDUMP_PKT_PATTERN;
488         skb_put_data(skb, &p, sizeof(p));
489
490         skb_queue_tail(&hdev->dump.dump_q, skb);
491         queue_work(hdev->workqueue, &hdev->dump.dump_rx);
492
493         return 0;
494 }
495 EXPORT_SYMBOL(hci_devcd_append_pattern);
496
497 int hci_devcd_complete(struct hci_dev *hdev)
498 {
499         struct sk_buff *skb;
500
501         if (!hci_devcd_enabled(hdev))
502                 return -EOPNOTSUPP;
503
504         skb = alloc_skb(0, GFP_ATOMIC);
505         if (!skb)
506                 return -ENOMEM;
507
508         hci_dmp_cb(skb)->pkt_type = HCI_DEVCOREDUMP_PKT_COMPLETE;
509
510         skb_queue_tail(&hdev->dump.dump_q, skb);
511         queue_work(hdev->workqueue, &hdev->dump.dump_rx);
512
513         return 0;
514 }
515 EXPORT_SYMBOL(hci_devcd_complete);
516
517 int hci_devcd_abort(struct hci_dev *hdev)
518 {
519         struct sk_buff *skb;
520
521         if (!hci_devcd_enabled(hdev))
522                 return -EOPNOTSUPP;
523
524         skb = alloc_skb(0, GFP_ATOMIC);
525         if (!skb)
526                 return -ENOMEM;
527
528         hci_dmp_cb(skb)->pkt_type = HCI_DEVCOREDUMP_PKT_ABORT;
529
530         skb_queue_tail(&hdev->dump.dump_q, skb);
531         queue_work(hdev->workqueue, &hdev->dump.dump_rx);
532
533         return 0;
534 }
535 EXPORT_SYMBOL(hci_devcd_abort);