Merge branch 'for-6.2/i2c' into for-linus
[platform/kernel/linux-rpi.git] / arch / arm64 / kernel / proton-pack.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Handle detection, reporting and mitigation of Spectre v1, v2, v3a and v4, as
4  * detailed at:
5  *
6  *   https://developer.arm.com/support/arm-security-updates/speculative-processor-vulnerability
7  *
8  * This code was originally written hastily under an awful lot of stress and so
9  * aspects of it are somewhat hacky. Unfortunately, changing anything in here
10  * instantly makes me feel ill. Thanks, Jann. Thann.
11  *
12  * Copyright (C) 2018 ARM Ltd, All Rights Reserved.
13  * Copyright (C) 2020 Google LLC
14  *
15  * "If there's something strange in your neighbourhood, who you gonna call?"
16  *
17  * Authors: Will Deacon <will@kernel.org> and Marc Zyngier <maz@kernel.org>
18  */
19
20 #include <linux/arm-smccc.h>
21 #include <linux/bpf.h>
22 #include <linux/cpu.h>
23 #include <linux/device.h>
24 #include <linux/nospec.h>
25 #include <linux/prctl.h>
26 #include <linux/sched/task_stack.h>
27
28 #include <asm/debug-monitors.h>
29 #include <asm/insn.h>
30 #include <asm/spectre.h>
31 #include <asm/traps.h>
32 #include <asm/vectors.h>
33 #include <asm/virt.h>
34
35 /*
36  * We try to ensure that the mitigation state can never change as the result of
37  * onlining a late CPU.
38  */
39 static void update_mitigation_state(enum mitigation_state *oldp,
40                                     enum mitigation_state new)
41 {
42         enum mitigation_state state;
43
44         do {
45                 state = READ_ONCE(*oldp);
46                 if (new <= state)
47                         break;
48
49                 /* Userspace almost certainly can't deal with this. */
50                 if (WARN_ON(system_capabilities_finalized()))
51                         break;
52         } while (cmpxchg_relaxed(oldp, state, new) != state);
53 }
54
55 /*
56  * Spectre v1.
57  *
58  * The kernel can't protect userspace for this one: it's each person for
59  * themselves. Advertise what we're doing and be done with it.
60  */
61 ssize_t cpu_show_spectre_v1(struct device *dev, struct device_attribute *attr,
62                             char *buf)
63 {
64         return sprintf(buf, "Mitigation: __user pointer sanitization\n");
65 }
66
67 /*
68  * Spectre v2.
69  *
70  * This one sucks. A CPU is either:
71  *
72  * - Mitigated in hardware and advertised by ID_AA64PFR0_EL1.CSV2.
73  * - Mitigated in hardware and listed in our "safe list".
74  * - Mitigated in software by firmware.
75  * - Mitigated in software by a CPU-specific dance in the kernel and a
76  *   firmware call at EL2.
77  * - Vulnerable.
78  *
79  * It's not unlikely for different CPUs in a big.LITTLE system to fall into
80  * different camps.
81  */
82 static enum mitigation_state spectre_v2_state;
83
84 static bool __read_mostly __nospectre_v2;
85 static int __init parse_spectre_v2_param(char *str)
86 {
87         __nospectre_v2 = true;
88         return 0;
89 }
90 early_param("nospectre_v2", parse_spectre_v2_param);
91
92 static bool spectre_v2_mitigations_off(void)
93 {
94         bool ret = __nospectre_v2 || cpu_mitigations_off();
95
96         if (ret)
97                 pr_info_once("spectre-v2 mitigation disabled by command line option\n");
98
99         return ret;
100 }
101
102 static const char *get_bhb_affected_string(enum mitigation_state bhb_state)
103 {
104         switch (bhb_state) {
105         case SPECTRE_UNAFFECTED:
106                 return "";
107         default:
108         case SPECTRE_VULNERABLE:
109                 return ", but not BHB";
110         case SPECTRE_MITIGATED:
111                 return ", BHB";
112         }
113 }
114
115 static bool _unprivileged_ebpf_enabled(void)
116 {
117 #ifdef CONFIG_BPF_SYSCALL
118         return !sysctl_unprivileged_bpf_disabled;
119 #else
120         return false;
121 #endif
122 }
123
124 ssize_t cpu_show_spectre_v2(struct device *dev, struct device_attribute *attr,
125                             char *buf)
126 {
127         enum mitigation_state bhb_state = arm64_get_spectre_bhb_state();
128         const char *bhb_str = get_bhb_affected_string(bhb_state);
129         const char *v2_str = "Branch predictor hardening";
130
131         switch (spectre_v2_state) {
132         case SPECTRE_UNAFFECTED:
133                 if (bhb_state == SPECTRE_UNAFFECTED)
134                         return sprintf(buf, "Not affected\n");
135
136                 /*
137                  * Platforms affected by Spectre-BHB can't report
138                  * "Not affected" for Spectre-v2.
139                  */
140                 v2_str = "CSV2";
141                 fallthrough;
142         case SPECTRE_MITIGATED:
143                 if (bhb_state == SPECTRE_MITIGATED && _unprivileged_ebpf_enabled())
144                         return sprintf(buf, "Vulnerable: Unprivileged eBPF enabled\n");
145
146                 return sprintf(buf, "Mitigation: %s%s\n", v2_str, bhb_str);
147         case SPECTRE_VULNERABLE:
148                 fallthrough;
149         default:
150                 return sprintf(buf, "Vulnerable\n");
151         }
152 }
153
154 static enum mitigation_state spectre_v2_get_cpu_hw_mitigation_state(void)
155 {
156         u64 pfr0;
157         static const struct midr_range spectre_v2_safe_list[] = {
158                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A35),
159                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A53),
160                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A55),
161                 MIDR_ALL_VERSIONS(MIDR_BRAHMA_B53),
162                 MIDR_ALL_VERSIONS(MIDR_HISI_TSV110),
163                 MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_2XX_SILVER),
164                 MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_3XX_SILVER),
165                 MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_4XX_SILVER),
166                 { /* sentinel */ }
167         };
168
169         /* If the CPU has CSV2 set, we're safe */
170         pfr0 = read_cpuid(ID_AA64PFR0_EL1);
171         if (cpuid_feature_extract_unsigned_field(pfr0, ID_AA64PFR0_EL1_CSV2_SHIFT))
172                 return SPECTRE_UNAFFECTED;
173
174         /* Alternatively, we have a list of unaffected CPUs */
175         if (is_midr_in_range_list(read_cpuid_id(), spectre_v2_safe_list))
176                 return SPECTRE_UNAFFECTED;
177
178         return SPECTRE_VULNERABLE;
179 }
180
181 static enum mitigation_state spectre_v2_get_cpu_fw_mitigation_state(void)
182 {
183         int ret;
184         struct arm_smccc_res res;
185
186         arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
187                              ARM_SMCCC_ARCH_WORKAROUND_1, &res);
188
189         ret = res.a0;
190         switch (ret) {
191         case SMCCC_RET_SUCCESS:
192                 return SPECTRE_MITIGATED;
193         case SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED:
194                 return SPECTRE_UNAFFECTED;
195         default:
196                 fallthrough;
197         case SMCCC_RET_NOT_SUPPORTED:
198                 return SPECTRE_VULNERABLE;
199         }
200 }
201
202 bool has_spectre_v2(const struct arm64_cpu_capabilities *entry, int scope)
203 {
204         WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
205
206         if (spectre_v2_get_cpu_hw_mitigation_state() == SPECTRE_UNAFFECTED)
207                 return false;
208
209         if (spectre_v2_get_cpu_fw_mitigation_state() == SPECTRE_UNAFFECTED)
210                 return false;
211
212         return true;
213 }
214
215 enum mitigation_state arm64_get_spectre_v2_state(void)
216 {
217         return spectre_v2_state;
218 }
219
220 DEFINE_PER_CPU_READ_MOSTLY(struct bp_hardening_data, bp_hardening_data);
221
222 static void install_bp_hardening_cb(bp_hardening_cb_t fn)
223 {
224         __this_cpu_write(bp_hardening_data.fn, fn);
225
226         /*
227          * Vinz Clortho takes the hyp_vecs start/end "keys" at
228          * the door when we're a guest. Skip the hyp-vectors work.
229          */
230         if (!is_hyp_mode_available())
231                 return;
232
233         __this_cpu_write(bp_hardening_data.slot, HYP_VECTOR_SPECTRE_DIRECT);
234 }
235
236 /* Called during entry so must be noinstr */
237 static noinstr void call_smc_arch_workaround_1(void)
238 {
239         arm_smccc_1_1_smc(ARM_SMCCC_ARCH_WORKAROUND_1, NULL);
240 }
241
242 /* Called during entry so must be noinstr */
243 static noinstr void call_hvc_arch_workaround_1(void)
244 {
245         arm_smccc_1_1_hvc(ARM_SMCCC_ARCH_WORKAROUND_1, NULL);
246 }
247
248 /* Called during entry so must be noinstr */
249 static noinstr void qcom_link_stack_sanitisation(void)
250 {
251         u64 tmp;
252
253         asm volatile("mov       %0, x30         \n"
254                      ".rept     16              \n"
255                      "bl        . + 4           \n"
256                      ".endr                     \n"
257                      "mov       x30, %0         \n"
258                      : "=&r" (tmp));
259 }
260
261 static bp_hardening_cb_t spectre_v2_get_sw_mitigation_cb(void)
262 {
263         u32 midr = read_cpuid_id();
264         if (((midr & MIDR_CPU_MODEL_MASK) != MIDR_QCOM_FALKOR) &&
265             ((midr & MIDR_CPU_MODEL_MASK) != MIDR_QCOM_FALKOR_V1))
266                 return NULL;
267
268         return qcom_link_stack_sanitisation;
269 }
270
271 static enum mitigation_state spectre_v2_enable_fw_mitigation(void)
272 {
273         bp_hardening_cb_t cb;
274         enum mitigation_state state;
275
276         state = spectre_v2_get_cpu_fw_mitigation_state();
277         if (state != SPECTRE_MITIGATED)
278                 return state;
279
280         if (spectre_v2_mitigations_off())
281                 return SPECTRE_VULNERABLE;
282
283         switch (arm_smccc_1_1_get_conduit()) {
284         case SMCCC_CONDUIT_HVC:
285                 cb = call_hvc_arch_workaround_1;
286                 break;
287
288         case SMCCC_CONDUIT_SMC:
289                 cb = call_smc_arch_workaround_1;
290                 break;
291
292         default:
293                 return SPECTRE_VULNERABLE;
294         }
295
296         /*
297          * Prefer a CPU-specific workaround if it exists. Note that we
298          * still rely on firmware for the mitigation at EL2.
299          */
300         cb = spectre_v2_get_sw_mitigation_cb() ?: cb;
301         install_bp_hardening_cb(cb);
302         return SPECTRE_MITIGATED;
303 }
304
305 void spectre_v2_enable_mitigation(const struct arm64_cpu_capabilities *__unused)
306 {
307         enum mitigation_state state;
308
309         WARN_ON(preemptible());
310
311         state = spectre_v2_get_cpu_hw_mitigation_state();
312         if (state == SPECTRE_VULNERABLE)
313                 state = spectre_v2_enable_fw_mitigation();
314
315         update_mitigation_state(&spectre_v2_state, state);
316 }
317
318 /*
319  * Spectre-v3a.
320  *
321  * Phew, there's not an awful lot to do here! We just instruct EL2 to use
322  * an indirect trampoline for the hyp vectors so that guests can't read
323  * VBAR_EL2 to defeat randomisation of the hypervisor VA layout.
324  */
325 bool has_spectre_v3a(const struct arm64_cpu_capabilities *entry, int scope)
326 {
327         static const struct midr_range spectre_v3a_unsafe_list[] = {
328                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A57),
329                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A72),
330                 {},
331         };
332
333         WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
334         return is_midr_in_range_list(read_cpuid_id(), spectre_v3a_unsafe_list);
335 }
336
337 void spectre_v3a_enable_mitigation(const struct arm64_cpu_capabilities *__unused)
338 {
339         struct bp_hardening_data *data = this_cpu_ptr(&bp_hardening_data);
340
341         if (this_cpu_has_cap(ARM64_SPECTRE_V3A))
342                 data->slot += HYP_VECTOR_INDIRECT;
343 }
344
345 /*
346  * Spectre v4.
347  *
348  * If you thought Spectre v2 was nasty, wait until you see this mess. A CPU is
349  * either:
350  *
351  * - Mitigated in hardware and listed in our "safe list".
352  * - Mitigated in hardware via PSTATE.SSBS.
353  * - Mitigated in software by firmware (sometimes referred to as SSBD).
354  *
355  * Wait, that doesn't sound so bad, does it? Keep reading...
356  *
357  * A major source of headaches is that the software mitigation is enabled both
358  * on a per-task basis, but can also be forced on for the kernel, necessitating
359  * both context-switch *and* entry/exit hooks. To make it even worse, some CPUs
360  * allow EL0 to toggle SSBS directly, which can end up with the prctl() state
361  * being stale when re-entering the kernel. The usual big.LITTLE caveats apply,
362  * so you can have systems that have both firmware and SSBS mitigations. This
363  * means we actually have to reject late onlining of CPUs with mitigations if
364  * all of the currently onlined CPUs are safelisted, as the mitigation tends to
365  * be opt-in for userspace. Yes, really, the cure is worse than the disease.
366  *
367  * The only good part is that if the firmware mitigation is present, then it is
368  * present for all CPUs, meaning we don't have to worry about late onlining of a
369  * vulnerable CPU if one of the boot CPUs is using the firmware mitigation.
370  *
371  * Give me a VAX-11/780 any day of the week...
372  */
373 static enum mitigation_state spectre_v4_state;
374
375 /* This is the per-cpu state tracking whether we need to talk to firmware */
376 DEFINE_PER_CPU_READ_MOSTLY(u64, arm64_ssbd_callback_required);
377
378 enum spectre_v4_policy {
379         SPECTRE_V4_POLICY_MITIGATION_DYNAMIC,
380         SPECTRE_V4_POLICY_MITIGATION_ENABLED,
381         SPECTRE_V4_POLICY_MITIGATION_DISABLED,
382 };
383
384 static enum spectre_v4_policy __read_mostly __spectre_v4_policy;
385
386 static const struct spectre_v4_param {
387         const char              *str;
388         enum spectre_v4_policy  policy;
389 } spectre_v4_params[] = {
390         { "force-on",   SPECTRE_V4_POLICY_MITIGATION_ENABLED, },
391         { "force-off",  SPECTRE_V4_POLICY_MITIGATION_DISABLED, },
392         { "kernel",     SPECTRE_V4_POLICY_MITIGATION_DYNAMIC, },
393 };
394 static int __init parse_spectre_v4_param(char *str)
395 {
396         int i;
397
398         if (!str || !str[0])
399                 return -EINVAL;
400
401         for (i = 0; i < ARRAY_SIZE(spectre_v4_params); i++) {
402                 const struct spectre_v4_param *param = &spectre_v4_params[i];
403
404                 if (strncmp(str, param->str, strlen(param->str)))
405                         continue;
406
407                 __spectre_v4_policy = param->policy;
408                 return 0;
409         }
410
411         return -EINVAL;
412 }
413 early_param("ssbd", parse_spectre_v4_param);
414
415 /*
416  * Because this was all written in a rush by people working in different silos,
417  * we've ended up with multiple command line options to control the same thing.
418  * Wrap these up in some helpers, which prefer disabling the mitigation if faced
419  * with contradictory parameters. The mitigation is always either "off",
420  * "dynamic" or "on".
421  */
422 static bool spectre_v4_mitigations_off(void)
423 {
424         bool ret = cpu_mitigations_off() ||
425                    __spectre_v4_policy == SPECTRE_V4_POLICY_MITIGATION_DISABLED;
426
427         if (ret)
428                 pr_info_once("spectre-v4 mitigation disabled by command-line option\n");
429
430         return ret;
431 }
432
433 /* Do we need to toggle the mitigation state on entry to/exit from the kernel? */
434 static bool spectre_v4_mitigations_dynamic(void)
435 {
436         return !spectre_v4_mitigations_off() &&
437                __spectre_v4_policy == SPECTRE_V4_POLICY_MITIGATION_DYNAMIC;
438 }
439
440 static bool spectre_v4_mitigations_on(void)
441 {
442         return !spectre_v4_mitigations_off() &&
443                __spectre_v4_policy == SPECTRE_V4_POLICY_MITIGATION_ENABLED;
444 }
445
446 ssize_t cpu_show_spec_store_bypass(struct device *dev,
447                                    struct device_attribute *attr, char *buf)
448 {
449         switch (spectre_v4_state) {
450         case SPECTRE_UNAFFECTED:
451                 return sprintf(buf, "Not affected\n");
452         case SPECTRE_MITIGATED:
453                 return sprintf(buf, "Mitigation: Speculative Store Bypass disabled via prctl\n");
454         case SPECTRE_VULNERABLE:
455                 fallthrough;
456         default:
457                 return sprintf(buf, "Vulnerable\n");
458         }
459 }
460
461 enum mitigation_state arm64_get_spectre_v4_state(void)
462 {
463         return spectre_v4_state;
464 }
465
466 static enum mitigation_state spectre_v4_get_cpu_hw_mitigation_state(void)
467 {
468         static const struct midr_range spectre_v4_safe_list[] = {
469                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A35),
470                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A53),
471                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A55),
472                 MIDR_ALL_VERSIONS(MIDR_BRAHMA_B53),
473                 MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_3XX_SILVER),
474                 MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_4XX_SILVER),
475                 { /* sentinel */ },
476         };
477
478         if (is_midr_in_range_list(read_cpuid_id(), spectre_v4_safe_list))
479                 return SPECTRE_UNAFFECTED;
480
481         /* CPU features are detected first */
482         if (this_cpu_has_cap(ARM64_SSBS))
483                 return SPECTRE_MITIGATED;
484
485         return SPECTRE_VULNERABLE;
486 }
487
488 static enum mitigation_state spectre_v4_get_cpu_fw_mitigation_state(void)
489 {
490         int ret;
491         struct arm_smccc_res res;
492
493         arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
494                              ARM_SMCCC_ARCH_WORKAROUND_2, &res);
495
496         ret = res.a0;
497         switch (ret) {
498         case SMCCC_RET_SUCCESS:
499                 return SPECTRE_MITIGATED;
500         case SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED:
501                 fallthrough;
502         case SMCCC_RET_NOT_REQUIRED:
503                 return SPECTRE_UNAFFECTED;
504         default:
505                 fallthrough;
506         case SMCCC_RET_NOT_SUPPORTED:
507                 return SPECTRE_VULNERABLE;
508         }
509 }
510
511 bool has_spectre_v4(const struct arm64_cpu_capabilities *cap, int scope)
512 {
513         enum mitigation_state state;
514
515         WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
516
517         state = spectre_v4_get_cpu_hw_mitigation_state();
518         if (state == SPECTRE_VULNERABLE)
519                 state = spectre_v4_get_cpu_fw_mitigation_state();
520
521         return state != SPECTRE_UNAFFECTED;
522 }
523
524 static int ssbs_emulation_handler(struct pt_regs *regs, u32 instr)
525 {
526         if (user_mode(regs))
527                 return 1;
528
529         if (instr & BIT(PSTATE_Imm_shift))
530                 regs->pstate |= PSR_SSBS_BIT;
531         else
532                 regs->pstate &= ~PSR_SSBS_BIT;
533
534         arm64_skip_faulting_instruction(regs, 4);
535         return 0;
536 }
537
538 static struct undef_hook ssbs_emulation_hook = {
539         .instr_mask     = ~(1U << PSTATE_Imm_shift),
540         .instr_val      = 0xd500401f | PSTATE_SSBS,
541         .fn             = ssbs_emulation_handler,
542 };
543
544 static enum mitigation_state spectre_v4_enable_hw_mitigation(void)
545 {
546         static bool undef_hook_registered = false;
547         static DEFINE_RAW_SPINLOCK(hook_lock);
548         enum mitigation_state state;
549
550         /*
551          * If the system is mitigated but this CPU doesn't have SSBS, then
552          * we must be on the safelist and there's nothing more to do.
553          */
554         state = spectre_v4_get_cpu_hw_mitigation_state();
555         if (state != SPECTRE_MITIGATED || !this_cpu_has_cap(ARM64_SSBS))
556                 return state;
557
558         raw_spin_lock(&hook_lock);
559         if (!undef_hook_registered) {
560                 register_undef_hook(&ssbs_emulation_hook);
561                 undef_hook_registered = true;
562         }
563         raw_spin_unlock(&hook_lock);
564
565         if (spectre_v4_mitigations_off()) {
566                 sysreg_clear_set(sctlr_el1, 0, SCTLR_ELx_DSSBS);
567                 set_pstate_ssbs(1);
568                 return SPECTRE_VULNERABLE;
569         }
570
571         /* SCTLR_EL1.DSSBS was initialised to 0 during boot */
572         set_pstate_ssbs(0);
573         return SPECTRE_MITIGATED;
574 }
575
576 /*
577  * Patch a branch over the Spectre-v4 mitigation code with a NOP so that
578  * we fallthrough and check whether firmware needs to be called on this CPU.
579  */
580 void __init spectre_v4_patch_fw_mitigation_enable(struct alt_instr *alt,
581                                                   __le32 *origptr,
582                                                   __le32 *updptr, int nr_inst)
583 {
584         BUG_ON(nr_inst != 1); /* Branch -> NOP */
585
586         if (spectre_v4_mitigations_off())
587                 return;
588
589         if (cpus_have_cap(ARM64_SSBS))
590                 return;
591
592         if (spectre_v4_mitigations_dynamic())
593                 *updptr = cpu_to_le32(aarch64_insn_gen_nop());
594 }
595
596 /*
597  * Patch a NOP in the Spectre-v4 mitigation code with an SMC/HVC instruction
598  * to call into firmware to adjust the mitigation state.
599  */
600 void __init smccc_patch_fw_mitigation_conduit(struct alt_instr *alt,
601                                                __le32 *origptr,
602                                                __le32 *updptr, int nr_inst)
603 {
604         u32 insn;
605
606         BUG_ON(nr_inst != 1); /* NOP -> HVC/SMC */
607
608         switch (arm_smccc_1_1_get_conduit()) {
609         case SMCCC_CONDUIT_HVC:
610                 insn = aarch64_insn_get_hvc_value();
611                 break;
612         case SMCCC_CONDUIT_SMC:
613                 insn = aarch64_insn_get_smc_value();
614                 break;
615         default:
616                 return;
617         }
618
619         *updptr = cpu_to_le32(insn);
620 }
621
622 static enum mitigation_state spectre_v4_enable_fw_mitigation(void)
623 {
624         enum mitigation_state state;
625
626         state = spectre_v4_get_cpu_fw_mitigation_state();
627         if (state != SPECTRE_MITIGATED)
628                 return state;
629
630         if (spectre_v4_mitigations_off()) {
631                 arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_WORKAROUND_2, false, NULL);
632                 return SPECTRE_VULNERABLE;
633         }
634
635         arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_WORKAROUND_2, true, NULL);
636
637         if (spectre_v4_mitigations_dynamic())
638                 __this_cpu_write(arm64_ssbd_callback_required, 1);
639
640         return SPECTRE_MITIGATED;
641 }
642
643 void spectre_v4_enable_mitigation(const struct arm64_cpu_capabilities *__unused)
644 {
645         enum mitigation_state state;
646
647         WARN_ON(preemptible());
648
649         state = spectre_v4_enable_hw_mitigation();
650         if (state == SPECTRE_VULNERABLE)
651                 state = spectre_v4_enable_fw_mitigation();
652
653         update_mitigation_state(&spectre_v4_state, state);
654 }
655
656 static void __update_pstate_ssbs(struct pt_regs *regs, bool state)
657 {
658         u64 bit = compat_user_mode(regs) ? PSR_AA32_SSBS_BIT : PSR_SSBS_BIT;
659
660         if (state)
661                 regs->pstate |= bit;
662         else
663                 regs->pstate &= ~bit;
664 }
665
666 void spectre_v4_enable_task_mitigation(struct task_struct *tsk)
667 {
668         struct pt_regs *regs = task_pt_regs(tsk);
669         bool ssbs = false, kthread = tsk->flags & PF_KTHREAD;
670
671         if (spectre_v4_mitigations_off())
672                 ssbs = true;
673         else if (spectre_v4_mitigations_dynamic() && !kthread)
674                 ssbs = !test_tsk_thread_flag(tsk, TIF_SSBD);
675
676         __update_pstate_ssbs(regs, ssbs);
677 }
678
679 /*
680  * The Spectre-v4 mitigation can be controlled via a prctl() from userspace.
681  * This is interesting because the "speculation disabled" behaviour can be
682  * configured so that it is preserved across exec(), which means that the
683  * prctl() may be necessary even when PSTATE.SSBS can be toggled directly
684  * from userspace.
685  */
686 static void ssbd_prctl_enable_mitigation(struct task_struct *task)
687 {
688         task_clear_spec_ssb_noexec(task);
689         task_set_spec_ssb_disable(task);
690         set_tsk_thread_flag(task, TIF_SSBD);
691 }
692
693 static void ssbd_prctl_disable_mitigation(struct task_struct *task)
694 {
695         task_clear_spec_ssb_noexec(task);
696         task_clear_spec_ssb_disable(task);
697         clear_tsk_thread_flag(task, TIF_SSBD);
698 }
699
700 static int ssbd_prctl_set(struct task_struct *task, unsigned long ctrl)
701 {
702         switch (ctrl) {
703         case PR_SPEC_ENABLE:
704                 /* Enable speculation: disable mitigation */
705                 /*
706                  * Force disabled speculation prevents it from being
707                  * re-enabled.
708                  */
709                 if (task_spec_ssb_force_disable(task))
710                         return -EPERM;
711
712                 /*
713                  * If the mitigation is forced on, then speculation is forced
714                  * off and we again prevent it from being re-enabled.
715                  */
716                 if (spectre_v4_mitigations_on())
717                         return -EPERM;
718
719                 ssbd_prctl_disable_mitigation(task);
720                 break;
721         case PR_SPEC_FORCE_DISABLE:
722                 /* Force disable speculation: force enable mitigation */
723                 /*
724                  * If the mitigation is forced off, then speculation is forced
725                  * on and we prevent it from being disabled.
726                  */
727                 if (spectre_v4_mitigations_off())
728                         return -EPERM;
729
730                 task_set_spec_ssb_force_disable(task);
731                 fallthrough;
732         case PR_SPEC_DISABLE:
733                 /* Disable speculation: enable mitigation */
734                 /* Same as PR_SPEC_FORCE_DISABLE */
735                 if (spectre_v4_mitigations_off())
736                         return -EPERM;
737
738                 ssbd_prctl_enable_mitigation(task);
739                 break;
740         case PR_SPEC_DISABLE_NOEXEC:
741                 /* Disable speculation until execve(): enable mitigation */
742                 /*
743                  * If the mitigation state is forced one way or the other, then
744                  * we must fail now before we try to toggle it on execve().
745                  */
746                 if (task_spec_ssb_force_disable(task) ||
747                     spectre_v4_mitigations_off() ||
748                     spectre_v4_mitigations_on()) {
749                         return -EPERM;
750                 }
751
752                 ssbd_prctl_enable_mitigation(task);
753                 task_set_spec_ssb_noexec(task);
754                 break;
755         default:
756                 return -ERANGE;
757         }
758
759         spectre_v4_enable_task_mitigation(task);
760         return 0;
761 }
762
763 int arch_prctl_spec_ctrl_set(struct task_struct *task, unsigned long which,
764                              unsigned long ctrl)
765 {
766         switch (which) {
767         case PR_SPEC_STORE_BYPASS:
768                 return ssbd_prctl_set(task, ctrl);
769         default:
770                 return -ENODEV;
771         }
772 }
773
774 static int ssbd_prctl_get(struct task_struct *task)
775 {
776         switch (spectre_v4_state) {
777         case SPECTRE_UNAFFECTED:
778                 return PR_SPEC_NOT_AFFECTED;
779         case SPECTRE_MITIGATED:
780                 if (spectre_v4_mitigations_on())
781                         return PR_SPEC_NOT_AFFECTED;
782
783                 if (spectre_v4_mitigations_dynamic())
784                         break;
785
786                 /* Mitigations are disabled, so we're vulnerable. */
787                 fallthrough;
788         case SPECTRE_VULNERABLE:
789                 fallthrough;
790         default:
791                 return PR_SPEC_ENABLE;
792         }
793
794         /* Check the mitigation state for this task */
795         if (task_spec_ssb_force_disable(task))
796                 return PR_SPEC_PRCTL | PR_SPEC_FORCE_DISABLE;
797
798         if (task_spec_ssb_noexec(task))
799                 return PR_SPEC_PRCTL | PR_SPEC_DISABLE_NOEXEC;
800
801         if (task_spec_ssb_disable(task))
802                 return PR_SPEC_PRCTL | PR_SPEC_DISABLE;
803
804         return PR_SPEC_PRCTL | PR_SPEC_ENABLE;
805 }
806
807 int arch_prctl_spec_ctrl_get(struct task_struct *task, unsigned long which)
808 {
809         switch (which) {
810         case PR_SPEC_STORE_BYPASS:
811                 return ssbd_prctl_get(task);
812         default:
813                 return -ENODEV;
814         }
815 }
816
817 /*
818  * Spectre BHB.
819  *
820  * A CPU is either:
821  * - Mitigated by a branchy loop a CPU specific number of times, and listed
822  *   in our "loop mitigated list".
823  * - Mitigated in software by the firmware Spectre v2 call.
824  * - Has the ClearBHB instruction to perform the mitigation.
825  * - Has the 'Exception Clears Branch History Buffer' (ECBHB) feature, so no
826  *   software mitigation in the vectors is needed.
827  * - Has CSV2.3, so is unaffected.
828  */
829 static enum mitigation_state spectre_bhb_state;
830
831 enum mitigation_state arm64_get_spectre_bhb_state(void)
832 {
833         return spectre_bhb_state;
834 }
835
836 enum bhb_mitigation_bits {
837         BHB_LOOP,
838         BHB_FW,
839         BHB_HW,
840         BHB_INSN,
841 };
842 static unsigned long system_bhb_mitigations;
843
844 /*
845  * This must be called with SCOPE_LOCAL_CPU for each type of CPU, before any
846  * SCOPE_SYSTEM call will give the right answer.
847  */
848 u8 spectre_bhb_loop_affected(int scope)
849 {
850         u8 k = 0;
851         static u8 max_bhb_k;
852
853         if (scope == SCOPE_LOCAL_CPU) {
854                 static const struct midr_range spectre_bhb_k32_list[] = {
855                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A78),
856                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A78AE),
857                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A78C),
858                         MIDR_ALL_VERSIONS(MIDR_CORTEX_X1),
859                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A710),
860                         MIDR_ALL_VERSIONS(MIDR_CORTEX_X2),
861                         MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N2),
862                         MIDR_ALL_VERSIONS(MIDR_NEOVERSE_V1),
863                         {},
864                 };
865                 static const struct midr_range spectre_bhb_k24_list[] = {
866                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A76),
867                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A77),
868                         MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N1),
869                         {},
870                 };
871                 static const struct midr_range spectre_bhb_k11_list[] = {
872                         MIDR_ALL_VERSIONS(MIDR_AMPERE1),
873                         {},
874                 };
875                 static const struct midr_range spectre_bhb_k8_list[] = {
876                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A72),
877                         MIDR_ALL_VERSIONS(MIDR_CORTEX_A57),
878                         {},
879                 };
880
881                 if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k32_list))
882                         k = 32;
883                 else if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k24_list))
884                         k = 24;
885                 else if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k11_list))
886                         k = 11;
887                 else if (is_midr_in_range_list(read_cpuid_id(), spectre_bhb_k8_list))
888                         k =  8;
889
890                 max_bhb_k = max(max_bhb_k, k);
891         } else {
892                 k = max_bhb_k;
893         }
894
895         return k;
896 }
897
898 static enum mitigation_state spectre_bhb_get_cpu_fw_mitigation_state(void)
899 {
900         int ret;
901         struct arm_smccc_res res;
902
903         arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
904                              ARM_SMCCC_ARCH_WORKAROUND_3, &res);
905
906         ret = res.a0;
907         switch (ret) {
908         case SMCCC_RET_SUCCESS:
909                 return SPECTRE_MITIGATED;
910         case SMCCC_ARCH_WORKAROUND_RET_UNAFFECTED:
911                 return SPECTRE_UNAFFECTED;
912         default:
913                 fallthrough;
914         case SMCCC_RET_NOT_SUPPORTED:
915                 return SPECTRE_VULNERABLE;
916         }
917 }
918
919 static bool is_spectre_bhb_fw_affected(int scope)
920 {
921         static bool system_affected;
922         enum mitigation_state fw_state;
923         bool has_smccc = arm_smccc_1_1_get_conduit() != SMCCC_CONDUIT_NONE;
924         static const struct midr_range spectre_bhb_firmware_mitigated_list[] = {
925                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A73),
926                 MIDR_ALL_VERSIONS(MIDR_CORTEX_A75),
927                 {},
928         };
929         bool cpu_in_list = is_midr_in_range_list(read_cpuid_id(),
930                                          spectre_bhb_firmware_mitigated_list);
931
932         if (scope != SCOPE_LOCAL_CPU)
933                 return system_affected;
934
935         fw_state = spectre_bhb_get_cpu_fw_mitigation_state();
936         if (cpu_in_list || (has_smccc && fw_state == SPECTRE_MITIGATED)) {
937                 system_affected = true;
938                 return true;
939         }
940
941         return false;
942 }
943
944 static bool supports_ecbhb(int scope)
945 {
946         u64 mmfr1;
947
948         if (scope == SCOPE_LOCAL_CPU)
949                 mmfr1 = read_sysreg_s(SYS_ID_AA64MMFR1_EL1);
950         else
951                 mmfr1 = read_sanitised_ftr_reg(SYS_ID_AA64MMFR1_EL1);
952
953         return cpuid_feature_extract_unsigned_field(mmfr1,
954                                                     ID_AA64MMFR1_EL1_ECBHB_SHIFT);
955 }
956
957 bool is_spectre_bhb_affected(const struct arm64_cpu_capabilities *entry,
958                              int scope)
959 {
960         WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
961
962         if (supports_csv2p3(scope))
963                 return false;
964
965         if (supports_clearbhb(scope))
966                 return true;
967
968         if (spectre_bhb_loop_affected(scope))
969                 return true;
970
971         if (is_spectre_bhb_fw_affected(scope))
972                 return true;
973
974         return false;
975 }
976
977 static void this_cpu_set_vectors(enum arm64_bp_harden_el1_vectors slot)
978 {
979         const char *v = arm64_get_bp_hardening_vector(slot);
980
981         if (slot < 0)
982                 return;
983
984         __this_cpu_write(this_cpu_vector, v);
985
986         /*
987          * When KPTI is in use, the vectors are switched when exiting to
988          * user-space.
989          */
990         if (arm64_kernel_unmapped_at_el0())
991                 return;
992
993         write_sysreg(v, vbar_el1);
994         isb();
995 }
996
997 static bool __read_mostly __nospectre_bhb;
998 static int __init parse_spectre_bhb_param(char *str)
999 {
1000         __nospectre_bhb = true;
1001         return 0;
1002 }
1003 early_param("nospectre_bhb", parse_spectre_bhb_param);
1004
1005 void spectre_bhb_enable_mitigation(const struct arm64_cpu_capabilities *entry)
1006 {
1007         bp_hardening_cb_t cpu_cb;
1008         enum mitigation_state fw_state, state = SPECTRE_VULNERABLE;
1009         struct bp_hardening_data *data = this_cpu_ptr(&bp_hardening_data);
1010
1011         if (!is_spectre_bhb_affected(entry, SCOPE_LOCAL_CPU))
1012                 return;
1013
1014         if (arm64_get_spectre_v2_state() == SPECTRE_VULNERABLE) {
1015                 /* No point mitigating Spectre-BHB alone. */
1016         } else if (!IS_ENABLED(CONFIG_MITIGATE_SPECTRE_BRANCH_HISTORY)) {
1017                 pr_info_once("spectre-bhb mitigation disabled by compile time option\n");
1018         } else if (cpu_mitigations_off() || __nospectre_bhb) {
1019                 pr_info_once("spectre-bhb mitigation disabled by command line option\n");
1020         } else if (supports_ecbhb(SCOPE_LOCAL_CPU)) {
1021                 state = SPECTRE_MITIGATED;
1022                 set_bit(BHB_HW, &system_bhb_mitigations);
1023         } else if (supports_clearbhb(SCOPE_LOCAL_CPU)) {
1024                 /*
1025                  * Ensure KVM uses the indirect vector which will have ClearBHB
1026                  * added.
1027                  */
1028                 if (!data->slot)
1029                         data->slot = HYP_VECTOR_INDIRECT;
1030
1031                 this_cpu_set_vectors(EL1_VECTOR_BHB_CLEAR_INSN);
1032                 state = SPECTRE_MITIGATED;
1033                 set_bit(BHB_INSN, &system_bhb_mitigations);
1034         } else if (spectre_bhb_loop_affected(SCOPE_LOCAL_CPU)) {
1035                 /*
1036                  * Ensure KVM uses the indirect vector which will have the
1037                  * branchy-loop added. A57/A72-r0 will already have selected
1038                  * the spectre-indirect vector, which is sufficient for BHB
1039                  * too.
1040                  */
1041                 if (!data->slot)
1042                         data->slot = HYP_VECTOR_INDIRECT;
1043
1044                 this_cpu_set_vectors(EL1_VECTOR_BHB_LOOP);
1045                 state = SPECTRE_MITIGATED;
1046                 set_bit(BHB_LOOP, &system_bhb_mitigations);
1047         } else if (is_spectre_bhb_fw_affected(SCOPE_LOCAL_CPU)) {
1048                 fw_state = spectre_bhb_get_cpu_fw_mitigation_state();
1049                 if (fw_state == SPECTRE_MITIGATED) {
1050                         /*
1051                          * Ensure KVM uses one of the spectre bp_hardening
1052                          * vectors. The indirect vector doesn't include the EL3
1053                          * call, so needs upgrading to
1054                          * HYP_VECTOR_SPECTRE_INDIRECT.
1055                          */
1056                         if (!data->slot || data->slot == HYP_VECTOR_INDIRECT)
1057                                 data->slot += 1;
1058
1059                         this_cpu_set_vectors(EL1_VECTOR_BHB_FW);
1060
1061                         /*
1062                          * The WA3 call in the vectors supersedes the WA1 call
1063                          * made during context-switch. Uninstall any firmware
1064                          * bp_hardening callback.
1065                          */
1066                         cpu_cb = spectre_v2_get_sw_mitigation_cb();
1067                         if (__this_cpu_read(bp_hardening_data.fn) != cpu_cb)
1068                                 __this_cpu_write(bp_hardening_data.fn, NULL);
1069
1070                         state = SPECTRE_MITIGATED;
1071                         set_bit(BHB_FW, &system_bhb_mitigations);
1072                 }
1073         }
1074
1075         update_mitigation_state(&spectre_bhb_state, state);
1076 }
1077
1078 /* Patched to NOP when enabled */
1079 void noinstr spectre_bhb_patch_loop_mitigation_enable(struct alt_instr *alt,
1080                                                      __le32 *origptr,
1081                                                       __le32 *updptr, int nr_inst)
1082 {
1083         BUG_ON(nr_inst != 1);
1084
1085         if (test_bit(BHB_LOOP, &system_bhb_mitigations))
1086                 *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
1087 }
1088
1089 /* Patched to NOP when enabled */
1090 void noinstr spectre_bhb_patch_fw_mitigation_enabled(struct alt_instr *alt,
1091                                                    __le32 *origptr,
1092                                                    __le32 *updptr, int nr_inst)
1093 {
1094         BUG_ON(nr_inst != 1);
1095
1096         if (test_bit(BHB_FW, &system_bhb_mitigations))
1097                 *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
1098 }
1099
1100 /* Patched to correct the immediate */
1101 void noinstr spectre_bhb_patch_loop_iter(struct alt_instr *alt,
1102                                    __le32 *origptr, __le32 *updptr, int nr_inst)
1103 {
1104         u8 rd;
1105         u32 insn;
1106         u16 loop_count = spectre_bhb_loop_affected(SCOPE_SYSTEM);
1107
1108         BUG_ON(nr_inst != 1); /* MOV -> MOV */
1109
1110         if (!IS_ENABLED(CONFIG_MITIGATE_SPECTRE_BRANCH_HISTORY))
1111                 return;
1112
1113         insn = le32_to_cpu(*origptr);
1114         rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, insn);
1115         insn = aarch64_insn_gen_movewide(rd, loop_count, 0,
1116                                          AARCH64_INSN_VARIANT_64BIT,
1117                                          AARCH64_INSN_MOVEWIDE_ZERO);
1118         *updptr++ = cpu_to_le32(insn);
1119 }
1120
1121 /* Patched to mov WA3 when supported */
1122 void noinstr spectre_bhb_patch_wa3(struct alt_instr *alt,
1123                                    __le32 *origptr, __le32 *updptr, int nr_inst)
1124 {
1125         u8 rd;
1126         u32 insn;
1127
1128         BUG_ON(nr_inst != 1); /* MOV -> MOV */
1129
1130         if (!IS_ENABLED(CONFIG_MITIGATE_SPECTRE_BRANCH_HISTORY) ||
1131             !test_bit(BHB_FW, &system_bhb_mitigations))
1132                 return;
1133
1134         insn = le32_to_cpu(*origptr);
1135         rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, insn);
1136
1137         insn = aarch64_insn_gen_logical_immediate(AARCH64_INSN_LOGIC_ORR,
1138                                                   AARCH64_INSN_VARIANT_32BIT,
1139                                                   AARCH64_INSN_REG_ZR, rd,
1140                                                   ARM_SMCCC_ARCH_WORKAROUND_3);
1141         if (WARN_ON_ONCE(insn == AARCH64_BREAK_FAULT))
1142                 return;
1143
1144         *updptr++ = cpu_to_le32(insn);
1145 }
1146
1147 /* Patched to NOP when not supported */
1148 void __init spectre_bhb_patch_clearbhb(struct alt_instr *alt,
1149                                    __le32 *origptr, __le32 *updptr, int nr_inst)
1150 {
1151         BUG_ON(nr_inst != 2);
1152
1153         if (test_bit(BHB_INSN, &system_bhb_mitigations))
1154                 return;
1155
1156         *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
1157         *updptr++ = cpu_to_le32(aarch64_insn_gen_nop());
1158 }
1159
1160 #ifdef CONFIG_BPF_SYSCALL
1161 #define EBPF_WARN "Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks!\n"
1162 void unpriv_ebpf_notify(int new_state)
1163 {
1164         if (spectre_v2_state == SPECTRE_VULNERABLE ||
1165             spectre_bhb_state != SPECTRE_MITIGATED)
1166                 return;
1167
1168         if (!new_state)
1169                 pr_err("WARNING: %s", EBPF_WARN);
1170 }
1171 #endif