ceph: no need to wait for transition RDCACHE|RD -> RD
[platform/kernel/linux-rpi.git] / kernel / static_call_inline.c
1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/init.h>
3 #include <linux/static_call.h>
4 #include <linux/bug.h>
5 #include <linux/smp.h>
6 #include <linux/sort.h>
7 #include <linux/slab.h>
8 #include <linux/module.h>
9 #include <linux/cpu.h>
10 #include <linux/processor.h>
11 #include <asm/sections.h>
12
13 extern struct static_call_site __start_static_call_sites[],
14                                __stop_static_call_sites[];
15 extern struct static_call_tramp_key __start_static_call_tramp_key[],
16                                     __stop_static_call_tramp_key[];
17
18 static bool static_call_initialized;
19
20 /* mutex to protect key modules/sites */
21 static DEFINE_MUTEX(static_call_mutex);
22
23 static void static_call_lock(void)
24 {
25         mutex_lock(&static_call_mutex);
26 }
27
28 static void static_call_unlock(void)
29 {
30         mutex_unlock(&static_call_mutex);
31 }
32
33 static inline void *static_call_addr(struct static_call_site *site)
34 {
35         return (void *)((long)site->addr + (long)&site->addr);
36 }
37
38 static inline unsigned long __static_call_key(const struct static_call_site *site)
39 {
40         return (long)site->key + (long)&site->key;
41 }
42
43 static inline struct static_call_key *static_call_key(const struct static_call_site *site)
44 {
45         return (void *)(__static_call_key(site) & ~STATIC_CALL_SITE_FLAGS);
46 }
47
48 /* These assume the key is word-aligned. */
49 static inline bool static_call_is_init(struct static_call_site *site)
50 {
51         return __static_call_key(site) & STATIC_CALL_SITE_INIT;
52 }
53
54 static inline bool static_call_is_tail(struct static_call_site *site)
55 {
56         return __static_call_key(site) & STATIC_CALL_SITE_TAIL;
57 }
58
59 static inline void static_call_set_init(struct static_call_site *site)
60 {
61         site->key = (__static_call_key(site) | STATIC_CALL_SITE_INIT) -
62                     (long)&site->key;
63 }
64
65 static int static_call_site_cmp(const void *_a, const void *_b)
66 {
67         const struct static_call_site *a = _a;
68         const struct static_call_site *b = _b;
69         const struct static_call_key *key_a = static_call_key(a);
70         const struct static_call_key *key_b = static_call_key(b);
71
72         if (key_a < key_b)
73                 return -1;
74
75         if (key_a > key_b)
76                 return 1;
77
78         return 0;
79 }
80
81 static void static_call_site_swap(void *_a, void *_b, int size)
82 {
83         long delta = (unsigned long)_a - (unsigned long)_b;
84         struct static_call_site *a = _a;
85         struct static_call_site *b = _b;
86         struct static_call_site tmp = *a;
87
88         a->addr = b->addr  - delta;
89         a->key  = b->key   - delta;
90
91         b->addr = tmp.addr + delta;
92         b->key  = tmp.key  + delta;
93 }
94
95 static inline void static_call_sort_entries(struct static_call_site *start,
96                                             struct static_call_site *stop)
97 {
98         sort(start, stop - start, sizeof(struct static_call_site),
99              static_call_site_cmp, static_call_site_swap);
100 }
101
102 static inline bool static_call_key_has_mods(struct static_call_key *key)
103 {
104         return !(key->type & 1);
105 }
106
107 static inline struct static_call_mod *static_call_key_next(struct static_call_key *key)
108 {
109         if (!static_call_key_has_mods(key))
110                 return NULL;
111
112         return key->mods;
113 }
114
115 static inline struct static_call_site *static_call_key_sites(struct static_call_key *key)
116 {
117         if (static_call_key_has_mods(key))
118                 return NULL;
119
120         return (struct static_call_site *)(key->type & ~1);
121 }
122
123 void __static_call_update(struct static_call_key *key, void *tramp, void *func)
124 {
125         struct static_call_site *site, *stop;
126         struct static_call_mod *site_mod, first;
127
128         cpus_read_lock();
129         static_call_lock();
130
131         if (key->func == func)
132                 goto done;
133
134         key->func = func;
135
136         arch_static_call_transform(NULL, tramp, func, false);
137
138         /*
139          * If uninitialized, we'll not update the callsites, but they still
140          * point to the trampoline and we just patched that.
141          */
142         if (WARN_ON_ONCE(!static_call_initialized))
143                 goto done;
144
145         first = (struct static_call_mod){
146                 .next = static_call_key_next(key),
147                 .mod = NULL,
148                 .sites = static_call_key_sites(key),
149         };
150
151         for (site_mod = &first; site_mod; site_mod = site_mod->next) {
152                 bool init = system_state < SYSTEM_RUNNING;
153                 struct module *mod = site_mod->mod;
154
155                 if (!site_mod->sites) {
156                         /*
157                          * This can happen if the static call key is defined in
158                          * a module which doesn't use it.
159                          *
160                          * It also happens in the has_mods case, where the
161                          * 'first' entry has no sites associated with it.
162                          */
163                         continue;
164                 }
165
166                 stop = __stop_static_call_sites;
167
168                 if (mod) {
169 #ifdef CONFIG_MODULES
170                         stop = mod->static_call_sites +
171                                mod->num_static_call_sites;
172                         init = mod->state == MODULE_STATE_COMING;
173 #endif
174                 }
175
176                 for (site = site_mod->sites;
177                      site < stop && static_call_key(site) == key; site++) {
178                         void *site_addr = static_call_addr(site);
179
180                         if (!init && static_call_is_init(site))
181                                 continue;
182
183                         if (!kernel_text_address((unsigned long)site_addr)) {
184                                 /*
185                                  * This skips patching built-in __exit, which
186                                  * is part of init_section_contains() but is
187                                  * not part of kernel_text_address().
188                                  *
189                                  * Skipping built-in __exit is fine since it
190                                  * will never be executed.
191                                  */
192                                 WARN_ONCE(!static_call_is_init(site),
193                                           "can't patch static call site at %pS",
194                                           site_addr);
195                                 continue;
196                         }
197
198                         arch_static_call_transform(site_addr, NULL, func,
199                                                    static_call_is_tail(site));
200                 }
201         }
202
203 done:
204         static_call_unlock();
205         cpus_read_unlock();
206 }
207 EXPORT_SYMBOL_GPL(__static_call_update);
208
209 static int __static_call_init(struct module *mod,
210                               struct static_call_site *start,
211                               struct static_call_site *stop)
212 {
213         struct static_call_site *site;
214         struct static_call_key *key, *prev_key = NULL;
215         struct static_call_mod *site_mod;
216
217         if (start == stop)
218                 return 0;
219
220         static_call_sort_entries(start, stop);
221
222         for (site = start; site < stop; site++) {
223                 void *site_addr = static_call_addr(site);
224
225                 if ((mod && within_module_init((unsigned long)site_addr, mod)) ||
226                     (!mod && init_section_contains(site_addr, 1)))
227                         static_call_set_init(site);
228
229                 key = static_call_key(site);
230                 if (key != prev_key) {
231                         prev_key = key;
232
233                         /*
234                          * For vmlinux (!mod) avoid the allocation by storing
235                          * the sites pointer in the key itself. Also see
236                          * __static_call_update()'s @first.
237                          *
238                          * This allows architectures (eg. x86) to call
239                          * static_call_init() before memory allocation works.
240                          */
241                         if (!mod) {
242                                 key->sites = site;
243                                 key->type |= 1;
244                                 goto do_transform;
245                         }
246
247                         site_mod = kzalloc(sizeof(*site_mod), GFP_KERNEL);
248                         if (!site_mod)
249                                 return -ENOMEM;
250
251                         /*
252                          * When the key has a direct sites pointer, extract
253                          * that into an explicit struct static_call_mod, so we
254                          * can have a list of modules.
255                          */
256                         if (static_call_key_sites(key)) {
257                                 site_mod->mod = NULL;
258                                 site_mod->next = NULL;
259                                 site_mod->sites = static_call_key_sites(key);
260
261                                 key->mods = site_mod;
262
263                                 site_mod = kzalloc(sizeof(*site_mod), GFP_KERNEL);
264                                 if (!site_mod)
265                                         return -ENOMEM;
266                         }
267
268                         site_mod->mod = mod;
269                         site_mod->sites = site;
270                         site_mod->next = static_call_key_next(key);
271                         key->mods = site_mod;
272                 }
273
274 do_transform:
275                 arch_static_call_transform(site_addr, NULL, key->func,
276                                 static_call_is_tail(site));
277         }
278
279         return 0;
280 }
281
282 static int addr_conflict(struct static_call_site *site, void *start, void *end)
283 {
284         unsigned long addr = (unsigned long)static_call_addr(site);
285
286         if (addr <= (unsigned long)end &&
287             addr + CALL_INSN_SIZE > (unsigned long)start)
288                 return 1;
289
290         return 0;
291 }
292
293 static int __static_call_text_reserved(struct static_call_site *iter_start,
294                                        struct static_call_site *iter_stop,
295                                        void *start, void *end, bool init)
296 {
297         struct static_call_site *iter = iter_start;
298
299         while (iter < iter_stop) {
300                 if (init || !static_call_is_init(iter)) {
301                         if (addr_conflict(iter, start, end))
302                                 return 1;
303                 }
304                 iter++;
305         }
306
307         return 0;
308 }
309
310 #ifdef CONFIG_MODULES
311
312 static int __static_call_mod_text_reserved(void *start, void *end)
313 {
314         struct module *mod;
315         int ret;
316
317         preempt_disable();
318         mod = __module_text_address((unsigned long)start);
319         WARN_ON_ONCE(__module_text_address((unsigned long)end) != mod);
320         if (!try_module_get(mod))
321                 mod = NULL;
322         preempt_enable();
323
324         if (!mod)
325                 return 0;
326
327         ret = __static_call_text_reserved(mod->static_call_sites,
328                         mod->static_call_sites + mod->num_static_call_sites,
329                         start, end, mod->state == MODULE_STATE_COMING);
330
331         module_put(mod);
332
333         return ret;
334 }
335
336 static unsigned long tramp_key_lookup(unsigned long addr)
337 {
338         struct static_call_tramp_key *start = __start_static_call_tramp_key;
339         struct static_call_tramp_key *stop = __stop_static_call_tramp_key;
340         struct static_call_tramp_key *tramp_key;
341
342         for (tramp_key = start; tramp_key != stop; tramp_key++) {
343                 unsigned long tramp;
344
345                 tramp = (long)tramp_key->tramp + (long)&tramp_key->tramp;
346                 if (tramp == addr)
347                         return (long)tramp_key->key + (long)&tramp_key->key;
348         }
349
350         return 0;
351 }
352
353 static int static_call_add_module(struct module *mod)
354 {
355         struct static_call_site *start = mod->static_call_sites;
356         struct static_call_site *stop = start + mod->num_static_call_sites;
357         struct static_call_site *site;
358
359         for (site = start; site != stop; site++) {
360                 unsigned long s_key = __static_call_key(site);
361                 unsigned long addr = s_key & ~STATIC_CALL_SITE_FLAGS;
362                 unsigned long key;
363
364                 /*
365                  * Is the key is exported, 'addr' points to the key, which
366                  * means modules are allowed to call static_call_update() on
367                  * it.
368                  *
369                  * Otherwise, the key isn't exported, and 'addr' points to the
370                  * trampoline so we need to lookup the key.
371                  *
372                  * We go through this dance to prevent crazy modules from
373                  * abusing sensitive static calls.
374                  */
375                 if (!kernel_text_address(addr))
376                         continue;
377
378                 key = tramp_key_lookup(addr);
379                 if (!key) {
380                         pr_warn("Failed to fixup __raw_static_call() usage at: %ps\n",
381                                 static_call_addr(site));
382                         return -EINVAL;
383                 }
384
385                 key |= s_key & STATIC_CALL_SITE_FLAGS;
386                 site->key = key - (long)&site->key;
387         }
388
389         return __static_call_init(mod, start, stop);
390 }
391
392 static void static_call_del_module(struct module *mod)
393 {
394         struct static_call_site *start = mod->static_call_sites;
395         struct static_call_site *stop = mod->static_call_sites +
396                                         mod->num_static_call_sites;
397         struct static_call_key *key, *prev_key = NULL;
398         struct static_call_mod *site_mod, **prev;
399         struct static_call_site *site;
400
401         for (site = start; site < stop; site++) {
402                 key = static_call_key(site);
403                 if (key == prev_key)
404                         continue;
405
406                 prev_key = key;
407
408                 for (prev = &key->mods, site_mod = key->mods;
409                      site_mod && site_mod->mod != mod;
410                      prev = &site_mod->next, site_mod = site_mod->next)
411                         ;
412
413                 if (!site_mod)
414                         continue;
415
416                 *prev = site_mod->next;
417                 kfree(site_mod);
418         }
419 }
420
421 static int static_call_module_notify(struct notifier_block *nb,
422                                      unsigned long val, void *data)
423 {
424         struct module *mod = data;
425         int ret = 0;
426
427         cpus_read_lock();
428         static_call_lock();
429
430         switch (val) {
431         case MODULE_STATE_COMING:
432                 ret = static_call_add_module(mod);
433                 if (ret) {
434                         WARN(1, "Failed to allocate memory for static calls");
435                         static_call_del_module(mod);
436                 }
437                 break;
438         case MODULE_STATE_GOING:
439                 static_call_del_module(mod);
440                 break;
441         }
442
443         static_call_unlock();
444         cpus_read_unlock();
445
446         return notifier_from_errno(ret);
447 }
448
449 static struct notifier_block static_call_module_nb = {
450         .notifier_call = static_call_module_notify,
451 };
452
453 #else
454
455 static inline int __static_call_mod_text_reserved(void *start, void *end)
456 {
457         return 0;
458 }
459
460 #endif /* CONFIG_MODULES */
461
462 int static_call_text_reserved(void *start, void *end)
463 {
464         bool init = system_state < SYSTEM_RUNNING;
465         int ret = __static_call_text_reserved(__start_static_call_sites,
466                         __stop_static_call_sites, start, end, init);
467
468         if (ret)
469                 return ret;
470
471         return __static_call_mod_text_reserved(start, end);
472 }
473
474 int __init static_call_init(void)
475 {
476         int ret;
477
478         if (static_call_initialized)
479                 return 0;
480
481         cpus_read_lock();
482         static_call_lock();
483         ret = __static_call_init(NULL, __start_static_call_sites,
484                                  __stop_static_call_sites);
485         static_call_unlock();
486         cpus_read_unlock();
487
488         if (ret) {
489                 pr_err("Failed to allocate memory for static_call!\n");
490                 BUG();
491         }
492
493         static_call_initialized = true;
494
495 #ifdef CONFIG_MODULES
496         register_module_notifier(&static_call_module_nb);
497 #endif
498         return 0;
499 }
500 early_initcall(static_call_init);
501
502 #ifdef CONFIG_STATIC_CALL_SELFTEST
503
504 static int func_a(int x)
505 {
506         return x+1;
507 }
508
509 static int func_b(int x)
510 {
511         return x+2;
512 }
513
514 DEFINE_STATIC_CALL(sc_selftest, func_a);
515
516 static struct static_call_data {
517       int (*func)(int);
518       int val;
519       int expect;
520 } static_call_data [] __initdata = {
521       { NULL,   2, 3 },
522       { func_b, 2, 4 },
523       { func_a, 2, 3 }
524 };
525
526 static int __init test_static_call_init(void)
527 {
528       int i;
529
530       for (i = 0; i < ARRAY_SIZE(static_call_data); i++ ) {
531               struct static_call_data *scd = &static_call_data[i];
532
533               if (scd->func)
534                       static_call_update(sc_selftest, scd->func);
535
536               WARN_ON(static_call(sc_selftest)(scd->val) != scd->expect);
537       }
538
539       return 0;
540 }
541 early_initcall(test_static_call_init);
542
543 #endif /* CONFIG_STATIC_CALL_SELFTEST */