clk: baikal-t1: Convert to platform device driver
[platform/kernel/linux-starfive.git] / tools / lib / perf / cpumap.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 #include <perf/cpumap.h>
3 #include <stdlib.h>
4 #include <linux/refcount.h>
5 #include <internal/cpumap.h>
6 #include <asm/bug.h>
7 #include <stdio.h>
8 #include <string.h>
9 #include <unistd.h>
10 #include <ctype.h>
11 #include <limits.h>
12
13 static struct perf_cpu_map *perf_cpu_map__alloc(int nr_cpus)
14 {
15         struct perf_cpu_map *cpus = malloc(sizeof(*cpus) + sizeof(struct perf_cpu) * nr_cpus);
16
17         if (cpus != NULL) {
18                 cpus->nr = nr_cpus;
19                 refcount_set(&cpus->refcnt, 1);
20
21         }
22         return cpus;
23 }
24
25 struct perf_cpu_map *perf_cpu_map__dummy_new(void)
26 {
27         struct perf_cpu_map *cpus = perf_cpu_map__alloc(1);
28
29         if (cpus)
30                 cpus->map[0].cpu = -1;
31
32         return cpus;
33 }
34
35 static void cpu_map__delete(struct perf_cpu_map *map)
36 {
37         if (map) {
38                 WARN_ONCE(refcount_read(&map->refcnt) != 0,
39                           "cpu_map refcnt unbalanced\n");
40                 free(map);
41         }
42 }
43
44 struct perf_cpu_map *perf_cpu_map__get(struct perf_cpu_map *map)
45 {
46         if (map)
47                 refcount_inc(&map->refcnt);
48         return map;
49 }
50
51 void perf_cpu_map__put(struct perf_cpu_map *map)
52 {
53         if (map && refcount_dec_and_test(&map->refcnt))
54                 cpu_map__delete(map);
55 }
56
57 static struct perf_cpu_map *cpu_map__default_new(void)
58 {
59         struct perf_cpu_map *cpus;
60         int nr_cpus;
61
62         nr_cpus = sysconf(_SC_NPROCESSORS_ONLN);
63         if (nr_cpus < 0)
64                 return NULL;
65
66         cpus = perf_cpu_map__alloc(nr_cpus);
67         if (cpus != NULL) {
68                 int i;
69
70                 for (i = 0; i < nr_cpus; ++i)
71                         cpus->map[i].cpu = i;
72         }
73
74         return cpus;
75 }
76
77 struct perf_cpu_map *perf_cpu_map__default_new(void)
78 {
79         return cpu_map__default_new();
80 }
81
82
83 static int cmp_cpu(const void *a, const void *b)
84 {
85         const struct perf_cpu *cpu_a = a, *cpu_b = b;
86
87         return cpu_a->cpu - cpu_b->cpu;
88 }
89
90 static struct perf_cpu_map *cpu_map__trim_new(int nr_cpus, const struct perf_cpu *tmp_cpus)
91 {
92         size_t payload_size = nr_cpus * sizeof(struct perf_cpu);
93         struct perf_cpu_map *cpus = perf_cpu_map__alloc(nr_cpus);
94         int i, j;
95
96         if (cpus != NULL) {
97                 memcpy(cpus->map, tmp_cpus, payload_size);
98                 qsort(cpus->map, nr_cpus, sizeof(struct perf_cpu), cmp_cpu);
99                 /* Remove dups */
100                 j = 0;
101                 for (i = 0; i < nr_cpus; i++) {
102                         if (i == 0 || cpus->map[i].cpu != cpus->map[i - 1].cpu)
103                                 cpus->map[j++].cpu = cpus->map[i].cpu;
104                 }
105                 cpus->nr = j;
106                 assert(j <= nr_cpus);
107         }
108         return cpus;
109 }
110
111 struct perf_cpu_map *perf_cpu_map__read(FILE *file)
112 {
113         struct perf_cpu_map *cpus = NULL;
114         int nr_cpus = 0;
115         struct perf_cpu *tmp_cpus = NULL, *tmp;
116         int max_entries = 0;
117         int n, cpu, prev;
118         char sep;
119
120         sep = 0;
121         prev = -1;
122         for (;;) {
123                 n = fscanf(file, "%u%c", &cpu, &sep);
124                 if (n <= 0)
125                         break;
126                 if (prev >= 0) {
127                         int new_max = nr_cpus + cpu - prev - 1;
128
129                         WARN_ONCE(new_max >= MAX_NR_CPUS, "Perf can support %d CPUs. "
130                                                           "Consider raising MAX_NR_CPUS\n", MAX_NR_CPUS);
131
132                         if (new_max >= max_entries) {
133                                 max_entries = new_max + MAX_NR_CPUS / 2;
134                                 tmp = realloc(tmp_cpus, max_entries * sizeof(struct perf_cpu));
135                                 if (tmp == NULL)
136                                         goto out_free_tmp;
137                                 tmp_cpus = tmp;
138                         }
139
140                         while (++prev < cpu)
141                                 tmp_cpus[nr_cpus++].cpu = prev;
142                 }
143                 if (nr_cpus == max_entries) {
144                         max_entries += MAX_NR_CPUS;
145                         tmp = realloc(tmp_cpus, max_entries * sizeof(struct perf_cpu));
146                         if (tmp == NULL)
147                                 goto out_free_tmp;
148                         tmp_cpus = tmp;
149                 }
150
151                 tmp_cpus[nr_cpus++].cpu = cpu;
152                 if (n == 2 && sep == '-')
153                         prev = cpu;
154                 else
155                         prev = -1;
156                 if (n == 1 || sep == '\n')
157                         break;
158         }
159
160         if (nr_cpus > 0)
161                 cpus = cpu_map__trim_new(nr_cpus, tmp_cpus);
162         else
163                 cpus = cpu_map__default_new();
164 out_free_tmp:
165         free(tmp_cpus);
166         return cpus;
167 }
168
169 static struct perf_cpu_map *cpu_map__read_all_cpu_map(void)
170 {
171         struct perf_cpu_map *cpus = NULL;
172         FILE *onlnf;
173
174         onlnf = fopen("/sys/devices/system/cpu/online", "r");
175         if (!onlnf)
176                 return cpu_map__default_new();
177
178         cpus = perf_cpu_map__read(onlnf);
179         fclose(onlnf);
180         return cpus;
181 }
182
183 struct perf_cpu_map *perf_cpu_map__new(const char *cpu_list)
184 {
185         struct perf_cpu_map *cpus = NULL;
186         unsigned long start_cpu, end_cpu = 0;
187         char *p = NULL;
188         int i, nr_cpus = 0;
189         struct perf_cpu *tmp_cpus = NULL, *tmp;
190         int max_entries = 0;
191
192         if (!cpu_list)
193                 return cpu_map__read_all_cpu_map();
194
195         /*
196          * must handle the case of empty cpumap to cover
197          * TOPOLOGY header for NUMA nodes with no CPU
198          * ( e.g., because of CPU hotplug)
199          */
200         if (!isdigit(*cpu_list) && *cpu_list != '\0')
201                 goto out;
202
203         while (isdigit(*cpu_list)) {
204                 p = NULL;
205                 start_cpu = strtoul(cpu_list, &p, 0);
206                 if (start_cpu >= INT_MAX
207                     || (*p != '\0' && *p != ',' && *p != '-'))
208                         goto invalid;
209
210                 if (*p == '-') {
211                         cpu_list = ++p;
212                         p = NULL;
213                         end_cpu = strtoul(cpu_list, &p, 0);
214
215                         if (end_cpu >= INT_MAX || (*p != '\0' && *p != ','))
216                                 goto invalid;
217
218                         if (end_cpu < start_cpu)
219                                 goto invalid;
220                 } else {
221                         end_cpu = start_cpu;
222                 }
223
224                 WARN_ONCE(end_cpu >= MAX_NR_CPUS, "Perf can support %d CPUs. "
225                                                   "Consider raising MAX_NR_CPUS\n", MAX_NR_CPUS);
226
227                 for (; start_cpu <= end_cpu; start_cpu++) {
228                         /* check for duplicates */
229                         for (i = 0; i < nr_cpus; i++)
230                                 if (tmp_cpus[i].cpu == (int)start_cpu)
231                                         goto invalid;
232
233                         if (nr_cpus == max_entries) {
234                                 max_entries += MAX_NR_CPUS;
235                                 tmp = realloc(tmp_cpus, max_entries * sizeof(struct perf_cpu));
236                                 if (tmp == NULL)
237                                         goto invalid;
238                                 tmp_cpus = tmp;
239                         }
240                         tmp_cpus[nr_cpus++].cpu = (int)start_cpu;
241                 }
242                 if (*p)
243                         ++p;
244
245                 cpu_list = p;
246         }
247
248         if (nr_cpus > 0)
249                 cpus = cpu_map__trim_new(nr_cpus, tmp_cpus);
250         else if (*cpu_list != '\0')
251                 cpus = cpu_map__default_new();
252         else
253                 cpus = perf_cpu_map__dummy_new();
254 invalid:
255         free(tmp_cpus);
256 out:
257         return cpus;
258 }
259
260 struct perf_cpu perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx)
261 {
262         struct perf_cpu result = {
263                 .cpu = -1
264         };
265
266         if (cpus && idx < cpus->nr)
267                 return cpus->map[idx];
268
269         return result;
270 }
271
272 int perf_cpu_map__nr(const struct perf_cpu_map *cpus)
273 {
274         return cpus ? cpus->nr : 1;
275 }
276
277 bool perf_cpu_map__empty(const struct perf_cpu_map *map)
278 {
279         return map ? map->map[0].cpu == -1 : true;
280 }
281
282 int perf_cpu_map__idx(const struct perf_cpu_map *cpus, struct perf_cpu cpu)
283 {
284         int low, high;
285
286         if (!cpus)
287                 return -1;
288
289         low = 0;
290         high = cpus->nr;
291         while (low < high) {
292                 int idx = (low + high) / 2;
293                 struct perf_cpu cpu_at_idx = cpus->map[idx];
294
295                 if (cpu_at_idx.cpu == cpu.cpu)
296                         return idx;
297
298                 if (cpu_at_idx.cpu > cpu.cpu)
299                         high = idx;
300                 else
301                         low = idx + 1;
302         }
303
304         return -1;
305 }
306
307 bool perf_cpu_map__has(const struct perf_cpu_map *cpus, struct perf_cpu cpu)
308 {
309         return perf_cpu_map__idx(cpus, cpu) != -1;
310 }
311
312 struct perf_cpu perf_cpu_map__max(struct perf_cpu_map *map)
313 {
314         struct perf_cpu result = {
315                 .cpu = -1
316         };
317
318         // cpu_map__trim_new() qsort()s it, cpu_map__default_new() sorts it as well.
319         return map->nr > 0 ? map->map[map->nr - 1] : result;
320 }
321
322 /** Is 'b' a subset of 'a'. */
323 bool perf_cpu_map__is_subset(const struct perf_cpu_map *a, const struct perf_cpu_map *b)
324 {
325         if (a == b || !b)
326                 return true;
327         if (!a || b->nr > a->nr)
328                 return false;
329
330         for (int i = 0, j = 0; i < a->nr; i++) {
331                 if (a->map[i].cpu > b->map[j].cpu)
332                         return false;
333                 if (a->map[i].cpu == b->map[j].cpu) {
334                         j++;
335                         if (j == b->nr)
336                                 return true;
337                 }
338         }
339         return false;
340 }
341
342 /*
343  * Merge two cpumaps
344  *
345  * orig either gets freed and replaced with a new map, or reused
346  * with no reference count change (similar to "realloc")
347  * other has its reference count increased.
348  */
349
350 struct perf_cpu_map *perf_cpu_map__merge(struct perf_cpu_map *orig,
351                                          struct perf_cpu_map *other)
352 {
353         struct perf_cpu *tmp_cpus;
354         int tmp_len;
355         int i, j, k;
356         struct perf_cpu_map *merged;
357
358         if (perf_cpu_map__is_subset(orig, other))
359                 return orig;
360         if (perf_cpu_map__is_subset(other, orig)) {
361                 perf_cpu_map__put(orig);
362                 return perf_cpu_map__get(other);
363         }
364
365         tmp_len = orig->nr + other->nr;
366         tmp_cpus = malloc(tmp_len * sizeof(struct perf_cpu));
367         if (!tmp_cpus)
368                 return NULL;
369
370         /* Standard merge algorithm from wikipedia */
371         i = j = k = 0;
372         while (i < orig->nr && j < other->nr) {
373                 if (orig->map[i].cpu <= other->map[j].cpu) {
374                         if (orig->map[i].cpu == other->map[j].cpu)
375                                 j++;
376                         tmp_cpus[k++] = orig->map[i++];
377                 } else
378                         tmp_cpus[k++] = other->map[j++];
379         }
380
381         while (i < orig->nr)
382                 tmp_cpus[k++] = orig->map[i++];
383
384         while (j < other->nr)
385                 tmp_cpus[k++] = other->map[j++];
386         assert(k <= tmp_len);
387
388         merged = cpu_map__trim_new(k, tmp_cpus);
389         free(tmp_cpus);
390         perf_cpu_map__put(orig);
391         return merged;
392 }