}
static struct perf_mmap*
-perf_evlist__map_get(struct perf_evlist *evlist, bool overwrite, int idx)
+perf_evlist__mmap_cb_get(struct perf_evlist *evlist, bool overwrite, int idx)
{
struct perf_mmap *map = &evlist->mmap[idx];
#define FD(e, x, y) (*(int *) xyarray__entry(e->fd, x, y))
static int
-mmap_per_evsel(struct perf_evlist *evlist, int idx,
- struct perf_mmap_param *mp, int cpu_idx,
+mmap_per_evsel(struct perf_evlist *evlist, struct perf_evlist_mmap_ops *ops,
+ int idx, struct perf_mmap_param *mp, int cpu_idx,
int thread, int *_output, int *_output_overwrite)
{
int evlist_cpu = perf_cpu_map__cpu(evlist->cpus, cpu_idx);
if (cpu == -1)
continue;
- map = perf_evlist__map_get(evlist, overwrite, idx);
+ map = ops->get(evlist, overwrite, idx);
if (map == NULL)
return -ENOMEM;
if (ops->idx)
ops->idx(evlist, mp, thread, false);
- if (mmap_per_evsel(evlist, thread, mp, 0, thread,
+ if (mmap_per_evsel(evlist, ops, thread, mp, 0, thread,
&output, &output_overwrite))
goto out_unmap;
}
ops->idx(evlist, mp, cpu, true);
for (thread = 0; thread < nr_threads; thread++) {
- if (mmap_per_evsel(evlist, cpu, mp, cpu,
+ if (mmap_per_evsel(evlist, ops, cpu, mp, cpu,
thread, &output, &output_overwrite))
goto out_unmap;
}
const struct perf_cpu_map *cpus = evlist->cpus;
const struct perf_thread_map *threads = evlist->threads;
- if (!ops)
+ if (!ops || !ops->get)
return -EINVAL;
if (!evlist->mmap)
int perf_evlist__mmap(struct perf_evlist *evlist, int pages)
{
struct perf_mmap_param mp;
- struct perf_evlist_mmap_ops ops = { 0 };
+ struct perf_evlist_mmap_ops ops = {
+ .get = perf_evlist__mmap_cb_get,
+ };
evlist->mmap_len = (pages + 1) * page_size;
mp.mask = evlist->mmap_len - page_size - 1;