Loading tools/perf/builtin-record.c +1 −1 Original line number Diff line number Diff line Loading @@ -427,7 +427,7 @@ static void mmap_read_all(void) { int i; for (i = 0; i < evsel_list->cpus->nr; i++) { for (i = 0; i < evsel_list->nr_mmaps; i++) { if (evsel_list->mmap[i].base) mmap_read(&evsel_list->mmap[i]); } Loading tools/perf/builtin-test.c +1 −1 Original line number Diff line number Diff line Loading @@ -549,7 +549,7 @@ static int test__basic_mmap(void) ++foo; } while ((event = perf_evlist__read_on_cpu(evlist, 0)) != NULL) { while ((event = perf_evlist__mmap_read(evlist, 0)) != NULL) { struct perf_sample sample; if (event->header.type != PERF_RECORD_SAMPLE) { Loading tools/perf/builtin-top.c +4 −4 Original line number Diff line number Diff line Loading @@ -801,12 +801,12 @@ static void perf_event__process_sample(const union perf_event *event, } } static void perf_session__mmap_read_cpu(struct perf_session *self, int cpu) static void perf_session__mmap_read_idx(struct perf_session *self, int idx) { struct perf_sample sample; union perf_event *event; while ((event = perf_evlist__read_on_cpu(top.evlist, cpu)) != NULL) { while ((event = perf_evlist__mmap_read(top.evlist, idx)) != NULL) { perf_session__parse_sample(self, event, &sample); if (event->header.type == PERF_RECORD_SAMPLE) Loading @@ -820,8 +820,8 @@ static void perf_session__mmap_read(struct perf_session *self) { int i; for (i = 0; i < top.evlist->cpus->nr; i++) perf_session__mmap_read_cpu(self, i); for (i = 0; i < top.evlist->nr_mmaps; i++) perf_session__mmap_read_idx(self, i); } static void start_counters(struct perf_evlist *evlist) Loading tools/perf/util/evlist.c +107 −46 Original line number Diff line number Diff line Loading @@ -166,11 +166,11 @@ struct perf_evsel *perf_evlist__id2evsel(struct perf_evlist *evlist, u64 id) return NULL; } union perf_event *perf_evlist__read_on_cpu(struct perf_evlist *evlist, int cpu) union perf_event *perf_evlist__mmap_read(struct perf_evlist *evlist, int idx) { /* XXX Move this to perf.c, making it generally available */ unsigned int page_size = sysconf(_SC_PAGE_SIZE); struct perf_mmap *md = &evlist->mmap[cpu]; struct perf_mmap *md = &evlist->mmap[idx]; unsigned int head = perf_mmap__read_head(md); unsigned int old = md->prev; unsigned char *data = md->base + page_size; Loading Loading @@ -235,31 +235,37 @@ union perf_event *perf_evlist__read_on_cpu(struct perf_evlist *evlist, int cpu) void perf_evlist__munmap(struct perf_evlist *evlist) { int cpu; int i; for (cpu = 0; cpu < evlist->cpus->nr; cpu++) { if (evlist->mmap[cpu].base != NULL) { munmap(evlist->mmap[cpu].base, evlist->mmap_len); evlist->mmap[cpu].base = NULL; for (i = 0; i < evlist->nr_mmaps; i++) { if (evlist->mmap[i].base != NULL) { munmap(evlist->mmap[i].base, evlist->mmap_len); evlist->mmap[i].base = NULL; } } free(evlist->mmap); evlist->mmap = NULL; } int perf_evlist__alloc_mmap(struct perf_evlist *evlist) { evlist->mmap = zalloc(evlist->cpus->nr * sizeof(struct perf_mmap)); evlist->nr_mmaps = evlist->cpus->nr; if (evlist->cpus->map[0] == -1) evlist->nr_mmaps = evlist->threads->nr; evlist->mmap = zalloc(evlist->nr_mmaps * sizeof(struct perf_mmap)); return evlist->mmap != NULL ? 0 : -ENOMEM; } static int __perf_evlist__mmap(struct perf_evlist *evlist, struct perf_evsel *evsel, int cpu, int prot, int mask, int fd) int idx, int prot, int mask, int fd) { evlist->mmap[cpu].prev = 0; evlist->mmap[cpu].mask = mask; evlist->mmap[cpu].base = mmap(NULL, evlist->mmap_len, prot, evlist->mmap[idx].prev = 0; evlist->mmap[idx].mask = mask; evlist->mmap[idx].base = mmap(NULL, evlist->mmap_len, prot, MAP_SHARED, fd, 0); if (evlist->mmap[cpu].base == MAP_FAILED) { if (evlist->cpus->map[cpu] == -1 && evsel->attr.inherit) if (evlist->mmap[idx].base == MAP_FAILED) { if (evlist->cpus->map[idx] == -1 && evsel->attr.inherit) ui__warning("Inherit is not allowed on per-task " "events using mmap.\n"); return -1; Loading @@ -269,6 +275,86 @@ static int __perf_evlist__mmap(struct perf_evlist *evlist, struct perf_evsel *ev return 0; } static int perf_evlist__mmap_per_cpu(struct perf_evlist *evlist, int prot, int mask) { struct perf_evsel *evsel; int cpu, thread; for (cpu = 0; cpu < evlist->cpus->nr; cpu++) { int output = -1; for (thread = 0; thread < evlist->threads->nr; thread++) { list_for_each_entry(evsel, &evlist->entries, node) { int fd = FD(evsel, cpu, thread); if (output == -1) { output = fd; if (__perf_evlist__mmap(evlist, evsel, cpu, prot, mask, output) < 0) goto out_unmap; } else { if (ioctl(fd, PERF_EVENT_IOC_SET_OUTPUT, output) != 0) goto out_unmap; } if ((evsel->attr.read_format & PERF_FORMAT_ID) && perf_evlist__id_add_fd(evlist, evsel, cpu, thread, fd) < 0) goto out_unmap; } } } return 0; out_unmap: for (cpu = 0; cpu < evlist->cpus->nr; cpu++) { if (evlist->mmap[cpu].base != NULL) { munmap(evlist->mmap[cpu].base, evlist->mmap_len); evlist->mmap[cpu].base = NULL; } } return -1; } static int perf_evlist__mmap_per_thread(struct perf_evlist *evlist, int prot, int mask) { struct perf_evsel *evsel; int thread; for (thread = 0; thread < evlist->threads->nr; thread++) { int output = -1; list_for_each_entry(evsel, &evlist->entries, node) { int fd = FD(evsel, 0, thread); if (output == -1) { output = fd; if (__perf_evlist__mmap(evlist, evsel, thread, prot, mask, output) < 0) goto out_unmap; } else { if (ioctl(fd, PERF_EVENT_IOC_SET_OUTPUT, output) != 0) goto out_unmap; } if ((evsel->attr.read_format & PERF_FORMAT_ID) && perf_evlist__id_add_fd(evlist, evsel, 0, thread, fd) < 0) goto out_unmap; } } return 0; out_unmap: for (thread = 0; thread < evlist->threads->nr; thread++) { if (evlist->mmap[thread].base != NULL) { munmap(evlist->mmap[thread].base, evlist->mmap_len); evlist->mmap[thread].base = NULL; } } return -1; } /** perf_evlist__mmap - Create per cpu maps to receive events * * @evlist - list of events Loading @@ -287,11 +373,11 @@ static int __perf_evlist__mmap(struct perf_evlist *evlist, struct perf_evsel *ev int perf_evlist__mmap(struct perf_evlist *evlist, int pages, bool overwrite) { unsigned int page_size = sysconf(_SC_PAGE_SIZE); int mask = pages * page_size - 1, cpu; struct perf_evsel *first_evsel, *evsel; int mask = pages * page_size - 1; struct perf_evsel *evsel; const struct cpu_map *cpus = evlist->cpus; const struct thread_map *threads = evlist->threads; int thread, prot = PROT_READ | (overwrite ? 0 : PROT_WRITE); int prot = PROT_READ | (overwrite ? 0 : PROT_WRITE); if (evlist->mmap == NULL && perf_evlist__alloc_mmap(evlist) < 0) return -ENOMEM; Loading @@ -301,43 +387,18 @@ int perf_evlist__mmap(struct perf_evlist *evlist, int pages, bool overwrite) evlist->overwrite = overwrite; evlist->mmap_len = (pages + 1) * page_size; first_evsel = list_entry(evlist->entries.next, struct perf_evsel, node); list_for_each_entry(evsel, &evlist->entries, node) { if ((evsel->attr.read_format & PERF_FORMAT_ID) && evsel->sample_id == NULL && perf_evsel__alloc_id(evsel, cpus->nr, threads->nr) < 0) return -ENOMEM; for (cpu = 0; cpu < cpus->nr; cpu++) { for (thread = 0; thread < threads->nr; thread++) { int fd = FD(evsel, cpu, thread); if (evsel->idx || thread) { if (ioctl(fd, PERF_EVENT_IOC_SET_OUTPUT, FD(first_evsel, cpu, 0)) != 0) goto out_unmap; } else if (__perf_evlist__mmap(evlist, evsel, cpu, prot, mask, fd) < 0) goto out_unmap; if ((evsel->attr.read_format & PERF_FORMAT_ID) && perf_evlist__id_add_fd(evlist, evsel, cpu, thread, fd) < 0) goto out_unmap; } } } return 0; if (evlist->cpus->map[0] == -1) return perf_evlist__mmap_per_thread(evlist, prot, mask); out_unmap: for (cpu = 0; cpu < cpus->nr; cpu++) { if (evlist->mmap[cpu].base != NULL) { munmap(evlist->mmap[cpu].base, evlist->mmap_len); evlist->mmap[cpu].base = NULL; } } return -1; return perf_evlist__mmap_per_cpu(evlist, prot, mask); } int perf_evlist__create_maps(struct perf_evlist *evlist, pid_t target_pid, Loading @@ -348,7 +409,7 @@ int perf_evlist__create_maps(struct perf_evlist *evlist, pid_t target_pid, if (evlist->threads == NULL) return -1; if (target_tid != -1) if (cpu_list == NULL && target_tid != -1) evlist->cpus = cpu_map__dummy_new(); else evlist->cpus = cpu_map__new(cpu_list); Loading tools/perf/util/evlist.h +2 −1 Original line number Diff line number Diff line Loading @@ -17,6 +17,7 @@ struct perf_evlist { struct hlist_head heads[PERF_EVLIST__HLIST_SIZE]; int nr_entries; int nr_fds; int nr_mmaps; int mmap_len; bool overwrite; union perf_event event_copy; Loading Loading @@ -46,7 +47,7 @@ void perf_evlist__add_pollfd(struct perf_evlist *evlist, int fd); struct perf_evsel *perf_evlist__id2evsel(struct perf_evlist *evlist, u64 id); union perf_event *perf_evlist__read_on_cpu(struct perf_evlist *self, int cpu); union perf_event *perf_evlist__mmap_read(struct perf_evlist *self, int idx); int perf_evlist__alloc_mmap(struct perf_evlist *evlist); int perf_evlist__mmap(struct perf_evlist *evlist, int pages, bool overwrite); Loading Loading
tools/perf/builtin-record.c +1 −1 Original line number Diff line number Diff line Loading @@ -427,7 +427,7 @@ static void mmap_read_all(void) { int i; for (i = 0; i < evsel_list->cpus->nr; i++) { for (i = 0; i < evsel_list->nr_mmaps; i++) { if (evsel_list->mmap[i].base) mmap_read(&evsel_list->mmap[i]); } Loading
tools/perf/builtin-test.c +1 −1 Original line number Diff line number Diff line Loading @@ -549,7 +549,7 @@ static int test__basic_mmap(void) ++foo; } while ((event = perf_evlist__read_on_cpu(evlist, 0)) != NULL) { while ((event = perf_evlist__mmap_read(evlist, 0)) != NULL) { struct perf_sample sample; if (event->header.type != PERF_RECORD_SAMPLE) { Loading
tools/perf/builtin-top.c +4 −4 Original line number Diff line number Diff line Loading @@ -801,12 +801,12 @@ static void perf_event__process_sample(const union perf_event *event, } } static void perf_session__mmap_read_cpu(struct perf_session *self, int cpu) static void perf_session__mmap_read_idx(struct perf_session *self, int idx) { struct perf_sample sample; union perf_event *event; while ((event = perf_evlist__read_on_cpu(top.evlist, cpu)) != NULL) { while ((event = perf_evlist__mmap_read(top.evlist, idx)) != NULL) { perf_session__parse_sample(self, event, &sample); if (event->header.type == PERF_RECORD_SAMPLE) Loading @@ -820,8 +820,8 @@ static void perf_session__mmap_read(struct perf_session *self) { int i; for (i = 0; i < top.evlist->cpus->nr; i++) perf_session__mmap_read_cpu(self, i); for (i = 0; i < top.evlist->nr_mmaps; i++) perf_session__mmap_read_idx(self, i); } static void start_counters(struct perf_evlist *evlist) Loading
tools/perf/util/evlist.c +107 −46 Original line number Diff line number Diff line Loading @@ -166,11 +166,11 @@ struct perf_evsel *perf_evlist__id2evsel(struct perf_evlist *evlist, u64 id) return NULL; } union perf_event *perf_evlist__read_on_cpu(struct perf_evlist *evlist, int cpu) union perf_event *perf_evlist__mmap_read(struct perf_evlist *evlist, int idx) { /* XXX Move this to perf.c, making it generally available */ unsigned int page_size = sysconf(_SC_PAGE_SIZE); struct perf_mmap *md = &evlist->mmap[cpu]; struct perf_mmap *md = &evlist->mmap[idx]; unsigned int head = perf_mmap__read_head(md); unsigned int old = md->prev; unsigned char *data = md->base + page_size; Loading Loading @@ -235,31 +235,37 @@ union perf_event *perf_evlist__read_on_cpu(struct perf_evlist *evlist, int cpu) void perf_evlist__munmap(struct perf_evlist *evlist) { int cpu; int i; for (cpu = 0; cpu < evlist->cpus->nr; cpu++) { if (evlist->mmap[cpu].base != NULL) { munmap(evlist->mmap[cpu].base, evlist->mmap_len); evlist->mmap[cpu].base = NULL; for (i = 0; i < evlist->nr_mmaps; i++) { if (evlist->mmap[i].base != NULL) { munmap(evlist->mmap[i].base, evlist->mmap_len); evlist->mmap[i].base = NULL; } } free(evlist->mmap); evlist->mmap = NULL; } int perf_evlist__alloc_mmap(struct perf_evlist *evlist) { evlist->mmap = zalloc(evlist->cpus->nr * sizeof(struct perf_mmap)); evlist->nr_mmaps = evlist->cpus->nr; if (evlist->cpus->map[0] == -1) evlist->nr_mmaps = evlist->threads->nr; evlist->mmap = zalloc(evlist->nr_mmaps * sizeof(struct perf_mmap)); return evlist->mmap != NULL ? 0 : -ENOMEM; } static int __perf_evlist__mmap(struct perf_evlist *evlist, struct perf_evsel *evsel, int cpu, int prot, int mask, int fd) int idx, int prot, int mask, int fd) { evlist->mmap[cpu].prev = 0; evlist->mmap[cpu].mask = mask; evlist->mmap[cpu].base = mmap(NULL, evlist->mmap_len, prot, evlist->mmap[idx].prev = 0; evlist->mmap[idx].mask = mask; evlist->mmap[idx].base = mmap(NULL, evlist->mmap_len, prot, MAP_SHARED, fd, 0); if (evlist->mmap[cpu].base == MAP_FAILED) { if (evlist->cpus->map[cpu] == -1 && evsel->attr.inherit) if (evlist->mmap[idx].base == MAP_FAILED) { if (evlist->cpus->map[idx] == -1 && evsel->attr.inherit) ui__warning("Inherit is not allowed on per-task " "events using mmap.\n"); return -1; Loading @@ -269,6 +275,86 @@ static int __perf_evlist__mmap(struct perf_evlist *evlist, struct perf_evsel *ev return 0; } static int perf_evlist__mmap_per_cpu(struct perf_evlist *evlist, int prot, int mask) { struct perf_evsel *evsel; int cpu, thread; for (cpu = 0; cpu < evlist->cpus->nr; cpu++) { int output = -1; for (thread = 0; thread < evlist->threads->nr; thread++) { list_for_each_entry(evsel, &evlist->entries, node) { int fd = FD(evsel, cpu, thread); if (output == -1) { output = fd; if (__perf_evlist__mmap(evlist, evsel, cpu, prot, mask, output) < 0) goto out_unmap; } else { if (ioctl(fd, PERF_EVENT_IOC_SET_OUTPUT, output) != 0) goto out_unmap; } if ((evsel->attr.read_format & PERF_FORMAT_ID) && perf_evlist__id_add_fd(evlist, evsel, cpu, thread, fd) < 0) goto out_unmap; } } } return 0; out_unmap: for (cpu = 0; cpu < evlist->cpus->nr; cpu++) { if (evlist->mmap[cpu].base != NULL) { munmap(evlist->mmap[cpu].base, evlist->mmap_len); evlist->mmap[cpu].base = NULL; } } return -1; } static int perf_evlist__mmap_per_thread(struct perf_evlist *evlist, int prot, int mask) { struct perf_evsel *evsel; int thread; for (thread = 0; thread < evlist->threads->nr; thread++) { int output = -1; list_for_each_entry(evsel, &evlist->entries, node) { int fd = FD(evsel, 0, thread); if (output == -1) { output = fd; if (__perf_evlist__mmap(evlist, evsel, thread, prot, mask, output) < 0) goto out_unmap; } else { if (ioctl(fd, PERF_EVENT_IOC_SET_OUTPUT, output) != 0) goto out_unmap; } if ((evsel->attr.read_format & PERF_FORMAT_ID) && perf_evlist__id_add_fd(evlist, evsel, 0, thread, fd) < 0) goto out_unmap; } } return 0; out_unmap: for (thread = 0; thread < evlist->threads->nr; thread++) { if (evlist->mmap[thread].base != NULL) { munmap(evlist->mmap[thread].base, evlist->mmap_len); evlist->mmap[thread].base = NULL; } } return -1; } /** perf_evlist__mmap - Create per cpu maps to receive events * * @evlist - list of events Loading @@ -287,11 +373,11 @@ static int __perf_evlist__mmap(struct perf_evlist *evlist, struct perf_evsel *ev int perf_evlist__mmap(struct perf_evlist *evlist, int pages, bool overwrite) { unsigned int page_size = sysconf(_SC_PAGE_SIZE); int mask = pages * page_size - 1, cpu; struct perf_evsel *first_evsel, *evsel; int mask = pages * page_size - 1; struct perf_evsel *evsel; const struct cpu_map *cpus = evlist->cpus; const struct thread_map *threads = evlist->threads; int thread, prot = PROT_READ | (overwrite ? 0 : PROT_WRITE); int prot = PROT_READ | (overwrite ? 0 : PROT_WRITE); if (evlist->mmap == NULL && perf_evlist__alloc_mmap(evlist) < 0) return -ENOMEM; Loading @@ -301,43 +387,18 @@ int perf_evlist__mmap(struct perf_evlist *evlist, int pages, bool overwrite) evlist->overwrite = overwrite; evlist->mmap_len = (pages + 1) * page_size; first_evsel = list_entry(evlist->entries.next, struct perf_evsel, node); list_for_each_entry(evsel, &evlist->entries, node) { if ((evsel->attr.read_format & PERF_FORMAT_ID) && evsel->sample_id == NULL && perf_evsel__alloc_id(evsel, cpus->nr, threads->nr) < 0) return -ENOMEM; for (cpu = 0; cpu < cpus->nr; cpu++) { for (thread = 0; thread < threads->nr; thread++) { int fd = FD(evsel, cpu, thread); if (evsel->idx || thread) { if (ioctl(fd, PERF_EVENT_IOC_SET_OUTPUT, FD(first_evsel, cpu, 0)) != 0) goto out_unmap; } else if (__perf_evlist__mmap(evlist, evsel, cpu, prot, mask, fd) < 0) goto out_unmap; if ((evsel->attr.read_format & PERF_FORMAT_ID) && perf_evlist__id_add_fd(evlist, evsel, cpu, thread, fd) < 0) goto out_unmap; } } } return 0; if (evlist->cpus->map[0] == -1) return perf_evlist__mmap_per_thread(evlist, prot, mask); out_unmap: for (cpu = 0; cpu < cpus->nr; cpu++) { if (evlist->mmap[cpu].base != NULL) { munmap(evlist->mmap[cpu].base, evlist->mmap_len); evlist->mmap[cpu].base = NULL; } } return -1; return perf_evlist__mmap_per_cpu(evlist, prot, mask); } int perf_evlist__create_maps(struct perf_evlist *evlist, pid_t target_pid, Loading @@ -348,7 +409,7 @@ int perf_evlist__create_maps(struct perf_evlist *evlist, pid_t target_pid, if (evlist->threads == NULL) return -1; if (target_tid != -1) if (cpu_list == NULL && target_tid != -1) evlist->cpus = cpu_map__dummy_new(); else evlist->cpus = cpu_map__new(cpu_list); Loading
tools/perf/util/evlist.h +2 −1 Original line number Diff line number Diff line Loading @@ -17,6 +17,7 @@ struct perf_evlist { struct hlist_head heads[PERF_EVLIST__HLIST_SIZE]; int nr_entries; int nr_fds; int nr_mmaps; int mmap_len; bool overwrite; union perf_event event_copy; Loading Loading @@ -46,7 +47,7 @@ void perf_evlist__add_pollfd(struct perf_evlist *evlist, int fd); struct perf_evsel *perf_evlist__id2evsel(struct perf_evlist *evlist, u64 id); union perf_event *perf_evlist__read_on_cpu(struct perf_evlist *self, int cpu); union perf_event *perf_evlist__mmap_read(struct perf_evlist *self, int idx); int perf_evlist__alloc_mmap(struct perf_evlist *evlist); int perf_evlist__mmap(struct perf_evlist *evlist, int pages, bool overwrite); Loading