1 #define _FILE_OFFSET_BITS 64
2
3 #include <sys/types.h>
4 #include <byteswap.h>
5 #include <unistd.h>
6 #include <stdio.h>
7 #include <stdlib.h>
8 /* ANDROID_CHANGE_BEGIN */
9 #if 0
10 #include <linux/list.h>
11 #include <linux/kernel.h>
12 #else
13 #include "include/linux/list.h"
14 #include "include/linux/kernel.h"
15 #endif
16 /* ANDROID_CHANGE_END */
17
18 #include "evlist.h"
19 #include "evsel.h"
20 #include "util.h"
21 #include "header.h"
22 #include "../perf.h"
23 #include "trace-event.h"
24 #include "session.h"
25 #include "symbol.h"
26 #include "debug.h"
27
28 static bool no_buildid_cache = false;
29
30 static int event_count;
31 static struct perf_trace_event_type *events;
32
perf_header__push_event(u64 id,const char * name)33 int perf_header__push_event(u64 id, const char *name)
34 {
35 if (strlen(name) > MAX_EVENT_NAME)
36 pr_warning("Event %s will be truncated\n", name);
37
38 if (!events) {
39 events = malloc(sizeof(struct perf_trace_event_type));
40 if (events == NULL)
41 return -ENOMEM;
42 } else {
43 struct perf_trace_event_type *nevents;
44
45 nevents = realloc(events, (event_count + 1) * sizeof(*events));
46 if (nevents == NULL)
47 return -ENOMEM;
48 events = nevents;
49 }
50 memset(&events[event_count], 0, sizeof(struct perf_trace_event_type));
51 events[event_count].event_id = id;
52 strncpy(events[event_count].name, name, MAX_EVENT_NAME - 1);
53 event_count++;
54 return 0;
55 }
56
perf_header__find_event(u64 id)57 char *perf_header__find_event(u64 id)
58 {
59 int i;
60 for (i = 0 ; i < event_count; i++) {
61 if (events[i].event_id == id)
62 return events[i].name;
63 }
64 return NULL;
65 }
66
67 /* ANDROID_CHANGE_BEGIN */
68 /*
69 * The string in the literal pool might not be aligned properly. ARM doesn't
70 * support unaligned loads with NEON registers
71 */
72 #if 0
73 static const char *__perf_magic = "PERFFILE";
74
75 #else
76 static const char const __perf_magic[9] __attribute__ ((aligned (8))) =
77 "PERFFILE";
78 #endif
79 /* ANDROID_CHANGE_END */
80
81 #define PERF_MAGIC (*(u64 *)__perf_magic)
82
83
84 struct perf_file_attr {
85 struct perf_event_attr attr;
86 struct perf_file_section ids;
87 };
88
perf_header__set_feat(struct perf_header * header,int feat)89 void perf_header__set_feat(struct perf_header *header, int feat)
90 {
91 set_bit(feat, header->adds_features);
92 }
93
perf_header__clear_feat(struct perf_header * header,int feat)94 void perf_header__clear_feat(struct perf_header *header, int feat)
95 {
96 clear_bit(feat, header->adds_features);
97 }
98
perf_header__has_feat(const struct perf_header * header,int feat)99 bool perf_header__has_feat(const struct perf_header *header, int feat)
100 {
101 return test_bit(feat, header->adds_features);
102 }
103
do_write(int fd,const void * buf,size_t size)104 static int do_write(int fd, const void *buf, size_t size)
105 {
106 while (size) {
107 int ret = write(fd, buf, size);
108
109 if (ret < 0)
110 return -errno;
111
112 size -= ret;
113 buf += ret;
114 }
115
116 return 0;
117 }
118
119 #define NAME_ALIGN 64
120
write_padded(int fd,const void * bf,size_t count,size_t count_aligned)121 static int write_padded(int fd, const void *bf, size_t count,
122 size_t count_aligned)
123 {
124 static const char zero_buf[NAME_ALIGN];
125 int err = do_write(fd, bf, count);
126
127 if (!err)
128 err = do_write(fd, zero_buf, count_aligned - count);
129
130 return err;
131 }
132
133 #define dsos__for_each_with_build_id(pos, head) \
134 list_for_each_entry(pos, head, node) \
135 if (!pos->has_build_id) \
136 continue; \
137 else
138
__dsos__write_buildid_table(struct list_head * head,pid_t pid,u16 misc,int fd)139 static int __dsos__write_buildid_table(struct list_head *head, pid_t pid,
140 u16 misc, int fd)
141 {
142 struct dso *pos;
143
144 dsos__for_each_with_build_id(pos, head) {
145 int err;
146 struct build_id_event b;
147 size_t len;
148
149 if (!pos->hit)
150 continue;
151 len = pos->long_name_len + 1;
152 /* ANDROID_CHANGE_BEGIN */
153 #if defined(__BIONIC__) || defined(__APPLE__)
154 len = KERNEL_ALIGN(len, NAME_ALIGN);
155 #else
156 len = ALIGN(len, NAME_ALIGN);
157 #endif
158 /* ANDROID_CHANGE_BEGIN */
159 memset(&b, 0, sizeof(b));
160 memcpy(&b.build_id, pos->build_id, sizeof(pos->build_id));
161 b.pid = pid;
162 b.header.misc = misc;
163 b.header.size = sizeof(b) + len;
164 err = do_write(fd, &b, sizeof(b));
165 if (err < 0)
166 return err;
167 err = write_padded(fd, pos->long_name,
168 pos->long_name_len + 1, len);
169 if (err < 0)
170 return err;
171 }
172
173 return 0;
174 }
175
machine__write_buildid_table(struct machine * machine,int fd)176 static int machine__write_buildid_table(struct machine *machine, int fd)
177 {
178 int err;
179 u16 kmisc = PERF_RECORD_MISC_KERNEL,
180 umisc = PERF_RECORD_MISC_USER;
181
182 if (!machine__is_host(machine)) {
183 kmisc = PERF_RECORD_MISC_GUEST_KERNEL;
184 umisc = PERF_RECORD_MISC_GUEST_USER;
185 }
186
187 err = __dsos__write_buildid_table(&machine->kernel_dsos, machine->pid,
188 kmisc, fd);
189 if (err == 0)
190 err = __dsos__write_buildid_table(&machine->user_dsos,
191 machine->pid, umisc, fd);
192 return err;
193 }
194
dsos__write_buildid_table(struct perf_header * header,int fd)195 static int dsos__write_buildid_table(struct perf_header *header, int fd)
196 {
197 struct perf_session *session = container_of(header,
198 struct perf_session, header);
199 struct rb_node *nd;
200 int err = machine__write_buildid_table(&session->host_machine, fd);
201
202 if (err)
203 return err;
204
205 for (nd = rb_first(&session->machines); nd; nd = rb_next(nd)) {
206 struct machine *pos = rb_entry(nd, struct machine, rb_node);
207 err = machine__write_buildid_table(pos, fd);
208 if (err)
209 break;
210 }
211 return err;
212 }
213
build_id_cache__add_s(const char * sbuild_id,const char * debugdir,const char * name,bool is_kallsyms)214 int build_id_cache__add_s(const char *sbuild_id, const char *debugdir,
215 const char *name, bool is_kallsyms)
216 {
217 const size_t size = PATH_MAX;
218 char *realname, *filename = malloc(size),
219 *linkname = malloc(size), *targetname;
220 int len, err = -1;
221
222 if (is_kallsyms) {
223 if (symbol_conf.kptr_restrict) {
224 pr_debug("Not caching a kptr_restrict'ed /proc/kallsyms\n");
225 return 0;
226 }
227 realname = (char *)name;
228 } else
229 realname = realpath(name, NULL);
230
231 if (realname == NULL || filename == NULL || linkname == NULL)
232 goto out_free;
233
234 len = snprintf(filename, size, "%s%s%s",
235 debugdir, is_kallsyms ? "/" : "", realname);
236 if (mkdir_p(filename, 0755))
237 goto out_free;
238
239 snprintf(filename + len, sizeof(filename) - len, "/%s", sbuild_id);
240
241 if (access(filename, F_OK)) {
242 if (is_kallsyms) {
243 if (copyfile("/proc/kallsyms", filename))
244 goto out_free;
245 } else if (link(realname, filename) && copyfile(name, filename))
246 goto out_free;
247 }
248
249 len = snprintf(linkname, size, "%s/.build-id/%.2s",
250 debugdir, sbuild_id);
251
252 if (access(linkname, X_OK) && mkdir_p(linkname, 0755))
253 goto out_free;
254
255 snprintf(linkname + len, size - len, "/%s", sbuild_id + 2);
256 targetname = filename + strlen(debugdir) - 5;
257 memcpy(targetname, "../..", 5);
258
259 if (symlink(targetname, linkname) == 0)
260 err = 0;
261 out_free:
262 if (!is_kallsyms)
263 free(realname);
264 free(filename);
265 free(linkname);
266 return err;
267 }
268
build_id_cache__add_b(const u8 * build_id,size_t build_id_size,const char * name,const char * debugdir,bool is_kallsyms)269 static int build_id_cache__add_b(const u8 *build_id, size_t build_id_size,
270 const char *name, const char *debugdir,
271 bool is_kallsyms)
272 {
273 char sbuild_id[BUILD_ID_SIZE * 2 + 1];
274
275 build_id__sprintf(build_id, build_id_size, sbuild_id);
276
277 return build_id_cache__add_s(sbuild_id, debugdir, name, is_kallsyms);
278 }
279
build_id_cache__remove_s(const char * sbuild_id,const char * debugdir)280 int build_id_cache__remove_s(const char *sbuild_id, const char *debugdir)
281 {
282 const size_t size = PATH_MAX;
283 char *filename = malloc(size),
284 *linkname = malloc(size);
285 int err = -1;
286
287 if (filename == NULL || linkname == NULL)
288 goto out_free;
289
290 snprintf(linkname, size, "%s/.build-id/%.2s/%s",
291 debugdir, sbuild_id, sbuild_id + 2);
292
293 if (access(linkname, F_OK))
294 goto out_free;
295
296 if (readlink(linkname, filename, size) < 0)
297 goto out_free;
298
299 if (unlink(linkname))
300 goto out_free;
301
302 /*
303 * Since the link is relative, we must make it absolute:
304 */
305 snprintf(linkname, size, "%s/.build-id/%.2s/%s",
306 debugdir, sbuild_id, filename);
307
308 if (unlink(linkname))
309 goto out_free;
310
311 err = 0;
312 out_free:
313 free(filename);
314 free(linkname);
315 return err;
316 }
317
dso__cache_build_id(struct dso * dso,const char * debugdir)318 static int dso__cache_build_id(struct dso *dso, const char *debugdir)
319 {
320 bool is_kallsyms = dso->kernel && dso->long_name[0] != '/';
321
322 return build_id_cache__add_b(dso->build_id, sizeof(dso->build_id),
323 dso->long_name, debugdir, is_kallsyms);
324 }
325
__dsos__cache_build_ids(struct list_head * head,const char * debugdir)326 static int __dsos__cache_build_ids(struct list_head *head, const char *debugdir)
327 {
328 struct dso *pos;
329 int err = 0;
330
331 dsos__for_each_with_build_id(pos, head)
332 if (dso__cache_build_id(pos, debugdir))
333 err = -1;
334
335 return err;
336 }
337
machine__cache_build_ids(struct machine * machine,const char * debugdir)338 static int machine__cache_build_ids(struct machine *machine, const char *debugdir)
339 {
340 int ret = __dsos__cache_build_ids(&machine->kernel_dsos, debugdir);
341 ret |= __dsos__cache_build_ids(&machine->user_dsos, debugdir);
342 return ret;
343 }
344
perf_session__cache_build_ids(struct perf_session * session)345 static int perf_session__cache_build_ids(struct perf_session *session)
346 {
347 struct rb_node *nd;
348 int ret;
349 char debugdir[PATH_MAX];
350
351 snprintf(debugdir, sizeof(debugdir), "%s", buildid_dir);
352
353 if (mkdir(debugdir, 0755) != 0 && errno != EEXIST)
354 return -1;
355
356 ret = machine__cache_build_ids(&session->host_machine, debugdir);
357
358 for (nd = rb_first(&session->machines); nd; nd = rb_next(nd)) {
359 struct machine *pos = rb_entry(nd, struct machine, rb_node);
360 ret |= machine__cache_build_ids(pos, debugdir);
361 }
362 return ret ? -1 : 0;
363 }
364
machine__read_build_ids(struct machine * machine,bool with_hits)365 static bool machine__read_build_ids(struct machine *machine, bool with_hits)
366 {
367 bool ret = __dsos__read_build_ids(&machine->kernel_dsos, with_hits);
368 ret |= __dsos__read_build_ids(&machine->user_dsos, with_hits);
369 return ret;
370 }
371
perf_session__read_build_ids(struct perf_session * session,bool with_hits)372 static bool perf_session__read_build_ids(struct perf_session *session, bool with_hits)
373 {
374 struct rb_node *nd;
375 bool ret = machine__read_build_ids(&session->host_machine, with_hits);
376
377 for (nd = rb_first(&session->machines); nd; nd = rb_next(nd)) {
378 struct machine *pos = rb_entry(nd, struct machine, rb_node);
379 ret |= machine__read_build_ids(pos, with_hits);
380 }
381
382 return ret;
383 }
384
perf_header__adds_write(struct perf_header * header,struct perf_evlist * evlist,int fd)385 static int perf_header__adds_write(struct perf_header *header,
386 struct perf_evlist *evlist, int fd)
387 {
388 int nr_sections;
389 struct perf_session *session;
390 struct perf_file_section *feat_sec;
391 int sec_size;
392 u64 sec_start;
393 int idx = 0, err;
394
395 session = container_of(header, struct perf_session, header);
396
397 if (perf_header__has_feat(header, HEADER_BUILD_ID &&
398 !perf_session__read_build_ids(session, true)))
399 perf_header__clear_feat(header, HEADER_BUILD_ID);
400
401 nr_sections = bitmap_weight(header->adds_features, HEADER_FEAT_BITS);
402 if (!nr_sections)
403 return 0;
404
405 feat_sec = calloc(sizeof(*feat_sec), nr_sections);
406 if (feat_sec == NULL)
407 return -ENOMEM;
408
409 sec_size = sizeof(*feat_sec) * nr_sections;
410
411 sec_start = header->data_offset + header->data_size;
412 lseek(fd, sec_start + sec_size, SEEK_SET);
413
414 if (perf_header__has_feat(header, HEADER_TRACE_INFO)) {
415 struct perf_file_section *trace_sec;
416
417 trace_sec = &feat_sec[idx++];
418
419 /* Write trace info */
420 trace_sec->offset = lseek(fd, 0, SEEK_CUR);
421 read_tracing_data(fd, &evlist->entries);
422 trace_sec->size = lseek(fd, 0, SEEK_CUR) - trace_sec->offset;
423 }
424
425 if (perf_header__has_feat(header, HEADER_BUILD_ID)) {
426 struct perf_file_section *buildid_sec;
427
428 buildid_sec = &feat_sec[idx++];
429
430 /* Write build-ids */
431 buildid_sec->offset = lseek(fd, 0, SEEK_CUR);
432 err = dsos__write_buildid_table(header, fd);
433 if (err < 0) {
434 pr_debug("failed to write buildid table\n");
435 goto out_free;
436 }
437 buildid_sec->size = lseek(fd, 0, SEEK_CUR) -
438 buildid_sec->offset;
439 if (!no_buildid_cache)
440 perf_session__cache_build_ids(session);
441 }
442
443 lseek(fd, sec_start, SEEK_SET);
444 err = do_write(fd, feat_sec, sec_size);
445 if (err < 0)
446 pr_debug("failed to write feature section\n");
447 out_free:
448 free(feat_sec);
449 return err;
450 }
451
perf_header__write_pipe(int fd)452 int perf_header__write_pipe(int fd)
453 {
454 struct perf_pipe_file_header f_header;
455 int err;
456
457 f_header = (struct perf_pipe_file_header){
458 .magic = PERF_MAGIC,
459 .size = sizeof(f_header),
460 };
461
462 err = do_write(fd, &f_header, sizeof(f_header));
463 if (err < 0) {
464 pr_debug("failed to write perf pipe header\n");
465 return err;
466 }
467
468 return 0;
469 }
470
perf_session__write_header(struct perf_session * session,struct perf_evlist * evlist,int fd,bool at_exit)471 int perf_session__write_header(struct perf_session *session,
472 struct perf_evlist *evlist,
473 int fd, bool at_exit)
474 {
475 struct perf_file_header f_header;
476 struct perf_file_attr f_attr;
477 struct perf_header *header = &session->header;
478 struct perf_evsel *attr, *pair = NULL;
479 int err;
480
481 lseek(fd, sizeof(f_header), SEEK_SET);
482
483 if (session->evlist != evlist)
484 pair = list_entry(session->evlist->entries.next, struct perf_evsel, node);
485
486 list_for_each_entry(attr, &evlist->entries, node) {
487 attr->id_offset = lseek(fd, 0, SEEK_CUR);
488 err = do_write(fd, attr->id, attr->ids * sizeof(u64));
489 if (err < 0) {
490 out_err_write:
491 pr_debug("failed to write perf header\n");
492 return err;
493 }
494 if (session->evlist != evlist) {
495 err = do_write(fd, pair->id, pair->ids * sizeof(u64));
496 if (err < 0)
497 goto out_err_write;
498 attr->ids += pair->ids;
499 pair = list_entry(pair->node.next, struct perf_evsel, node);
500 }
501 }
502
503 header->attr_offset = lseek(fd, 0, SEEK_CUR);
504
505 list_for_each_entry(attr, &evlist->entries, node) {
506 f_attr = (struct perf_file_attr){
507 .attr = attr->attr,
508 .ids = {
509 .offset = attr->id_offset,
510 .size = attr->ids * sizeof(u64),
511 }
512 };
513 err = do_write(fd, &f_attr, sizeof(f_attr));
514 if (err < 0) {
515 pr_debug("failed to write perf header attribute\n");
516 return err;
517 }
518 }
519
520 header->event_offset = lseek(fd, 0, SEEK_CUR);
521 header->event_size = event_count * sizeof(struct perf_trace_event_type);
522 if (events) {
523 err = do_write(fd, events, header->event_size);
524 if (err < 0) {
525 pr_debug("failed to write perf header events\n");
526 return err;
527 }
528 }
529
530 header->data_offset = lseek(fd, 0, SEEK_CUR);
531
532 if (at_exit) {
533 err = perf_header__adds_write(header, evlist, fd);
534 if (err < 0)
535 return err;
536 }
537
538 f_header = (struct perf_file_header){
539 .magic = PERF_MAGIC,
540 .size = sizeof(f_header),
541 .attr_size = sizeof(f_attr),
542 .attrs = {
543 .offset = header->attr_offset,
544 .size = evlist->nr_entries * sizeof(f_attr),
545 },
546 .data = {
547 .offset = header->data_offset,
548 .size = header->data_size,
549 },
550 .event_types = {
551 .offset = header->event_offset,
552 .size = header->event_size,
553 },
554 };
555
556 memcpy(&f_header.adds_features, &header->adds_features, sizeof(header->adds_features));
557
558 lseek(fd, 0, SEEK_SET);
559 err = do_write(fd, &f_header, sizeof(f_header));
560 if (err < 0) {
561 pr_debug("failed to write perf header\n");
562 return err;
563 }
564 lseek(fd, header->data_offset + header->data_size, SEEK_SET);
565
566 header->frozen = 1;
567 return 0;
568 }
569
perf_header__getbuffer64(struct perf_header * header,int fd,void * buf,size_t size)570 static int perf_header__getbuffer64(struct perf_header *header,
571 int fd, void *buf, size_t size)
572 {
573 if (readn(fd, buf, size) <= 0)
574 return -1;
575
576 if (header->needs_swap)
577 mem_bswap_64(buf, size);
578
579 return 0;
580 }
581
perf_header__process_sections(struct perf_header * header,int fd,int (* process)(struct perf_file_section * section,struct perf_header * ph,int feat,int fd))582 int perf_header__process_sections(struct perf_header *header, int fd,
583 int (*process)(struct perf_file_section *section,
584 struct perf_header *ph,
585 int feat, int fd))
586 {
587 struct perf_file_section *feat_sec;
588 int nr_sections;
589 int sec_size;
590 int idx = 0;
591 int err = -1, feat = 1;
592
593 nr_sections = bitmap_weight(header->adds_features, HEADER_FEAT_BITS);
594 if (!nr_sections)
595 return 0;
596
597 feat_sec = calloc(sizeof(*feat_sec), nr_sections);
598 if (!feat_sec)
599 return -1;
600
601 sec_size = sizeof(*feat_sec) * nr_sections;
602
603 lseek(fd, header->data_offset + header->data_size, SEEK_SET);
604
605 if (perf_header__getbuffer64(header, fd, feat_sec, sec_size))
606 goto out_free;
607
608 err = 0;
609 while (idx < nr_sections && feat < HEADER_LAST_FEATURE) {
610 if (perf_header__has_feat(header, feat)) {
611 struct perf_file_section *sec = &feat_sec[idx++];
612
613 err = process(sec, header, feat, fd);
614 if (err < 0)
615 break;
616 }
617 ++feat;
618 }
619 out_free:
620 free(feat_sec);
621 return err;
622 }
623
perf_file_header__read(struct perf_file_header * header,struct perf_header * ph,int fd)624 int perf_file_header__read(struct perf_file_header *header,
625 struct perf_header *ph, int fd)
626 {
627 lseek(fd, 0, SEEK_SET);
628
629 if (readn(fd, header, sizeof(*header)) <= 0 ||
630 memcmp(&header->magic, __perf_magic, sizeof(header->magic)))
631 return -1;
632
633 if (header->attr_size != sizeof(struct perf_file_attr)) {
634 u64 attr_size = bswap_64(header->attr_size);
635
636 if (attr_size != sizeof(struct perf_file_attr))
637 return -1;
638
639 mem_bswap_64(header, offsetof(struct perf_file_header,
640 adds_features));
641 ph->needs_swap = true;
642 }
643
644 if (header->size != sizeof(*header)) {
645 /* Support the previous format */
646 if (header->size == offsetof(typeof(*header), adds_features))
647 bitmap_zero(header->adds_features, HEADER_FEAT_BITS);
648 else
649 return -1;
650 }
651
652 memcpy(&ph->adds_features, &header->adds_features,
653 sizeof(ph->adds_features));
654 /*
655 * FIXME: hack that assumes that if we need swap the perf.data file
656 * may be coming from an arch with a different word-size, ergo different
657 * DEFINE_BITMAP format, investigate more later, but for now its mostly
658 * safe to assume that we have a build-id section. Trace files probably
659 * have several other issues in this realm anyway...
660 */
661 if (ph->needs_swap) {
662 memset(&ph->adds_features, 0, sizeof(ph->adds_features));
663 perf_header__set_feat(ph, HEADER_BUILD_ID);
664 }
665
666 ph->event_offset = header->event_types.offset;
667 ph->event_size = header->event_types.size;
668 ph->data_offset = header->data.offset;
669 ph->data_size = header->data.size;
670 return 0;
671 }
672
__event_process_build_id(struct build_id_event * bev,char * filename,struct perf_session * session)673 static int __event_process_build_id(struct build_id_event *bev,
674 char *filename,
675 struct perf_session *session)
676 {
677 int err = -1;
678 struct list_head *head;
679 struct machine *machine;
680 u16 misc;
681 struct dso *dso;
682 enum dso_kernel_type dso_type;
683
684 machine = perf_session__findnew_machine(session, bev->pid);
685 if (!machine)
686 goto out;
687
688 misc = bev->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;
689
690 switch (misc) {
691 case PERF_RECORD_MISC_KERNEL:
692 dso_type = DSO_TYPE_KERNEL;
693 head = &machine->kernel_dsos;
694 break;
695 case PERF_RECORD_MISC_GUEST_KERNEL:
696 dso_type = DSO_TYPE_GUEST_KERNEL;
697 head = &machine->kernel_dsos;
698 break;
699 case PERF_RECORD_MISC_USER:
700 case PERF_RECORD_MISC_GUEST_USER:
701 dso_type = DSO_TYPE_USER;
702 head = &machine->user_dsos;
703 break;
704 default:
705 goto out;
706 }
707
708 dso = __dsos__findnew(head, filename);
709 if (dso != NULL) {
710 char sbuild_id[BUILD_ID_SIZE * 2 + 1];
711
712 dso__set_build_id(dso, &bev->build_id);
713
714 if (filename[0] == '[')
715 dso->kernel = dso_type;
716
717 build_id__sprintf(dso->build_id, sizeof(dso->build_id),
718 sbuild_id);
719 pr_debug("build id event received for %s: %s\n",
720 dso->long_name, sbuild_id);
721 }
722
723 err = 0;
724 out:
725 return err;
726 }
727
perf_header__read_build_ids_abi_quirk(struct perf_header * header,int input,u64 offset,u64 size)728 static int perf_header__read_build_ids_abi_quirk(struct perf_header *header,
729 int input, u64 offset, u64 size)
730 {
731 struct perf_session *session = container_of(header, struct perf_session, header);
732 struct {
733 struct perf_event_header header;
734 /* ANDROID_CHANGE_BEGIN */
735 #if defined(__BIONIC__) || defined(__APPLE__)
736 u8 build_id[KERNEL_ALIGN(BUILD_ID_SIZE, sizeof(u64))];
737 #else
738 u8 build_id[ALIGN(BUILD_ID_SIZE, sizeof(u64))];
739 #endif
740 /* ANDROID_CHANGE_END */
741 char filename[0];
742 } old_bev;
743 struct build_id_event bev;
744 char filename[PATH_MAX];
745 u64 limit = offset + size;
746
747 while (offset < limit) {
748 ssize_t len;
749
750 if (read(input, &old_bev, sizeof(old_bev)) != sizeof(old_bev))
751 return -1;
752
753 if (header->needs_swap)
754 perf_event_header__bswap(&old_bev.header);
755
756 len = old_bev.header.size - sizeof(old_bev);
757 if (read(input, filename, len) != len)
758 return -1;
759
760 bev.header = old_bev.header;
761 bev.pid = 0;
762 memcpy(bev.build_id, old_bev.build_id, sizeof(bev.build_id));
763 __event_process_build_id(&bev, filename, session);
764
765 offset += bev.header.size;
766 }
767
768 return 0;
769 }
770
perf_header__read_build_ids(struct perf_header * header,int input,u64 offset,u64 size)771 static int perf_header__read_build_ids(struct perf_header *header,
772 int input, u64 offset, u64 size)
773 {
774 struct perf_session *session = container_of(header, struct perf_session, header);
775 struct build_id_event bev;
776 char filename[PATH_MAX];
777 u64 limit = offset + size, orig_offset = offset;
778 int err = -1;
779
780 while (offset < limit) {
781 ssize_t len;
782
783 if (read(input, &bev, sizeof(bev)) != sizeof(bev))
784 goto out;
785
786 if (header->needs_swap)
787 perf_event_header__bswap(&bev.header);
788
789 len = bev.header.size - sizeof(bev);
790 if (read(input, filename, len) != len)
791 goto out;
792 /*
793 * The a1645ce1 changeset:
794 *
795 * "perf: 'perf kvm' tool for monitoring guest performance from host"
796 *
797 * Added a field to struct build_id_event that broke the file
798 * format.
799 *
800 * Since the kernel build-id is the first entry, process the
801 * table using the old format if the well known
802 * '[kernel.kallsyms]' string for the kernel build-id has the
803 * first 4 characters chopped off (where the pid_t sits).
804 */
805 if (memcmp(filename, "nel.kallsyms]", 13) == 0) {
806 if (lseek(input, orig_offset, SEEK_SET) == (off_t)-1)
807 return -1;
808 return perf_header__read_build_ids_abi_quirk(header, input, offset, size);
809 }
810
811 __event_process_build_id(&bev, filename, session);
812
813 offset += bev.header.size;
814 }
815 err = 0;
816 out:
817 return err;
818 }
819
perf_file_section__process(struct perf_file_section * section,struct perf_header * ph,int feat,int fd)820 static int perf_file_section__process(struct perf_file_section *section,
821 struct perf_header *ph,
822 int feat, int fd)
823 {
824 if (lseek(fd, section->offset, SEEK_SET) == (off_t)-1) {
825 pr_debug("Failed to lseek to %" PRIu64 " offset for feature "
826 "%d, continuing...\n", section->offset, feat);
827 return 0;
828 }
829
830 switch (feat) {
831 case HEADER_TRACE_INFO:
832 trace_report(fd, false);
833 break;
834
835 case HEADER_BUILD_ID:
836 if (perf_header__read_build_ids(ph, fd, section->offset, section->size))
837 pr_debug("Failed to read buildids, continuing...\n");
838 break;
839 default:
840 pr_debug("unknown feature %d, continuing...\n", feat);
841 }
842
843 return 0;
844 }
845
perf_file_header__read_pipe(struct perf_pipe_file_header * header,struct perf_header * ph,int fd,bool repipe)846 static int perf_file_header__read_pipe(struct perf_pipe_file_header *header,
847 struct perf_header *ph, int fd,
848 bool repipe)
849 {
850 if (readn(fd, header, sizeof(*header)) <= 0 ||
851 memcmp(&header->magic, __perf_magic, sizeof(header->magic)))
852 return -1;
853
854 if (repipe && do_write(STDOUT_FILENO, header, sizeof(*header)) < 0)
855 return -1;
856
857 if (header->size != sizeof(*header)) {
858 u64 size = bswap_64(header->size);
859
860 if (size != sizeof(*header))
861 return -1;
862
863 ph->needs_swap = true;
864 }
865
866 return 0;
867 }
868
perf_header__read_pipe(struct perf_session * session,int fd)869 static int perf_header__read_pipe(struct perf_session *session, int fd)
870 {
871 struct perf_header *header = &session->header;
872 struct perf_pipe_file_header f_header;
873
874 if (perf_file_header__read_pipe(&f_header, header, fd,
875 session->repipe) < 0) {
876 pr_debug("incompatible file format\n");
877 return -EINVAL;
878 }
879
880 session->fd = fd;
881
882 return 0;
883 }
884
perf_session__read_header(struct perf_session * session,int fd)885 int perf_session__read_header(struct perf_session *session, int fd)
886 {
887 struct perf_header *header = &session->header;
888 struct perf_file_header f_header;
889 struct perf_file_attr f_attr;
890 u64 f_id;
891 int nr_attrs, nr_ids, i, j;
892
893 session->evlist = perf_evlist__new(NULL, NULL);
894 if (session->evlist == NULL)
895 return -ENOMEM;
896
897 if (session->fd_pipe)
898 return perf_header__read_pipe(session, fd);
899
900 if (perf_file_header__read(&f_header, header, fd) < 0) {
901 pr_debug("incompatible file format\n");
902 return -EINVAL;
903 }
904
905 nr_attrs = f_header.attrs.size / sizeof(f_attr);
906 lseek(fd, f_header.attrs.offset, SEEK_SET);
907
908 for (i = 0; i < nr_attrs; i++) {
909 struct perf_evsel *evsel;
910 off_t tmp;
911
912 if (readn(fd, &f_attr, sizeof(f_attr)) <= 0)
913 goto out_errno;
914
915 if (header->needs_swap)
916 perf_event__attr_swap(&f_attr.attr);
917
918 tmp = lseek(fd, 0, SEEK_CUR);
919 evsel = perf_evsel__new(&f_attr.attr, i);
920
921 if (evsel == NULL)
922 goto out_delete_evlist;
923 /*
924 * Do it before so that if perf_evsel__alloc_id fails, this
925 * entry gets purged too at perf_evlist__delete().
926 */
927 perf_evlist__add(session->evlist, evsel);
928
929 nr_ids = f_attr.ids.size / sizeof(u64);
930 /*
931 * We don't have the cpu and thread maps on the header, so
932 * for allocating the perf_sample_id table we fake 1 cpu and
933 * hattr->ids threads.
934 */
935 if (perf_evsel__alloc_id(evsel, 1, nr_ids))
936 goto out_delete_evlist;
937
938 lseek(fd, f_attr.ids.offset, SEEK_SET);
939
940 for (j = 0; j < nr_ids; j++) {
941 if (perf_header__getbuffer64(header, fd, &f_id, sizeof(f_id)))
942 goto out_errno;
943
944 perf_evlist__id_add(session->evlist, evsel, 0, j, f_id);
945 }
946
947 lseek(fd, tmp, SEEK_SET);
948 }
949
950 if (f_header.event_types.size) {
951 lseek(fd, f_header.event_types.offset, SEEK_SET);
952 events = malloc(f_header.event_types.size);
953 if (events == NULL)
954 return -ENOMEM;
955 if (perf_header__getbuffer64(header, fd, events,
956 f_header.event_types.size))
957 goto out_errno;
958 event_count = f_header.event_types.size / sizeof(struct perf_trace_event_type);
959 }
960
961 perf_header__process_sections(header, fd, perf_file_section__process);
962
963 lseek(fd, header->data_offset, SEEK_SET);
964
965 header->frozen = 1;
966 return 0;
967 out_errno:
968 return -errno;
969
970 out_delete_evlist:
971 perf_evlist__delete(session->evlist);
972 session->evlist = NULL;
973 return -ENOMEM;
974 }
975
perf_event__synthesize_attr(struct perf_event_attr * attr,u16 ids,u64 * id,perf_event__handler_t process,struct perf_session * session)976 int perf_event__synthesize_attr(struct perf_event_attr *attr, u16 ids, u64 *id,
977 perf_event__handler_t process,
978 struct perf_session *session)
979 {
980 union perf_event *ev;
981 size_t size;
982 int err;
983
984 size = sizeof(struct perf_event_attr);
985 /* ANDROID_CHANGE_BEGIN */
986 #if defined(__BIONIC__) || defined(__APPLE__)
987 size = KERNEL_ALIGN(size, sizeof(u64));
988 #else
989 size = ALIGN(size, sizeof(u64));
990 #endif
991 /* ANDROID_CHANGE_END */
992 size += sizeof(struct perf_event_header);
993 size += ids * sizeof(u64);
994
995 ev = malloc(size);
996
997 if (ev == NULL)
998 return -ENOMEM;
999
1000 ev->attr.attr = *attr;
1001 memcpy(ev->attr.id, id, ids * sizeof(u64));
1002
1003 ev->attr.header.type = PERF_RECORD_HEADER_ATTR;
1004 ev->attr.header.size = size;
1005
1006 err = process(ev, NULL, session);
1007
1008 free(ev);
1009
1010 return err;
1011 }
1012
perf_session__synthesize_attrs(struct perf_session * session,perf_event__handler_t process)1013 int perf_session__synthesize_attrs(struct perf_session *session,
1014 perf_event__handler_t process)
1015 {
1016 struct perf_evsel *attr;
1017 int err = 0;
1018
1019 list_for_each_entry(attr, &session->evlist->entries, node) {
1020 err = perf_event__synthesize_attr(&attr->attr, attr->ids,
1021 attr->id, process, session);
1022 if (err) {
1023 pr_debug("failed to create perf header attribute\n");
1024 return err;
1025 }
1026 }
1027
1028 return err;
1029 }
1030
perf_event__process_attr(union perf_event * event,struct perf_session * session)1031 int perf_event__process_attr(union perf_event *event,
1032 struct perf_session *session)
1033 {
1034 unsigned int i, ids, n_ids;
1035 struct perf_evsel *evsel;
1036
1037 if (session->evlist == NULL) {
1038 session->evlist = perf_evlist__new(NULL, NULL);
1039 if (session->evlist == NULL)
1040 return -ENOMEM;
1041 }
1042
1043 evsel = perf_evsel__new(&event->attr.attr,
1044 session->evlist->nr_entries);
1045 if (evsel == NULL)
1046 return -ENOMEM;
1047
1048 perf_evlist__add(session->evlist, evsel);
1049
1050 ids = event->header.size;
1051 ids -= (void *)&event->attr.id - (void *)event;
1052 n_ids = ids / sizeof(u64);
1053 /*
1054 * We don't have the cpu and thread maps on the header, so
1055 * for allocating the perf_sample_id table we fake 1 cpu and
1056 * hattr->ids threads.
1057 */
1058 if (perf_evsel__alloc_id(evsel, 1, n_ids))
1059 return -ENOMEM;
1060
1061 for (i = 0; i < n_ids; i++) {
1062 perf_evlist__id_add(session->evlist, evsel, 0, i,
1063 event->attr.id[i]);
1064 }
1065
1066 perf_session__update_sample_type(session);
1067
1068 return 0;
1069 }
1070
perf_event__synthesize_event_type(u64 event_id,char * name,perf_event__handler_t process,struct perf_session * session)1071 int perf_event__synthesize_event_type(u64 event_id, char *name,
1072 perf_event__handler_t process,
1073 struct perf_session *session)
1074 {
1075 union perf_event ev;
1076 size_t size = 0;
1077 int err = 0;
1078
1079 memset(&ev, 0, sizeof(ev));
1080
1081 ev.event_type.event_type.event_id = event_id;
1082 memset(ev.event_type.event_type.name, 0, MAX_EVENT_NAME);
1083 strncpy(ev.event_type.event_type.name, name, MAX_EVENT_NAME - 1);
1084
1085 ev.event_type.header.type = PERF_RECORD_HEADER_EVENT_TYPE;
1086 size = strlen(name);
1087 /* ANDROID_CHANGE_BEGIN */
1088 #if defined(__BIONIC__) || defined(__APPLE__)
1089 size = KERNEL_ALIGN(size, sizeof(u64));
1090 #else
1091 size = ALIGN(size, sizeof(u64));
1092 #endif
1093 /* ANDROID_CHANGE_END */
1094 ev.event_type.header.size = sizeof(ev.event_type) -
1095 (sizeof(ev.event_type.event_type.name) - size);
1096
1097 err = process(&ev, NULL, session);
1098
1099 return err;
1100 }
1101
perf_event__synthesize_event_types(perf_event__handler_t process,struct perf_session * session)1102 int perf_event__synthesize_event_types(perf_event__handler_t process,
1103 struct perf_session *session)
1104 {
1105 struct perf_trace_event_type *type;
1106 int i, err = 0;
1107
1108 for (i = 0; i < event_count; i++) {
1109 type = &events[i];
1110
1111 err = perf_event__synthesize_event_type(type->event_id,
1112 type->name, process,
1113 session);
1114 if (err) {
1115 pr_debug("failed to create perf header event type\n");
1116 return err;
1117 }
1118 }
1119
1120 return err;
1121 }
1122
perf_event__process_event_type(union perf_event * event,struct perf_session * session __unused)1123 int perf_event__process_event_type(union perf_event *event,
1124 struct perf_session *session __unused)
1125 {
1126 if (perf_header__push_event(event->event_type.event_type.event_id,
1127 event->event_type.event_type.name) < 0)
1128 return -ENOMEM;
1129
1130 return 0;
1131 }
1132
perf_event__synthesize_tracing_data(int fd,struct perf_evlist * evlist,perf_event__handler_t process,struct perf_session * session __unused)1133 int perf_event__synthesize_tracing_data(int fd, struct perf_evlist *evlist,
1134 perf_event__handler_t process,
1135 struct perf_session *session __unused)
1136 {
1137 union perf_event ev;
1138 ssize_t size = 0, aligned_size = 0, padding;
1139 int err __used = 0;
1140
1141 memset(&ev, 0, sizeof(ev));
1142
1143 ev.tracing_data.header.type = PERF_RECORD_HEADER_TRACING_DATA;
1144 size = read_tracing_data_size(fd, &evlist->entries);
1145 if (size <= 0)
1146 return size;
1147 /* ANDROID_CHANGE_BEGIN */
1148 #if defined(__BIONIC__) || defined(__APPLE__)
1149 aligned_size = KERNEL_ALIGN(size, sizeof(u64));
1150 #else
1151 aligned_size = ALIGN(size, sizeof(u64));
1152 #endif
1153 /* ANDROID_CHANGE_END */
1154 padding = aligned_size - size;
1155 ev.tracing_data.header.size = sizeof(ev.tracing_data);
1156 ev.tracing_data.size = aligned_size;
1157
1158 process(&ev, NULL, session);
1159
1160 err = read_tracing_data(fd, &evlist->entries);
1161 write_padded(fd, NULL, 0, padding);
1162
1163 return aligned_size;
1164 }
1165
perf_event__process_tracing_data(union perf_event * event,struct perf_session * session)1166 int perf_event__process_tracing_data(union perf_event *event,
1167 struct perf_session *session)
1168 {
1169 ssize_t size_read, padding, size = event->tracing_data.size;
1170 off_t offset = lseek(session->fd, 0, SEEK_CUR);
1171 char buf[BUFSIZ];
1172
1173 /* setup for reading amidst mmap */
1174 lseek(session->fd, offset + sizeof(struct tracing_data_event),
1175 SEEK_SET);
1176
1177 size_read = trace_report(session->fd, session->repipe);
1178
1179 /* ANDROID_CHANGE_BEGIN */
1180 #if defined(__BIONIC__) || defined(__APPLE__)
1181 padding = KERNEL_ALIGN(size_read, sizeof(u64)) - size_read;
1182 #else
1183 padding = ALIGN(size_read, sizeof(u64)) - size_read;
1184 #endif
1185 /* ANDROID_CHANGE_END */
1186
1187 if (read(session->fd, buf, padding) < 0)
1188 die("reading input file");
1189 if (session->repipe) {
1190 int retw = write(STDOUT_FILENO, buf, padding);
1191 if (retw <= 0 || retw != padding)
1192 die("repiping tracing data padding");
1193 }
1194
1195 if (size_read + padding != size)
1196 die("tracing data size mismatch");
1197
1198 return size_read + padding;
1199 }
1200
perf_event__synthesize_build_id(struct dso * pos,u16 misc,perf_event__handler_t process,struct machine * machine,struct perf_session * session)1201 int perf_event__synthesize_build_id(struct dso *pos, u16 misc,
1202 perf_event__handler_t process,
1203 struct machine *machine,
1204 struct perf_session *session)
1205 {
1206 union perf_event ev;
1207 size_t len;
1208 int err = 0;
1209
1210 if (!pos->hit)
1211 return err;
1212
1213 memset(&ev, 0, sizeof(ev));
1214
1215 len = pos->long_name_len + 1;
1216 /* ANDROID_CHANGE_BEGIN */
1217 #if defined(__BIONIC__) || defined(__APPLE__)
1218 len = KERNEL_ALIGN(len, NAME_ALIGN);
1219 #else
1220 len = ALIGN(len, NAME_ALIGN);
1221 #endif
1222 /* ANDROID_CHANGE_END */
1223 memcpy(&ev.build_id.build_id, pos->build_id, sizeof(pos->build_id));
1224 ev.build_id.header.type = PERF_RECORD_HEADER_BUILD_ID;
1225 ev.build_id.header.misc = misc;
1226 ev.build_id.pid = machine->pid;
1227 ev.build_id.header.size = sizeof(ev.build_id) + len;
1228 memcpy(&ev.build_id.filename, pos->long_name, pos->long_name_len);
1229
1230 err = process(&ev, NULL, session);
1231
1232 return err;
1233 }
1234
perf_event__process_build_id(union perf_event * event,struct perf_session * session)1235 int perf_event__process_build_id(union perf_event *event,
1236 struct perf_session *session)
1237 {
1238 __event_process_build_id(&event->build_id,
1239 event->build_id.filename,
1240 session);
1241 return 0;
1242 }
1243
disable_buildid_cache(void)1244 void disable_buildid_cache(void)
1245 {
1246 no_buildid_cache = true;
1247 }
1248