• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2018 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #include "src/traced/probes/ftrace/ftrace_config_muxer.h"
18 
19 #include <string.h>
20 #include <sys/types.h>
21 #include <unistd.h>
22 #include <cstdint>
23 
24 #include <algorithm>
25 #include <iterator>
26 #include <limits>
27 
28 #include "perfetto/base/compiler.h"
29 #include "perfetto/ext/base/utils.h"
30 #include "src/traced/probes/ftrace/atrace_wrapper.h"
31 #include "src/traced/probes/ftrace/compact_sched.h"
32 #include "src/traced/probes/ftrace/ftrace_config_utils.h"
33 #include "src/traced/probes/ftrace/ftrace_stats.h"
34 
35 #include "protos/perfetto/trace/ftrace/ftrace_event.pbzero.h"
36 
37 namespace perfetto {
38 namespace {
39 
40 constexpr uint64_t kDefaultLowRamPerCpuBufferSizeKb = 2 * (1ULL << 10);   // 2mb
41 constexpr uint64_t kDefaultHighRamPerCpuBufferSizeKb = 8 * (1ULL << 10);  // 8mb
42 constexpr uint64_t kMaxPerCpuBufferSizeKb = 64 * (1ULL << 10);  // 64mb
43 
44 // Threshold for physical ram size used when deciding on default kernel buffer
45 // sizes. We want to detect 8 GB, but the size reported through sysconf is
46 // usually lower.
47 constexpr uint64_t kHighMemBytes = 7 * (1ULL << 30);  // 7gb
48 
49 // A fake "syscall id" that indicates all syscalls should be recorded. This
50 // allows us to distinguish between the case where `syscall_events` is empty
51 // because raw_syscalls aren't enabled, or the case where it is and we want to
52 // record all events.
53 constexpr size_t kAllSyscallsId = kMaxSyscalls + 1;
54 
55 // trace_clocks in preference order.
56 // If this list is changed, the FtraceClocks enum in ftrace_event_bundle.proto
57 // and FtraceConfigMuxer::SetupClock() should be also changed accordingly.
58 constexpr const char* kClocks[] = {"boot", "global", "local"};
59 
60 // optional monotonic raw clock.
61 // Enabled by the "use_monotonic_raw_clock" option in the ftrace config.
62 constexpr const char* kClockMonoRaw = "mono_raw";
63 
AddEventGroup(const ProtoTranslationTable * table,const std::string & group,std::set<GroupAndName> * to)64 void AddEventGroup(const ProtoTranslationTable* table,
65                    const std::string& group,
66                    std::set<GroupAndName>* to) {
67   const std::vector<const Event*>* events = table->GetEventsByGroup(group);
68   if (!events)
69     return;
70   for (const Event* event : *events)
71     to->insert(GroupAndName(group, event->name));
72 }
73 
ReadEventsInGroupFromFs(const FtraceProcfs & ftrace_procfs,const std::string & group)74 std::set<GroupAndName> ReadEventsInGroupFromFs(
75     const FtraceProcfs& ftrace_procfs,
76     const std::string& group) {
77   std::set<std::string> names =
78       ftrace_procfs.GetEventNamesForGroup("events/" + group);
79   std::set<GroupAndName> events;
80   for (const auto& name : names)
81     events.insert(GroupAndName(group, name));
82   return events;
83 }
84 
EventToStringGroupAndName(const std::string & event)85 std::pair<std::string, std::string> EventToStringGroupAndName(
86     const std::string& event) {
87   auto slash_pos = event.find('/');
88   if (slash_pos == std::string::npos)
89     return std::make_pair("", event);
90   return std::make_pair(event.substr(0, slash_pos),
91                         event.substr(slash_pos + 1));
92 }
93 
UnionInPlace(const std::vector<std::string> & unsorted_a,std::vector<std::string> * out)94 void UnionInPlace(const std::vector<std::string>& unsorted_a,
95                   std::vector<std::string>* out) {
96   std::vector<std::string> a = unsorted_a;
97   std::sort(a.begin(), a.end());
98   std::sort(out->begin(), out->end());
99   std::vector<std::string> v;
100   std::set_union(a.begin(), a.end(), out->begin(), out->end(),
101                  std::back_inserter(v));
102   *out = std::move(v);
103 }
104 
IntersectInPlace(const std::vector<std::string> & unsorted_a,std::vector<std::string> * out)105 void IntersectInPlace(const std::vector<std::string>& unsorted_a,
106                       std::vector<std::string>* out) {
107   std::vector<std::string> a = unsorted_a;
108   std::sort(a.begin(), a.end());
109   std::sort(out->begin(), out->end());
110   std::vector<std::string> v;
111   std::set_intersection(a.begin(), a.end(), out->begin(), out->end(),
112                         std::back_inserter(v));
113   *out = std::move(v);
114 }
115 
116 // This is just to reduce binary size and stack frame size of the insertions.
117 // It effectively undoes STL's set::insert inlining.
InsertEvent(const char * group,const char * name,std::set<GroupAndName> * dst)118 void PERFETTO_NO_INLINE InsertEvent(const char* group,
119                                     const char* name,
120                                     std::set<GroupAndName>* dst) {
121   dst->insert(GroupAndName(group, name));
122 }
123 
124 }  // namespace
125 
GetFtraceEvents(const FtraceConfig & request,const ProtoTranslationTable * table)126 std::set<GroupAndName> FtraceConfigMuxer::GetFtraceEvents(
127     const FtraceConfig& request,
128     const ProtoTranslationTable* table) {
129   std::set<GroupAndName> events;
130   for (const auto& config_value : request.ftrace_events()) {
131     std::string group;
132     std::string name;
133     std::tie(group, name) = EventToStringGroupAndName(config_value);
134     if (name == "*") {
135       for (const auto& event : ReadEventsInGroupFromFs(*ftrace_, group))
136         events.insert(event);
137     } else if (group.empty()) {
138       // If there is no group specified, find an event with that name and
139       // use it's group.
140       const Event* e = table->GetEventByName(name);
141       if (!e) {
142         PERFETTO_DLOG(
143             "Event doesn't exist: %s. Include the group in the config to allow "
144             "the event to be output as a generic event.",
145             name.c_str());
146         continue;
147       }
148       events.insert(GroupAndName(e->group, e->name));
149     } else {
150       events.insert(GroupAndName(group, name));
151     }
152   }
153   if (RequiresAtrace(request)) {
154     InsertEvent("ftrace", "print", &events);
155 
156     // Ideally we should keep this code in sync with:
157     // platform/frameworks/native/cmds/atrace/atrace.cpp
158     // It's not a disaster if they go out of sync, we can always add the ftrace
159     // categories manually server side but this is user friendly and reduces the
160     // size of the configs.
161     for (const std::string& category : request.atrace_categories()) {
162       if (category == "gfx") {
163         AddEventGroup(table, "mdss", &events);
164         InsertEvent("mdss", "rotator_bw_ao_as_context", &events);
165         InsertEvent("mdss", "mdp_trace_counter", &events);
166         InsertEvent("mdss", "tracing_mark_write", &events);
167         InsertEvent("mdss", "mdp_cmd_wait_pingpong", &events);
168         InsertEvent("mdss", "mdp_cmd_kickoff", &events);
169         InsertEvent("mdss", "mdp_cmd_release_bw", &events);
170         InsertEvent("mdss", "mdp_cmd_readptr_done", &events);
171         InsertEvent("mdss", "mdp_cmd_pingpong_done", &events);
172         InsertEvent("mdss", "mdp_misr_crc", &events);
173         InsertEvent("mdss", "mdp_compare_bw", &events);
174         InsertEvent("mdss", "mdp_perf_update_bus", &events);
175         InsertEvent("mdss", "mdp_video_underrun_done", &events);
176         InsertEvent("mdss", "mdp_commit", &events);
177         InsertEvent("mdss", "mdp_mixer_update", &events);
178         InsertEvent("mdss", "mdp_perf_prefill_calc", &events);
179         InsertEvent("mdss", "mdp_perf_set_ot", &events);
180         InsertEvent("mdss", "mdp_perf_set_wm_levels", &events);
181         InsertEvent("mdss", "mdp_perf_set_panic_luts", &events);
182         InsertEvent("mdss", "mdp_perf_set_qos_luts", &events);
183         InsertEvent("mdss", "mdp_sspp_change", &events);
184         InsertEvent("mdss", "mdp_sspp_set", &events);
185         AddEventGroup(table, "mali", &events);
186         InsertEvent("mali", "tracing_mark_write", &events);
187 
188         AddEventGroup(table, "sde", &events);
189         InsertEvent("sde", "tracing_mark_write", &events);
190         InsertEvent("sde", "sde_perf_update_bus", &events);
191         InsertEvent("sde", "sde_perf_set_qos_luts", &events);
192         InsertEvent("sde", "sde_perf_set_ot", &events);
193         InsertEvent("sde", "sde_perf_set_danger_luts", &events);
194         InsertEvent("sde", "sde_perf_crtc_update", &events);
195         InsertEvent("sde", "sde_perf_calc_crtc", &events);
196         InsertEvent("sde", "sde_evtlog", &events);
197         InsertEvent("sde", "sde_encoder_underrun", &events);
198         InsertEvent("sde", "sde_cmd_release_bw", &events);
199 
200         AddEventGroup(table, "dpu", &events);
201         InsertEvent("dpu", "tracing_mark_write", &events);
202 
203         AddEventGroup(table, "g2d", &events);
204         InsertEvent("g2d", "tracing_mark_write", &events);
205         InsertEvent("g2d", "g2d_perf_update_qos", &events);
206 
207         AddEventGroup(table, "panel", &events);
208         InsertEvent("panel", "panel_write_generic", &events);
209         continue;
210       }
211 
212       if (category == "ion") {
213         InsertEvent("kmem", "ion_alloc_buffer_start", &events);
214         continue;
215       }
216 
217       // Note: sched_wakeup intentionally removed (diverging from atrace), as it
218       // is high-volume, but mostly redundant when sched_waking is also enabled.
219       // The event can still be enabled explicitly when necessary.
220       if (category == "sched") {
221         InsertEvent("sched", "sched_switch", &events);
222         InsertEvent("sched", "sched_waking", &events);
223         InsertEvent("sched", "sched_blocked_reason", &events);
224         InsertEvent("sched", "sched_cpu_hotplug", &events);
225         InsertEvent("sched", "sched_pi_setprio", &events);
226         InsertEvent("sched", "sched_process_exit", &events);
227         AddEventGroup(table, "cgroup", &events);
228         InsertEvent("cgroup", "cgroup_transfer_tasks", &events);
229         InsertEvent("cgroup", "cgroup_setup_root", &events);
230         InsertEvent("cgroup", "cgroup_rmdir", &events);
231         InsertEvent("cgroup", "cgroup_rename", &events);
232         InsertEvent("cgroup", "cgroup_remount", &events);
233         InsertEvent("cgroup", "cgroup_release", &events);
234         InsertEvent("cgroup", "cgroup_mkdir", &events);
235         InsertEvent("cgroup", "cgroup_destroy_root", &events);
236         InsertEvent("cgroup", "cgroup_attach_task", &events);
237         InsertEvent("oom", "oom_score_adj_update", &events);
238         InsertEvent("task", "task_rename", &events);
239         InsertEvent("task", "task_newtask", &events);
240 
241         AddEventGroup(table, "systrace", &events);
242         InsertEvent("systrace", "0", &events);
243 
244         AddEventGroup(table, "scm", &events);
245         InsertEvent("scm", "scm_call_start", &events);
246         InsertEvent("scm", "scm_call_end", &events);
247         continue;
248       }
249 
250       if (category == "irq") {
251         AddEventGroup(table, "irq", &events);
252         InsertEvent("irq", "tasklet_hi_exit", &events);
253         InsertEvent("irq", "tasklet_hi_entry", &events);
254         InsertEvent("irq", "tasklet_exit", &events);
255         InsertEvent("irq", "tasklet_entry", &events);
256         InsertEvent("irq", "softirq_raise", &events);
257         InsertEvent("irq", "softirq_exit", &events);
258         InsertEvent("irq", "softirq_entry", &events);
259         InsertEvent("irq", "irq_handler_exit", &events);
260         InsertEvent("irq", "irq_handler_entry", &events);
261         AddEventGroup(table, "ipi", &events);
262         InsertEvent("ipi", "ipi_raise", &events);
263         InsertEvent("ipi", "ipi_exit", &events);
264         InsertEvent("ipi", "ipi_entry", &events);
265         continue;
266       }
267 
268       if (category == "irqoff") {
269         InsertEvent("preemptirq", "irq_enable", &events);
270         InsertEvent("preemptirq", "irq_disable", &events);
271         continue;
272       }
273 
274       if (category == "preemptoff") {
275         InsertEvent("preemptirq", "preempt_enable", &events);
276         InsertEvent("preemptirq", "preempt_disable", &events);
277         continue;
278       }
279 
280       if (category == "i2c") {
281         AddEventGroup(table, "i2c", &events);
282         InsertEvent("i2c", "i2c_read", &events);
283         InsertEvent("i2c", "i2c_write", &events);
284         InsertEvent("i2c", "i2c_result", &events);
285         InsertEvent("i2c", "i2c_reply", &events);
286         InsertEvent("i2c", "smbus_read", &events);
287         InsertEvent("i2c", "smbus_write", &events);
288         InsertEvent("i2c", "smbus_result", &events);
289         InsertEvent("i2c", "smbus_reply", &events);
290         continue;
291       }
292 
293       if (category == "freq") {
294         InsertEvent("power", "cpu_frequency", &events);
295         InsertEvent("power", "gpu_frequency", &events);
296         InsertEvent("power", "clock_set_rate", &events);
297         InsertEvent("power", "clock_disable", &events);
298         InsertEvent("power", "clock_enable", &events);
299         InsertEvent("clk", "clk_set_rate", &events);
300         InsertEvent("clk", "clk_disable", &events);
301         InsertEvent("clk", "clk_enable", &events);
302         InsertEvent("power", "cpu_frequency_limits", &events);
303         InsertEvent("power", "suspend_resume", &events);
304         InsertEvent("cpuhp", "cpuhp_enter", &events);
305         InsertEvent("cpuhp", "cpuhp_exit", &events);
306         InsertEvent("cpuhp", "cpuhp_pause", &events);
307         AddEventGroup(table, "msm_bus", &events);
308         InsertEvent("msm_bus", "bus_update_request_end", &events);
309         InsertEvent("msm_bus", "bus_update_request", &events);
310         InsertEvent("msm_bus", "bus_rules_matches", &events);
311         InsertEvent("msm_bus", "bus_max_votes", &events);
312         InsertEvent("msm_bus", "bus_client_status", &events);
313         InsertEvent("msm_bus", "bus_bke_params", &events);
314         InsertEvent("msm_bus", "bus_bimc_config_limiter", &events);
315         InsertEvent("msm_bus", "bus_avail_bw", &events);
316         InsertEvent("msm_bus", "bus_agg_bw", &events);
317         continue;
318       }
319 
320       if (category == "membus") {
321         AddEventGroup(table, "memory_bus", &events);
322         continue;
323       }
324 
325       if (category == "idle") {
326         InsertEvent("power", "cpu_idle", &events);
327         continue;
328       }
329 
330       if (category == "disk") {
331         InsertEvent("f2fs", "f2fs_sync_file_enter", &events);
332         InsertEvent("f2fs", "f2fs_sync_file_exit", &events);
333         InsertEvent("f2fs", "f2fs_write_begin", &events);
334         InsertEvent("f2fs", "f2fs_write_end", &events);
335         InsertEvent("f2fs", "f2fs_iostat", &events);
336         InsertEvent("f2fs", "f2fs_iostat_latency", &events);
337         InsertEvent("ext4", "ext4_da_write_begin", &events);
338         InsertEvent("ext4", "ext4_da_write_end", &events);
339         InsertEvent("ext4", "ext4_sync_file_enter", &events);
340         InsertEvent("ext4", "ext4_sync_file_exit", &events);
341         InsertEvent("block", "block_bio_queue", &events);
342         InsertEvent("block", "block_bio_complete", &events);
343         InsertEvent("ufs", "ufshcd_command", &events);
344         continue;
345       }
346 
347       if (category == "mmc") {
348         AddEventGroup(table, "mmc", &events);
349         continue;
350       }
351 
352       if (category == "load") {
353         AddEventGroup(table, "cpufreq_interactive", &events);
354         continue;
355       }
356 
357       if (category == "sync") {
358         // linux kernel < 4.9
359         AddEventGroup(table, "sync", &events);
360         InsertEvent("sync", "sync_pt", &events);
361         InsertEvent("sync", "sync_timeline", &events);
362         InsertEvent("sync", "sync_wait", &events);
363         // linux kernel == 4.9.x
364         AddEventGroup(table, "fence", &events);
365         InsertEvent("fence", "fence_annotate_wait_on", &events);
366         InsertEvent("fence", "fence_destroy", &events);
367         InsertEvent("fence", "fence_emit", &events);
368         InsertEvent("fence", "fence_enable_signal", &events);
369         InsertEvent("fence", "fence_init", &events);
370         InsertEvent("fence", "fence_signaled", &events);
371         InsertEvent("fence", "fence_wait_end", &events);
372         InsertEvent("fence", "fence_wait_start", &events);
373         // linux kernel > 4.9
374         AddEventGroup(table, "dma_fence", &events);
375         continue;
376       }
377 
378       if (category == "workq") {
379         AddEventGroup(table, "workqueue", &events);
380         InsertEvent("workqueue", "workqueue_queue_work", &events);
381         InsertEvent("workqueue", "workqueue_execute_start", &events);
382         InsertEvent("workqueue", "workqueue_execute_end", &events);
383         InsertEvent("workqueue", "workqueue_activate_work", &events);
384         continue;
385       }
386 
387       if (category == "memreclaim") {
388         InsertEvent("vmscan", "mm_vmscan_direct_reclaim_begin", &events);
389         InsertEvent("vmscan", "mm_vmscan_direct_reclaim_end", &events);
390         InsertEvent("vmscan", "mm_vmscan_kswapd_wake", &events);
391         InsertEvent("vmscan", "mm_vmscan_kswapd_sleep", &events);
392         AddEventGroup(table, "lowmemorykiller", &events);
393         InsertEvent("lowmemorykiller", "lowmemory_kill", &events);
394         continue;
395       }
396 
397       if (category == "regulators") {
398         AddEventGroup(table, "regulator", &events);
399         events.insert(
400             GroupAndName("regulator", "regulator_set_voltage_complete"));
401         InsertEvent("regulator", "regulator_set_voltage", &events);
402         InsertEvent("regulator", "regulator_enable_delay", &events);
403         InsertEvent("regulator", "regulator_enable_complete", &events);
404         InsertEvent("regulator", "regulator_enable", &events);
405         InsertEvent("regulator", "regulator_disable_complete", &events);
406         InsertEvent("regulator", "regulator_disable", &events);
407         continue;
408       }
409 
410       if (category == "binder_driver") {
411         InsertEvent("binder", "binder_transaction", &events);
412         InsertEvent("binder", "binder_transaction_received", &events);
413         InsertEvent("binder", "binder_transaction_alloc_buf", &events);
414         InsertEvent("binder", "binder_set_priority", &events);
415         continue;
416       }
417 
418       if (category == "binder_lock") {
419         InsertEvent("binder", "binder_lock", &events);
420         InsertEvent("binder", "binder_locked", &events);
421         InsertEvent("binder", "binder_unlock", &events);
422         continue;
423       }
424 
425       if (category == "pagecache") {
426         AddEventGroup(table, "filemap", &events);
427         events.insert(
428             GroupAndName("filemap", "mm_filemap_delete_from_page_cache"));
429         InsertEvent("filemap", "mm_filemap_add_to_page_cache", &events);
430         InsertEvent("filemap", "filemap_set_wb_err", &events);
431         InsertEvent("filemap", "file_check_and_advance_wb_err", &events);
432         continue;
433       }
434 
435       if (category == "memory") {
436         // Use rss_stat_throttled if supported
437         if (ftrace_->SupportsRssStatThrottled()) {
438           InsertEvent("synthetic", "rss_stat_throttled", &events);
439         } else {
440           InsertEvent("kmem", "rss_stat", &events);
441         }
442         InsertEvent("kmem", "ion_heap_grow", &events);
443         InsertEvent("kmem", "ion_heap_shrink", &events);
444         // ion_stat supersedes ion_heap_grow / shrink for kernel 4.19+
445         InsertEvent("ion", "ion_stat", &events);
446         InsertEvent("mm_event", "mm_event_record", &events);
447         InsertEvent("dmabuf_heap", "dma_heap_stat", &events);
448         InsertEvent("gpu_mem", "gpu_mem_total", &events);
449         continue;
450       }
451 
452       if (category == "thermal") {
453         InsertEvent("thermal", "thermal_temperature", &events);
454         InsertEvent("thermal", "cdev_update", &events);
455         continue;
456       }
457 
458       if (category == "camera") {
459         AddEventGroup(table, "lwis", &events);
460         InsertEvent("lwis", "tracing_mark_write", &events);
461         continue;
462       }
463     }
464   }
465 
466   // recording a subset of syscalls -> enable the backing events
467   if (request.syscall_events_size() > 0) {
468     InsertEvent("raw_syscalls", "sys_enter", &events);
469     InsertEvent("raw_syscalls", "sys_exit", &events);
470   }
471 
472   // function_graph tracer emits two builtin ftrace events
473   if (request.enable_function_graph()) {
474     InsertEvent("ftrace", "funcgraph_entry", &events);
475     InsertEvent("ftrace", "funcgraph_exit", &events);
476   }
477 
478   // If throttle_rss_stat: true, use the rss_stat_throttled event if supported
479   if (request.throttle_rss_stat() && ftrace_->SupportsRssStatThrottled()) {
480     auto it = std::find_if(
481         events.begin(), events.end(), [](const GroupAndName& event) {
482           return event.group() == "kmem" && event.name() == "rss_stat";
483         });
484 
485     if (it != events.end()) {
486       events.erase(it);
487       InsertEvent("synthetic", "rss_stat_throttled", &events);
488     }
489   }
490 
491   return events;
492 }
493 
GetSyscallsReturningFds(const SyscallTable & syscalls)494 base::FlatSet<int64_t> FtraceConfigMuxer::GetSyscallsReturningFds(
495     const SyscallTable& syscalls) {
496   auto insertSyscallId = [&syscalls](base::FlatSet<int64_t>& set,
497                                      const char* syscall) {
498     auto syscall_id = syscalls.GetByName(syscall);
499     if (syscall_id)
500       set.insert(static_cast<int64_t>(*syscall_id));
501   };
502 
503   base::FlatSet<int64_t> call_ids;
504   insertSyscallId(call_ids, "sys_open");
505   insertSyscallId(call_ids, "sys_openat");
506   insertSyscallId(call_ids, "sys_socket");
507   insertSyscallId(call_ids, "sys_dup");
508   insertSyscallId(call_ids, "sys_dup2");
509   insertSyscallId(call_ids, "sys_dup3");
510   return call_ids;
511 }
512 
FilterHasGroup(const EventFilter & filter,const std::string & group)513 bool FtraceConfigMuxer::FilterHasGroup(const EventFilter& filter,
514                                        const std::string& group) {
515   const std::vector<const Event*>* events = table_->GetEventsByGroup(group);
516   if (!events) {
517     return false;
518   }
519 
520   for (const Event* event : *events) {
521     if (filter.IsEventEnabled(event->ftrace_event_id)) {
522       return true;
523     }
524   }
525   return false;
526 }
527 
BuildSyscallFilter(const EventFilter & ftrace_filter,const FtraceConfig & request)528 EventFilter FtraceConfigMuxer::BuildSyscallFilter(
529     const EventFilter& ftrace_filter,
530     const FtraceConfig& request) {
531   EventFilter output;
532 
533   if (!FilterHasGroup(ftrace_filter, "raw_syscalls")) {
534     return output;
535   }
536 
537   if (request.syscall_events().empty()) {
538     output.AddEnabledEvent(kAllSyscallsId);
539     return output;
540   }
541 
542   for (const std::string& syscall : request.syscall_events()) {
543     std::optional<size_t> id = syscalls_.GetByName(syscall);
544     if (!id.has_value()) {
545       PERFETTO_ELOG("Can't enable %s, syscall not known", syscall.c_str());
546       continue;
547     }
548     output.AddEnabledEvent(*id);
549   }
550 
551   return output;
552 }
553 
SetSyscallEventFilter(const EventFilter & extra_syscalls)554 bool FtraceConfigMuxer::SetSyscallEventFilter(
555     const EventFilter& extra_syscalls) {
556   EventFilter syscall_filter;
557 
558   syscall_filter.EnableEventsFrom(extra_syscalls);
559   for (const auto& id_config : ds_configs_) {
560     const perfetto::FtraceDataSourceConfig& config = id_config.second;
561     syscall_filter.EnableEventsFrom(config.syscall_filter);
562   }
563 
564   std::set<size_t> filter_set = syscall_filter.GetEnabledEvents();
565   if (syscall_filter.IsEventEnabled(kAllSyscallsId)) {
566     filter_set.clear();
567   }
568 
569   if (current_state_.syscall_filter != filter_set) {
570     if (!ftrace_->SetSyscallFilter(filter_set)) {
571       return false;
572     }
573 
574     current_state_.syscall_filter = filter_set;
575   }
576 
577   return true;
578 }
579 
FtraceConfigMuxer(FtraceProcfs * ftrace,AtraceWrapper * atrace_wrapper,ProtoTranslationTable * table,SyscallTable syscalls,std::map<std::string,std::vector<GroupAndName>> vendor_events,bool secondary_instance)580 FtraceConfigMuxer::FtraceConfigMuxer(
581     FtraceProcfs* ftrace,
582     AtraceWrapper* atrace_wrapper,
583     ProtoTranslationTable* table,
584     SyscallTable syscalls,
585     std::map<std::string, std::vector<GroupAndName>> vendor_events,
586     bool secondary_instance)
587     : ftrace_(ftrace),
588       atrace_wrapper_(atrace_wrapper),
589       table_(table),
590       syscalls_(std::move(syscalls)),
591       current_state_(),
592       ds_configs_(),
593       vendor_events_(std::move(vendor_events)),
594       secondary_instance_(secondary_instance) {}
595 FtraceConfigMuxer::~FtraceConfigMuxer() = default;
596 
SetupConfig(FtraceConfigId id,const FtraceConfig & request,FtraceSetupErrors * errors)597 bool FtraceConfigMuxer::SetupConfig(FtraceConfigId id,
598                                     const FtraceConfig& request,
599                                     FtraceSetupErrors* errors) {
600   EventFilter filter;
601   if (ds_configs_.empty()) {
602     PERFETTO_DCHECK(active_configs_.empty());
603 
604     // If someone outside of perfetto is using a non-nop tracer, yield. We can't
605     // realistically figure out all notions of "in use" even if we look at
606     // set_event or events/enable, so this is all we check for.
607     if (!request.preserve_ftrace_buffer() && !ftrace_->IsTracingAvailable()) {
608       PERFETTO_ELOG(
609           "ftrace in use by non-Perfetto. Check that %s current_tracer is nop.",
610           ftrace_->GetRootPath().c_str());
611       return false;
612     }
613 
614     // Clear tracefs state, remembering which value of "tracing_on" to restore
615     // to after we're done, though we won't restore the rest of the tracefs
616     // state.
617     current_state_.saved_tracing_on = ftrace_->GetTracingOn();
618     if (!request.preserve_ftrace_buffer()) {
619       ftrace_->SetTracingOn(false);
620       // This will fail on release ("user") builds due to ACLs, but that's
621       // acceptable since the per-event enabling/disabling should still be
622       // balanced.
623       ftrace_->DisableAllEvents();
624       ftrace_->ClearTrace();
625     }
626 
627     // Set up the rest of the tracefs state, without starting it.
628     // Notes:
629     // * resizing buffers can be quite slow (up to hundreds of ms).
630     // * resizing buffers may truncate existing contents if the new size is
631     // smaller, which matters to the preserve_ftrace_buffer option.
632     if (!request.preserve_ftrace_buffer()) {
633       SetupClock(request);
634       SetupBufferSize(request);
635     }
636   }
637 
638   std::set<GroupAndName> events = GetFtraceEvents(request, table_);
639 
640   // Vendors can provide a set of extra ftrace categories to be enabled when a
641   // specific atrace category is used (e.g. "gfx" -> ["my_hw/my_custom_event",
642   // "my_hw/my_special_gpu"]). Merge them with the hard coded events for each
643   // categories.
644   for (const std::string& category : request.atrace_categories()) {
645     if (vendor_events_.count(category)) {
646       for (const GroupAndName& event : vendor_events_[category]) {
647         events.insert(event);
648       }
649     }
650   }
651 
652   if (RequiresAtrace(request)) {
653     if (secondary_instance_) {
654       PERFETTO_ELOG(
655           "Secondary ftrace instances do not support atrace_categories and "
656           "atrace_apps options as they affect global state");
657       return false;
658     }
659     if (!atrace_wrapper_->SupportsUserspaceOnly() && !ds_configs_.empty()) {
660       PERFETTO_ELOG(
661           "Concurrent atrace sessions are not supported before Android P, "
662           "bailing out.");
663       return false;
664     }
665     UpdateAtrace(request, errors ? &errors->atrace_errors : nullptr);
666   }
667 
668   for (const auto& group_and_name : events) {
669     const Event* event = table_->GetOrCreateEvent(group_and_name);
670     if (!event) {
671       PERFETTO_DLOG("Can't enable %s, event not known",
672                     group_and_name.ToString().c_str());
673       if (errors)
674         errors->unknown_ftrace_events.push_back(group_and_name.ToString());
675       continue;
676     }
677     // Niche option to skip events that are in the config, but don't have a
678     // dedicated proto for the event in perfetto. Otherwise such events will be
679     // encoded as GenericFtraceEvent.
680     if (request.disable_generic_events() &&
681         event->proto_field_id ==
682             protos::pbzero::FtraceEvent::kGenericFieldNumber) {
683       if (errors)
684         errors->failed_ftrace_events.push_back(group_and_name.ToString());
685       continue;
686     }
687     // Note: ftrace events are always implicitly enabled (and don't have an
688     // "enable" file). So they aren't tracked by the central event filter (but
689     // still need to be added to the per data source event filter to retain
690     // the events during parsing).
691     if (current_state_.ftrace_events.IsEventEnabled(event->ftrace_event_id) ||
692         std::string("ftrace") == event->group) {
693       filter.AddEnabledEvent(event->ftrace_event_id);
694       continue;
695     }
696     if (ftrace_->EnableEvent(event->group, event->name)) {
697       current_state_.ftrace_events.AddEnabledEvent(event->ftrace_event_id);
698       filter.AddEnabledEvent(event->ftrace_event_id);
699     } else {
700       PERFETTO_DPLOG("Failed to enable %s.", group_and_name.ToString().c_str());
701       if (errors)
702         errors->failed_ftrace_events.push_back(group_and_name.ToString());
703     }
704   }
705 
706   EventFilter syscall_filter = BuildSyscallFilter(filter, request);
707   if (!SetSyscallEventFilter(syscall_filter)) {
708     PERFETTO_ELOG("Failed to set raw_syscall ftrace filter in SetupConfig");
709     return false;
710   }
711 
712   // Kernel function tracing (function_graph).
713   // Note 1: there is no cleanup in |RemoveConfig| because tracers cannot be
714   // changed while tracing pipes are opened. So we'll keep the current_tracer
715   // until all data sources are gone, at which point ftrace_controller will
716   // make an explicit call to |ResetCurrentTracer|.
717   // Note 2: we don't track the set of filters ourselves and instead let the
718   // kernel statefully collate them, hence the use of |AppendFunctionFilters|.
719   // This is because each concurrent data source that wants funcgraph will get
720   // all of the enabled functions (we don't go as far as doing per-DS event
721   // steering in the parser), and we don't want to remove functions midway
722   // through a trace (but some might get added).
723   if (request.enable_function_graph()) {
724     if (!current_state_.funcgraph_on && !ftrace_->ClearFunctionFilters())
725       return false;
726     if (!current_state_.funcgraph_on && !ftrace_->ClearFunctionGraphFilters())
727       return false;
728     if (!ftrace_->AppendFunctionFilters(request.function_filters()))
729       return false;
730     if (!ftrace_->AppendFunctionGraphFilters(request.function_graph_roots()))
731       return false;
732     if (!current_state_.funcgraph_on &&
733         !ftrace_->SetCurrentTracer("function_graph")) {
734       PERFETTO_LOG(
735           "Unable to enable function_graph tracing since a concurrent ftrace "
736           "data source is using a different tracer");
737       return false;
738     }
739     current_state_.funcgraph_on = true;
740   }
741   const auto& compact_format = table_->compact_sched_format();
742   auto compact_sched = CreateCompactSchedConfig(
743       request, filter.IsEventEnabled(compact_format.sched_switch.event_id),
744       compact_format);
745   if (errors && !compact_format.format_valid) {
746     errors->failed_ftrace_events.emplace_back(
747         "perfetto/compact_sched (unexpected sched event format)");
748   }
749 
750   std::optional<FtracePrintFilterConfig> ftrace_print_filter;
751   if (request.has_print_filter()) {
752     ftrace_print_filter =
753         FtracePrintFilterConfig::Create(request.print_filter(), table_);
754     if (!ftrace_print_filter.has_value()) {
755       if (errors) {
756         errors->failed_ftrace_events.emplace_back(
757             "ftrace/print (unexpected format for filtering)");
758       }
759     }
760   }
761 
762   std::vector<std::string> apps(request.atrace_apps());
763   std::vector<std::string> categories(request.atrace_categories());
764   ds_configs_.emplace(
765       std::piecewise_construct, std::forward_as_tuple(id),
766       std::forward_as_tuple(
767           std::move(filter), std::move(syscall_filter), compact_sched,
768           std::move(ftrace_print_filter), std::move(apps),
769           std::move(categories), request.symbolize_ksyms(),
770           request.drain_buffer_percent(), GetSyscallsReturningFds(syscalls_)));
771   return true;
772 }
773 
ActivateConfig(FtraceConfigId id)774 bool FtraceConfigMuxer::ActivateConfig(FtraceConfigId id) {
775   if (!id || ds_configs_.count(id) == 0) {
776     PERFETTO_DFATAL("Config not found");
777     return false;
778   }
779 
780   bool first_config = active_configs_.empty();
781   active_configs_.insert(id);
782 
783   // Pick the lowest buffer_percent across the new set of active configs.
784   if (!UpdateBufferPercent()) {
785     PERFETTO_ELOG(
786         "Invalid FtraceConfig.drain_buffer_percent or "
787         "/sys/kernel/tracing/buffer_percent file permissions.");
788     // carry on, non-critical error
789   }
790 
791   // Enable kernel event writer.
792   if (first_config) {
793     if (!ftrace_->SetTracingOn(true)) {
794       PERFETTO_ELOG("Failed to enable ftrace.");
795       active_configs_.erase(id);
796       return false;
797     }
798   }
799   return true;
800 }
801 
RemoveConfig(FtraceConfigId config_id)802 bool FtraceConfigMuxer::RemoveConfig(FtraceConfigId config_id) {
803   if (!config_id || !ds_configs_.erase(config_id))
804     return false;
805   EventFilter expected_ftrace_events;
806   std::vector<std::string> expected_apps;
807   std::vector<std::string> expected_categories;
808   for (const auto& id_config : ds_configs_) {
809     const perfetto::FtraceDataSourceConfig& config = id_config.second;
810     expected_ftrace_events.EnableEventsFrom(config.event_filter);
811     UnionInPlace(config.atrace_apps, &expected_apps);
812     UnionInPlace(config.atrace_categories, &expected_categories);
813   }
814   // At this point expected_{apps,categories} contains the union of the
815   // leftover configs (if any) that should be still on. However we did not
816   // necessarily succeed in turning on atrace for each of those configs
817   // previously so we now intersect the {apps,categories} that we *did* manage
818   // to turn on with those we want on to determine the new state we should aim
819   // for:
820   IntersectInPlace(current_state_.atrace_apps, &expected_apps);
821   IntersectInPlace(current_state_.atrace_categories, &expected_categories);
822   // Work out if there is any difference between the current state and the
823   // desired state: It's sufficient to compare sizes here (since we know from
824   // above that expected_{apps,categories} is now a subset of
825   // atrace_{apps,categories}:
826   bool atrace_changed =
827       (current_state_.atrace_apps.size() != expected_apps.size()) ||
828       (current_state_.atrace_categories.size() != expected_categories.size());
829 
830   if (!SetSyscallEventFilter(/*extra_syscalls=*/{})) {
831     PERFETTO_ELOG("Failed to set raw_syscall ftrace filter in RemoveConfig");
832   }
833 
834   // Disable any events that are currently enabled, but are not in any configs
835   // anymore.
836   std::set<size_t> event_ids = current_state_.ftrace_events.GetEnabledEvents();
837   for (size_t id : event_ids) {
838     if (expected_ftrace_events.IsEventEnabled(id))
839       continue;
840     const Event* event = table_->GetEventById(id);
841     // Any event that was enabled must exist.
842     PERFETTO_DCHECK(event);
843     if (ftrace_->DisableEvent(event->group, event->name))
844       current_state_.ftrace_events.DisableEvent(event->ftrace_event_id);
845   }
846 
847   auto active_it = active_configs_.find(config_id);
848   if (active_it != active_configs_.end()) {
849     active_configs_.erase(active_it);
850     if (active_configs_.empty()) {
851       // This was the last active config for now, but potentially more dormant
852       // configs need to be activated. We are not interested in reading while no
853       // active configs so diasble tracing_on here.
854       ftrace_->SetTracingOn(false);
855     }
856   }
857 
858   // Update buffer_percent to the minimum of the remaining configs.
859   UpdateBufferPercent();
860 
861   // Even if we don't have any other active configs, we might still have idle
862   // configs around. Tear down the rest of the ftrace config only if all
863   // configs are removed.
864   if (ds_configs_.empty()) {
865     if (ftrace_->SetCpuBufferSizeInPages(1))
866       current_state_.cpu_buffer_size_pages = 1;
867     ftrace_->SetBufferPercent(50);
868     ftrace_->DisableAllEvents();
869     ftrace_->ClearTrace();
870     ftrace_->SetTracingOn(current_state_.saved_tracing_on);
871   }
872 
873   if (current_state_.atrace_on) {
874     if (expected_apps.empty() && expected_categories.empty()) {
875       DisableAtrace();
876     } else if (atrace_changed) {
877       // Update atrace to remove the no longer wanted categories/apps. For
878       // some categories this won't disable them (e.g. categories that just
879       // enable ftrace events) for those there is nothing we can do till the
880       // last ftrace config is removed.
881       if (StartAtrace(expected_apps, expected_categories,
882                       /*atrace_errors=*/nullptr)) {
883         // Update current_state_ to reflect this change.
884         current_state_.atrace_apps = expected_apps;
885         current_state_.atrace_categories = expected_categories;
886       }
887     }
888   }
889 
890   return true;
891 }
892 
ResetCurrentTracer()893 bool FtraceConfigMuxer::ResetCurrentTracer() {
894   if (!current_state_.funcgraph_on)
895     return true;
896   if (!ftrace_->ResetCurrentTracer()) {
897     PERFETTO_PLOG("Failed to reset current_tracer to nop");
898     return false;
899   }
900   current_state_.funcgraph_on = false;
901   if (!ftrace_->ClearFunctionFilters()) {
902     PERFETTO_PLOG("Failed to reset set_ftrace_filter.");
903     return false;
904   }
905   if (!ftrace_->ClearFunctionGraphFilters()) {
906     PERFETTO_PLOG("Failed to reset set_function_graph.");
907     return false;
908   }
909   return true;
910 }
911 
GetDataSourceConfig(FtraceConfigId id)912 const FtraceDataSourceConfig* FtraceConfigMuxer::GetDataSourceConfig(
913     FtraceConfigId id) {
914   if (!ds_configs_.count(id))
915     return nullptr;
916   return &ds_configs_.at(id);
917 }
918 
SetupClock(const FtraceConfig & config)919 void FtraceConfigMuxer::SetupClock(const FtraceConfig& config) {
920   std::string current_clock = ftrace_->GetClock();
921   std::set<std::string> clocks = ftrace_->AvailableClocks();
922 
923   if (config.has_use_monotonic_raw_clock() &&
924       config.use_monotonic_raw_clock() && clocks.count(kClockMonoRaw)) {
925     ftrace_->SetClock(kClockMonoRaw);
926     current_clock = kClockMonoRaw;
927   } else {
928     for (size_t i = 0; i < base::ArraySize(kClocks); i++) {
929       std::string clock = std::string(kClocks[i]);
930       if (!clocks.count(clock))
931         continue;
932       if (current_clock == clock)
933         break;
934       ftrace_->SetClock(clock);
935       current_clock = clock;
936       break;
937     }
938   }
939 
940   namespace pb0 = protos::pbzero;
941   if (current_clock == "boot") {
942     // "boot" is the default expectation on modern kernels, which is why we
943     // don't have an explicit FTRACE_CLOCK_BOOT enum and leave it unset.
944     // See comments in ftrace_event_bundle.proto.
945     current_state_.ftrace_clock = pb0::FTRACE_CLOCK_UNSPECIFIED;
946   } else if (current_clock == "global") {
947     current_state_.ftrace_clock = pb0::FTRACE_CLOCK_GLOBAL;
948   } else if (current_clock == "local") {
949     current_state_.ftrace_clock = pb0::FTRACE_CLOCK_LOCAL;
950   } else if (current_clock == kClockMonoRaw) {
951     current_state_.ftrace_clock = pb0::FTRACE_CLOCK_MONO_RAW;
952   } else {
953     current_state_.ftrace_clock = pb0::FTRACE_CLOCK_UNKNOWN;
954   }
955 }
956 
SetupBufferSize(const FtraceConfig & request)957 void FtraceConfigMuxer::SetupBufferSize(const FtraceConfig& request) {
958   int64_t phys_ram_pages = sysconf(_SC_PHYS_PAGES);
959   size_t pages = ComputeCpuBufferSizeInPages(request.buffer_size_kb(),
960                                              request.buffer_size_lower_bound(),
961                                              phys_ram_pages);
962   ftrace_->SetCpuBufferSizeInPages(pages);
963   current_state_.cpu_buffer_size_pages = pages;
964 }
965 
966 // Post-conditions:
967 // 1. result >= 1 (should have at least one page per CPU)
968 // 2. result < kMaxTotalBufferSizeKb / (page_size / 1024)
969 // 3. If input is 0 output is a good default number
ComputeCpuBufferSizeInPages(size_t requested_buffer_size_kb,bool buffer_size_lower_bound,int64_t sysconf_phys_pages)970 size_t ComputeCpuBufferSizeInPages(size_t requested_buffer_size_kb,
971                                    bool buffer_size_lower_bound,
972                                    int64_t sysconf_phys_pages) {
973   uint32_t page_sz = base::GetSysPageSize();
974   uint64_t default_size_kb =
975       (sysconf_phys_pages > 0 &&
976        (static_cast<uint64_t>(sysconf_phys_pages) >= (kHighMemBytes / page_sz)))
977           ? kDefaultHighRamPerCpuBufferSizeKb
978           : kDefaultLowRamPerCpuBufferSizeKb;
979 
980   size_t actual_size_kb = requested_buffer_size_kb;
981   if ((requested_buffer_size_kb == 0) ||
982       (buffer_size_lower_bound && default_size_kb > requested_buffer_size_kb)) {
983     actual_size_kb = default_size_kb;
984   }
985 
986   if (actual_size_kb > kMaxPerCpuBufferSizeKb) {
987     PERFETTO_ELOG(
988         "The requested ftrace buf size (%zu KB) is too big, capping to %" PRIu64
989         " KB",
990         actual_size_kb, kMaxPerCpuBufferSizeKb);
991     actual_size_kb = kMaxPerCpuBufferSizeKb;
992   }
993 
994   size_t pages = actual_size_kb / (page_sz / 1024);
995   return pages ? pages : 1;
996 }
997 
GetPerCpuBufferSizePages()998 size_t FtraceConfigMuxer::GetPerCpuBufferSizePages() {
999   return current_state_.cpu_buffer_size_pages;
1000 }
1001 
1002 // If new_cfg_id is set, consider it in addition to already active configs
1003 // as we're trying to activate it.
UpdateBufferPercent()1004 bool FtraceConfigMuxer::UpdateBufferPercent() {
1005   uint32_t kUnsetPercent = std::numeric_limits<uint32_t>::max();
1006   uint32_t min_percent = kUnsetPercent;
1007   for (auto cfg_id : active_configs_) {
1008     auto ds_it = ds_configs_.find(cfg_id);
1009     if (ds_it != ds_configs_.end() && ds_it->second.buffer_percent > 0) {
1010       min_percent = std::min(min_percent, ds_it->second.buffer_percent);
1011     }
1012   }
1013   if (min_percent == kUnsetPercent)
1014     return true;
1015   // Let the kernel ignore values >100.
1016   return ftrace_->SetBufferPercent(min_percent);
1017 }
1018 
UpdateAtrace(const FtraceConfig & request,std::string * atrace_errors)1019 void FtraceConfigMuxer::UpdateAtrace(const FtraceConfig& request,
1020                                      std::string* atrace_errors) {
1021   // We want to avoid poisoning current_state_.atrace_{categories, apps}
1022   // if for some reason these args make atrace unhappy so we stash the
1023   // union into temps and only update current_state_ if we successfully
1024   // run atrace.
1025 
1026   std::vector<std::string> combined_categories = request.atrace_categories();
1027   UnionInPlace(current_state_.atrace_categories, &combined_categories);
1028 
1029   std::vector<std::string> combined_apps = request.atrace_apps();
1030   UnionInPlace(current_state_.atrace_apps, &combined_apps);
1031 
1032   if (current_state_.atrace_on &&
1033       combined_apps.size() == current_state_.atrace_apps.size() &&
1034       combined_categories.size() == current_state_.atrace_categories.size()) {
1035     return;
1036   }
1037 
1038   if (StartAtrace(combined_apps, combined_categories, atrace_errors)) {
1039     current_state_.atrace_categories = combined_categories;
1040     current_state_.atrace_apps = combined_apps;
1041     current_state_.atrace_on = true;
1042   }
1043 }
1044 
StartAtrace(const std::vector<std::string> & apps,const std::vector<std::string> & categories,std::string * atrace_errors)1045 bool FtraceConfigMuxer::StartAtrace(const std::vector<std::string>& apps,
1046                                     const std::vector<std::string>& categories,
1047                                     std::string* atrace_errors) {
1048   PERFETTO_DLOG("Update atrace config...");
1049 
1050   std::vector<std::string> args;
1051   args.push_back("atrace");  // argv0 for exec()
1052   args.push_back("--async_start");
1053   if (atrace_wrapper_->SupportsUserspaceOnly())
1054     args.push_back("--only_userspace");
1055 
1056   for (const auto& category : categories)
1057     args.push_back(category);
1058 
1059   if (!apps.empty()) {
1060     args.push_back("-a");
1061     std::string arg = "";
1062     for (const auto& app : apps) {
1063       arg += app;
1064       arg += ",";
1065     }
1066     arg.resize(arg.size() - 1);
1067     args.push_back(arg);
1068   }
1069 
1070   bool result = atrace_wrapper_->RunAtrace(args, atrace_errors);
1071   PERFETTO_DLOG("...done (%s)", result ? "success" : "fail");
1072   return result;
1073 }
1074 
DisableAtrace()1075 void FtraceConfigMuxer::DisableAtrace() {
1076   PERFETTO_DCHECK(current_state_.atrace_on);
1077 
1078   PERFETTO_DLOG("Stop atrace...");
1079 
1080   std::vector<std::string> args{"atrace", "--async_stop"};
1081   if (atrace_wrapper_->SupportsUserspaceOnly())
1082     args.push_back("--only_userspace");
1083   if (atrace_wrapper_->RunAtrace(args, /*atrace_errors=*/nullptr)) {
1084     current_state_.atrace_categories.clear();
1085     current_state_.atrace_apps.clear();
1086     current_state_.atrace_on = false;
1087   }
1088 
1089   PERFETTO_DLOG("...done");
1090 }
1091 
1092 }  // namespace perfetto
1093