• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (c) 2018 Google, Inc.
3  *
4  * SPDX-License-Identifier: GPL-2.0-or-later
5  *
6  * A CFS task is affined to a particular CPU. The task runs as a CPU hog for a
7  * while then as a very small task for a while. The latency for the CPU
8  * frequency of the CPU to reach max and then min is verified.
9  */
10 
11 #define _GNU_SOURCE
12 #include <errno.h>
13 #include <pthread.h>
14 #include <sched.h>
15 #include <time.h>
16 
17 #include "tst_test.h"
18 #include "tst_safe_file_ops.h"
19 #include "tst_safe_pthread.h"
20 
21 #include "trace_parse.h"
22 #include "util.h"
23 
24 #define TRACE_EVENTS "sched_process_exit sched_process_fork cpu_frequency"
25 
26 #define MAX_FREQ_INCREASE_LATENCY_US 70000
27 #define MAX_FREQ_DECREASE_LATENCY_US 70000
28 
29 static int test_cpu;
30 
31 #define BURN_MSEC 500
burn_fn(void * arg LTP_ATTRIBUTE_UNUSED)32 static void *burn_fn(void *arg LTP_ATTRIBUTE_UNUSED)
33 {
34 	int i = 0;
35 	unsigned int scaling_min_freq, scaling_cur_freq;
36 	char scaling_freq_file[60];
37 
38 	affine(test_cpu);
39 
40 	/*
41 	 * wait a bit to allow any hacks to boost frequency on migration
42 	 * to take effect
43 	 */
44 	usleep(200);
45 
46 	sprintf(scaling_freq_file,
47 		"/sys/devices/system/cpu/cpu%d/cpufreq/scaling_min_freq",
48 		test_cpu);
49 	SAFE_FILE_SCANF(scaling_freq_file, "%d", &scaling_min_freq);
50 
51 	sprintf(scaling_freq_file,
52 		"/sys/devices/system/cpu/cpu%d/cpufreq/scaling_cur_freq",
53 		test_cpu);
54 
55 	/* wait for test_cpu to reach scaling_min_freq */
56 	while(i++ < 10) {
57 		usleep(100 * 1000);
58 		SAFE_FILE_SCANF(scaling_freq_file, "%d",
59 				&scaling_cur_freq);
60 		if (scaling_cur_freq == scaling_min_freq)
61 			break;
62 	}
63 	if (i >= 10) {
64 		printf("Unable to reach scaling_min_freq before test!\n");
65 		return NULL;
66 	}
67 
68 	SAFE_FILE_PRINTF(TRACING_DIR "trace_marker", "affined");
69 	burn(BURN_MSEC * 1000, 0);
70 	SAFE_FILE_PRINTF(TRACING_DIR "trace_marker", "small task");
71 	burn(BURN_MSEC * 1000, 1);
72 
73 	return NULL;
74 }
75 
parse_results(void)76 static int parse_results(void)
77 {
78 	int i;
79 
80 	int start_idx;
81 	int sleep_idx;
82 	unsigned int max_freq_seen = 0;
83 	int max_freq_seen_idx;
84 	unsigned int min_freq_seen = UINT_MAX;
85 	int min_freq_seen_idx;
86 
87 	char scaling_freq_file[60];
88 	unsigned int scaling_max_freq;
89 	unsigned int scaling_min_freq;
90 
91 	unsigned int increase_latency_usec;
92 	unsigned int decrease_latency_usec;
93 
94 	/* find starting timestamp of test */
95 	for (i = 0; i < num_trace_records; i++)
96 		if (trace[i].event_type == TRACE_RECORD_TRACING_MARK_WRITE &&
97 		    !strcmp(trace[i].event_data, "affined"))
98 			break;
99 	if (i == num_trace_records) {
100 		printf("Did not find start of burn thread in trace!\n");
101 		return -1;
102 	}
103 	start_idx = i;
104 
105 	/* find timestamp when burn thread sleeps */
106 	for (; i < num_trace_records; i++)
107 		if (trace[i].event_type == TRACE_RECORD_TRACING_MARK_WRITE &&
108 		    !strcmp(trace[i].event_data, "small task"))
109 				break;
110 	if (i == num_trace_records) {
111 		printf("Did not find switch to small task of burn thread in "
112 		       "trace!\n");
113 		return -1;
114 	}
115 	sleep_idx = i;
116 
117 	/* find highest CPU frequency bewteen start and sleep timestamp */
118 	for (i = start_idx; i < sleep_idx; i++)
119 		if (trace[i].event_type == TRACE_RECORD_CPU_FREQUENCY) {
120 			struct trace_cpu_frequency *t = trace[i].event_data;
121 			if (t->cpu == test_cpu && t->state > max_freq_seen) {
122 				max_freq_seen = t->state;
123 				max_freq_seen_idx = i;
124 			}
125 		}
126 	if (max_freq_seen == 0) {
127 		printf("No freq events between start and sleep!\n");
128 		return -1;
129 	}
130 
131 	/* find lowest CPU frequency between sleep timestamp and end */
132 	for (; i < num_trace_records; i++)
133 		if (trace[i].event_type == TRACE_RECORD_CPU_FREQUENCY) {
134 			struct trace_cpu_frequency *t = trace[i].event_data;
135 			if (t->cpu == test_cpu && t->state < min_freq_seen) {
136 				min_freq_seen = t->state;
137 				min_freq_seen_idx = i;
138 			}
139 		}
140 	if (min_freq_seen == UINT_MAX) {
141 		printf("No freq events between sleep and end!\n");
142 		return -1;
143 	}
144 
145 	/* is highest CPU freq equal or greater than highest reported in
146 	 * scaling_max_freq? */
147 	sprintf(scaling_freq_file,
148 		"/sys/devices/system/cpu/cpu%d/cpufreq/scaling_max_freq",
149 		test_cpu);
150 	SAFE_FILE_SCANF(scaling_freq_file, "%d", &scaling_max_freq);
151 	if (max_freq_seen < scaling_max_freq) {
152 		printf("CPU%d did not reach scaling_max_freq!\n",
153 		       test_cpu);
154 		return -1;
155 	} else {
156 		printf("CPU%d reached %d MHz during test "
157 		       "(scaling_max_freq %d MHz).\n", test_cpu,
158 		       max_freq_seen / 1000, scaling_max_freq / 1000);
159 	}
160 
161 	/* is lowest CPU freq equal or less than scaling_min_freq? */
162 	sprintf(scaling_freq_file,
163 		"/sys/devices/system/cpu/cpu%d/cpufreq/scaling_min_freq",
164 		test_cpu);
165 	SAFE_FILE_SCANF(scaling_freq_file, "%d", &scaling_min_freq);
166 	if (min_freq_seen > scaling_min_freq) {
167 		printf("CPU%d did not reach scaling_min_freq!\n",
168 		       test_cpu);
169 		return -1;
170 	} else {
171 		printf("CPU%d reached %d MHz after test "
172 		       "(scaling_min_freq %d Mhz).\n",
173 		       test_cpu, min_freq_seen / 1000,
174 		       scaling_min_freq / 1000);
175 	}
176 
177 	/* calculate and check latencies */
178 	increase_latency_usec = trace[max_freq_seen_idx].ts.sec * USEC_PER_SEC +
179 		trace[max_freq_seen_idx].ts.usec;
180 	increase_latency_usec -= trace[start_idx].ts.sec * USEC_PER_SEC +
181 		trace[start_idx].ts.usec;
182 
183 	decrease_latency_usec = trace[min_freq_seen_idx].ts.sec * USEC_PER_SEC +
184 		trace[min_freq_seen_idx].ts.usec;
185 	decrease_latency_usec -= trace[sleep_idx].ts.sec * USEC_PER_SEC +
186 		trace[sleep_idx].ts.usec;
187 
188 	printf("Increase latency: %d usec\n", increase_latency_usec);
189 	printf("Decrease latency: %d usec\n", decrease_latency_usec);
190 
191 	return (increase_latency_usec > MAX_FREQ_INCREASE_LATENCY_US ||
192 		decrease_latency_usec > MAX_FREQ_DECREASE_LATENCY_US);
193 }
194 
run(void)195 static void run(void)
196 {
197 	pthread_t burn_thread;
198 
199 	tst_res(TINFO, "Max acceptable latency to fmax: %d usec\n",
200 		MAX_FREQ_INCREASE_LATENCY_US);
201 	tst_res(TINFO, "Max acceptable latency to fmin: %d usec\n",
202 		MAX_FREQ_DECREASE_LATENCY_US);
203 
204 	test_cpu = tst_ncpus() - 1;
205 	printf("CPU hog will be bound to CPU %d.\n", test_cpu);
206 
207 	/* configure and enable tracing */
208 	SAFE_FILE_PRINTF(TRACING_DIR "tracing_on", "0");
209 	SAFE_FILE_PRINTF(TRACING_DIR "buffer_size_kb", "16384");
210 	SAFE_FILE_PRINTF(TRACING_DIR "set_event", TRACE_EVENTS);
211 	SAFE_FILE_PRINTF(TRACING_DIR "trace", "\n");
212 	SAFE_FILE_PRINTF(TRACING_DIR "tracing_on", "1");
213 
214 	SAFE_PTHREAD_CREATE(&burn_thread, NULL, burn_fn, NULL);
215 	SAFE_PTHREAD_JOIN(burn_thread, NULL);
216 
217 	/* disable tracing */
218 	SAFE_FILE_PRINTF(TRACING_DIR "tracing_on", "0");
219 	LOAD_TRACE();
220 
221 	if (parse_results())
222 		tst_res(TFAIL, "Governor did not meet latency targets.\n");
223 	else
224 		tst_res(TPASS, "Governor met latency targets.\n");
225 }
226 
227 static struct tst_test test = {
228 	.test_all = run,
229 	.cleanup = trace_cleanup,
230 };
231