1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * DAMON Debugfs Interface
4 *
5 * Author: SeongJae Park <sjpark@amazon.de>
6 */
7
8 #define pr_fmt(fmt) "damon-dbgfs: " fmt
9
10 #include <linux/damon.h>
11 #include <linux/debugfs.h>
12 #include <linux/file.h>
13 #include <linux/mm.h>
14 #include <linux/module.h>
15 #include <linux/page_idle.h>
16 #include <linux/slab.h>
17
18 static struct damon_ctx **dbgfs_ctxs;
19 static int dbgfs_nr_ctxs;
20 static struct dentry **dbgfs_dirs;
21 static DEFINE_MUTEX(damon_dbgfs_lock);
22
23 /*
24 * Returns non-empty string on success, negative error code otherwise.
25 */
user_input_str(const char __user * buf,size_t count,loff_t * ppos)26 static char *user_input_str(const char __user *buf, size_t count, loff_t *ppos)
27 {
28 char *kbuf;
29 ssize_t ret;
30
31 /* We do not accept continuous write */
32 if (*ppos)
33 return ERR_PTR(-EINVAL);
34
35 kbuf = kmalloc(count + 1, GFP_KERNEL | __GFP_NOWARN);
36 if (!kbuf)
37 return ERR_PTR(-ENOMEM);
38
39 ret = simple_write_to_buffer(kbuf, count + 1, ppos, buf, count);
40 if (ret != count) {
41 kfree(kbuf);
42 return ERR_PTR(-EIO);
43 }
44 kbuf[ret] = '\0';
45
46 return kbuf;
47 }
48
dbgfs_attrs_read(struct file * file,char __user * buf,size_t count,loff_t * ppos)49 static ssize_t dbgfs_attrs_read(struct file *file,
50 char __user *buf, size_t count, loff_t *ppos)
51 {
52 struct damon_ctx *ctx = file->private_data;
53 char kbuf[128];
54 int ret;
55
56 mutex_lock(&ctx->kdamond_lock);
57 ret = scnprintf(kbuf, ARRAY_SIZE(kbuf), "%lu %lu %lu %lu %lu\n",
58 ctx->sample_interval, ctx->aggr_interval,
59 ctx->primitive_update_interval, ctx->min_nr_regions,
60 ctx->max_nr_regions);
61 mutex_unlock(&ctx->kdamond_lock);
62
63 return simple_read_from_buffer(buf, count, ppos, kbuf, ret);
64 }
65
dbgfs_attrs_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)66 static ssize_t dbgfs_attrs_write(struct file *file,
67 const char __user *buf, size_t count, loff_t *ppos)
68 {
69 struct damon_ctx *ctx = file->private_data;
70 unsigned long s, a, r, minr, maxr;
71 char *kbuf;
72 ssize_t ret;
73
74 kbuf = user_input_str(buf, count, ppos);
75 if (IS_ERR(kbuf))
76 return PTR_ERR(kbuf);
77
78 if (sscanf(kbuf, "%lu %lu %lu %lu %lu",
79 &s, &a, &r, &minr, &maxr) != 5) {
80 ret = -EINVAL;
81 goto out;
82 }
83
84 mutex_lock(&ctx->kdamond_lock);
85 if (ctx->kdamond) {
86 ret = -EBUSY;
87 goto unlock_out;
88 }
89
90 ret = damon_set_attrs(ctx, s, a, r, minr, maxr);
91 if (!ret)
92 ret = count;
93 unlock_out:
94 mutex_unlock(&ctx->kdamond_lock);
95 out:
96 kfree(kbuf);
97 return ret;
98 }
99
sprint_schemes(struct damon_ctx * c,char * buf,ssize_t len)100 static ssize_t sprint_schemes(struct damon_ctx *c, char *buf, ssize_t len)
101 {
102 struct damos *s;
103 int written = 0;
104 int rc;
105
106 damon_for_each_scheme(s, c) {
107 rc = scnprintf(&buf[written], len - written,
108 "%lu %lu %u %u %u %u %d %lu %lu %lu %u %u %u %d %lu %lu %lu %lu %lu %lu %lu %lu %lu\n",
109 s->min_sz_region, s->max_sz_region,
110 s->min_nr_accesses, s->max_nr_accesses,
111 s->min_age_region, s->max_age_region,
112 s->action,
113 s->quota.ms, s->quota.sz,
114 s->quota.reset_interval,
115 s->quota.weight_sz,
116 s->quota.weight_nr_accesses,
117 s->quota.weight_age,
118 s->wmarks.metric, s->wmarks.interval,
119 s->wmarks.high, s->wmarks.mid, s->wmarks.low,
120 s->stat.nr_tried, s->stat.sz_tried,
121 s->stat.nr_applied, s->stat.sz_applied,
122 s->stat.qt_exceeds);
123 if (!rc)
124 return -ENOMEM;
125
126 written += rc;
127 }
128 return written;
129 }
130
dbgfs_schemes_read(struct file * file,char __user * buf,size_t count,loff_t * ppos)131 static ssize_t dbgfs_schemes_read(struct file *file, char __user *buf,
132 size_t count, loff_t *ppos)
133 {
134 struct damon_ctx *ctx = file->private_data;
135 char *kbuf;
136 ssize_t len;
137
138 kbuf = kmalloc(count, GFP_KERNEL | __GFP_NOWARN);
139 if (!kbuf)
140 return -ENOMEM;
141
142 mutex_lock(&ctx->kdamond_lock);
143 len = sprint_schemes(ctx, kbuf, count);
144 mutex_unlock(&ctx->kdamond_lock);
145 if (len < 0)
146 goto out;
147 len = simple_read_from_buffer(buf, count, ppos, kbuf, len);
148
149 out:
150 kfree(kbuf);
151 return len;
152 }
153
free_schemes_arr(struct damos ** schemes,ssize_t nr_schemes)154 static void free_schemes_arr(struct damos **schemes, ssize_t nr_schemes)
155 {
156 ssize_t i;
157
158 for (i = 0; i < nr_schemes; i++)
159 kfree(schemes[i]);
160 kfree(schemes);
161 }
162
damos_action_valid(int action)163 static bool damos_action_valid(int action)
164 {
165 switch (action) {
166 case DAMOS_WILLNEED:
167 case DAMOS_COLD:
168 case DAMOS_PAGEOUT:
169 case DAMOS_HUGEPAGE:
170 case DAMOS_NOHUGEPAGE:
171 case DAMOS_STAT:
172 return true;
173 default:
174 return false;
175 }
176 }
177
178 /*
179 * Converts a string into an array of struct damos pointers
180 *
181 * Returns an array of struct damos pointers that converted if the conversion
182 * success, or NULL otherwise.
183 */
str_to_schemes(const char * str,ssize_t len,ssize_t * nr_schemes)184 static struct damos **str_to_schemes(const char *str, ssize_t len,
185 ssize_t *nr_schemes)
186 {
187 struct damos *scheme, **schemes;
188 const int max_nr_schemes = 256;
189 int pos = 0, parsed, ret;
190 unsigned long min_sz, max_sz;
191 unsigned int min_nr_a, max_nr_a, min_age, max_age;
192 unsigned int action;
193
194 schemes = kmalloc_array(max_nr_schemes, sizeof(scheme),
195 GFP_KERNEL);
196 if (!schemes)
197 return NULL;
198
199 *nr_schemes = 0;
200 while (pos < len && *nr_schemes < max_nr_schemes) {
201 struct damos_quota quota = {};
202 struct damos_watermarks wmarks;
203
204 ret = sscanf(&str[pos],
205 "%lu %lu %u %u %u %u %u %lu %lu %lu %u %u %u %u %lu %lu %lu %lu%n",
206 &min_sz, &max_sz, &min_nr_a, &max_nr_a,
207 &min_age, &max_age, &action, "a.ms,
208 "a.sz, "a.reset_interval,
209 "a.weight_sz, "a.weight_nr_accesses,
210 "a.weight_age, &wmarks.metric,
211 &wmarks.interval, &wmarks.high, &wmarks.mid,
212 &wmarks.low, &parsed);
213 if (ret != 18)
214 break;
215 if (!damos_action_valid(action))
216 goto fail;
217
218 if (min_sz > max_sz || min_nr_a > max_nr_a || min_age > max_age)
219 goto fail;
220
221 if (wmarks.high < wmarks.mid || wmarks.high < wmarks.low ||
222 wmarks.mid < wmarks.low)
223 goto fail;
224
225 pos += parsed;
226 scheme = damon_new_scheme(min_sz, max_sz, min_nr_a, max_nr_a,
227 min_age, max_age, action, "a, &wmarks);
228 if (!scheme)
229 goto fail;
230
231 schemes[*nr_schemes] = scheme;
232 *nr_schemes += 1;
233 }
234 return schemes;
235 fail:
236 free_schemes_arr(schemes, *nr_schemes);
237 return NULL;
238 }
239
dbgfs_schemes_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)240 static ssize_t dbgfs_schemes_write(struct file *file, const char __user *buf,
241 size_t count, loff_t *ppos)
242 {
243 struct damon_ctx *ctx = file->private_data;
244 char *kbuf;
245 struct damos **schemes;
246 ssize_t nr_schemes = 0, ret;
247
248 kbuf = user_input_str(buf, count, ppos);
249 if (IS_ERR(kbuf))
250 return PTR_ERR(kbuf);
251
252 schemes = str_to_schemes(kbuf, count, &nr_schemes);
253 if (!schemes) {
254 ret = -EINVAL;
255 goto out;
256 }
257
258 mutex_lock(&ctx->kdamond_lock);
259 if (ctx->kdamond) {
260 ret = -EBUSY;
261 goto unlock_out;
262 }
263
264 ret = damon_set_schemes(ctx, schemes, nr_schemes);
265 if (!ret) {
266 ret = count;
267 nr_schemes = 0;
268 }
269
270 unlock_out:
271 mutex_unlock(&ctx->kdamond_lock);
272 free_schemes_arr(schemes, nr_schemes);
273 out:
274 kfree(kbuf);
275 return ret;
276 }
277
targetid_is_pid(const struct damon_ctx * ctx)278 static inline bool targetid_is_pid(const struct damon_ctx *ctx)
279 {
280 return ctx->primitive.target_valid == damon_va_target_valid;
281 }
282
sprint_target_ids(struct damon_ctx * ctx,char * buf,ssize_t len)283 static ssize_t sprint_target_ids(struct damon_ctx *ctx, char *buf, ssize_t len)
284 {
285 struct damon_target *t;
286 unsigned long id;
287 int written = 0;
288 int rc;
289
290 damon_for_each_target(t, ctx) {
291 id = t->id;
292 if (targetid_is_pid(ctx))
293 /* Show pid numbers to debugfs users */
294 id = (unsigned long)pid_vnr((struct pid *)id);
295
296 rc = scnprintf(&buf[written], len - written, "%lu ", id);
297 if (!rc)
298 return -ENOMEM;
299 written += rc;
300 }
301 if (written)
302 written -= 1;
303 written += scnprintf(&buf[written], len - written, "\n");
304 return written;
305 }
306
dbgfs_target_ids_read(struct file * file,char __user * buf,size_t count,loff_t * ppos)307 static ssize_t dbgfs_target_ids_read(struct file *file,
308 char __user *buf, size_t count, loff_t *ppos)
309 {
310 struct damon_ctx *ctx = file->private_data;
311 ssize_t len;
312 char ids_buf[320];
313
314 mutex_lock(&ctx->kdamond_lock);
315 len = sprint_target_ids(ctx, ids_buf, 320);
316 mutex_unlock(&ctx->kdamond_lock);
317 if (len < 0)
318 return len;
319
320 return simple_read_from_buffer(buf, count, ppos, ids_buf, len);
321 }
322
323 /*
324 * Converts a string into an array of unsigned long integers
325 *
326 * Returns an array of unsigned long integers if the conversion success, or
327 * NULL otherwise.
328 */
str_to_target_ids(const char * str,ssize_t len,ssize_t * nr_ids)329 static unsigned long *str_to_target_ids(const char *str, ssize_t len,
330 ssize_t *nr_ids)
331 {
332 unsigned long *ids;
333 const int max_nr_ids = 32;
334 unsigned long id;
335 int pos = 0, parsed, ret;
336
337 *nr_ids = 0;
338 ids = kmalloc_array(max_nr_ids, sizeof(id), GFP_KERNEL);
339 if (!ids)
340 return NULL;
341 while (*nr_ids < max_nr_ids && pos < len) {
342 ret = sscanf(&str[pos], "%lu%n", &id, &parsed);
343 pos += parsed;
344 if (ret != 1)
345 break;
346 ids[*nr_ids] = id;
347 *nr_ids += 1;
348 }
349
350 return ids;
351 }
352
dbgfs_put_pids(unsigned long * ids,int nr_ids)353 static void dbgfs_put_pids(unsigned long *ids, int nr_ids)
354 {
355 int i;
356
357 for (i = 0; i < nr_ids; i++)
358 put_pid((struct pid *)ids[i]);
359 }
360
dbgfs_target_ids_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)361 static ssize_t dbgfs_target_ids_write(struct file *file,
362 const char __user *buf, size_t count, loff_t *ppos)
363 {
364 struct damon_ctx *ctx = file->private_data;
365 struct damon_target *t, *next_t;
366 bool id_is_pid = true;
367 char *kbuf;
368 unsigned long *targets;
369 ssize_t nr_targets;
370 ssize_t ret;
371 int i;
372
373 kbuf = user_input_str(buf, count, ppos);
374 if (IS_ERR(kbuf))
375 return PTR_ERR(kbuf);
376
377 if (!strncmp(kbuf, "paddr\n", count)) {
378 id_is_pid = false;
379 /* target id is meaningless here, but we set it just for fun */
380 scnprintf(kbuf, count, "42 ");
381 }
382
383 targets = str_to_target_ids(kbuf, count, &nr_targets);
384 if (!targets) {
385 ret = -ENOMEM;
386 goto out;
387 }
388
389 if (id_is_pid) {
390 for (i = 0; i < nr_targets; i++) {
391 targets[i] = (unsigned long)find_get_pid(
392 (int)targets[i]);
393 if (!targets[i]) {
394 dbgfs_put_pids(targets, i);
395 ret = -EINVAL;
396 goto free_targets_out;
397 }
398 }
399 }
400
401 mutex_lock(&ctx->kdamond_lock);
402 if (ctx->kdamond) {
403 if (id_is_pid)
404 dbgfs_put_pids(targets, nr_targets);
405 ret = -EBUSY;
406 goto unlock_out;
407 }
408
409 /* remove previously set targets */
410 damon_for_each_target_safe(t, next_t, ctx) {
411 if (targetid_is_pid(ctx))
412 put_pid((struct pid *)t->id);
413 damon_destroy_target(t);
414 }
415
416 /* Configure the context for the address space type */
417 if (id_is_pid)
418 damon_va_set_primitives(ctx);
419 else
420 damon_pa_set_primitives(ctx);
421
422 ret = damon_set_targets(ctx, targets, nr_targets);
423 if (ret) {
424 if (id_is_pid)
425 dbgfs_put_pids(targets, nr_targets);
426 } else {
427 ret = count;
428 }
429
430 unlock_out:
431 mutex_unlock(&ctx->kdamond_lock);
432 free_targets_out:
433 kfree(targets);
434 out:
435 kfree(kbuf);
436 return ret;
437 }
438
sprint_init_regions(struct damon_ctx * c,char * buf,ssize_t len)439 static ssize_t sprint_init_regions(struct damon_ctx *c, char *buf, ssize_t len)
440 {
441 struct damon_target *t;
442 struct damon_region *r;
443 int written = 0;
444 int rc;
445
446 damon_for_each_target(t, c) {
447 damon_for_each_region(r, t) {
448 rc = scnprintf(&buf[written], len - written,
449 "%lu %lu %lu\n",
450 t->id, r->ar.start, r->ar.end);
451 if (!rc)
452 return -ENOMEM;
453 written += rc;
454 }
455 }
456 return written;
457 }
458
dbgfs_init_regions_read(struct file * file,char __user * buf,size_t count,loff_t * ppos)459 static ssize_t dbgfs_init_regions_read(struct file *file, char __user *buf,
460 size_t count, loff_t *ppos)
461 {
462 struct damon_ctx *ctx = file->private_data;
463 char *kbuf;
464 ssize_t len;
465
466 kbuf = kmalloc(count, GFP_KERNEL | __GFP_NOWARN);
467 if (!kbuf)
468 return -ENOMEM;
469
470 mutex_lock(&ctx->kdamond_lock);
471 if (ctx->kdamond) {
472 mutex_unlock(&ctx->kdamond_lock);
473 len = -EBUSY;
474 goto out;
475 }
476
477 len = sprint_init_regions(ctx, kbuf, count);
478 mutex_unlock(&ctx->kdamond_lock);
479 if (len < 0)
480 goto out;
481 len = simple_read_from_buffer(buf, count, ppos, kbuf, len);
482
483 out:
484 kfree(kbuf);
485 return len;
486 }
487
add_init_region(struct damon_ctx * c,unsigned long target_id,struct damon_addr_range * ar)488 static int add_init_region(struct damon_ctx *c,
489 unsigned long target_id, struct damon_addr_range *ar)
490 {
491 struct damon_target *t;
492 struct damon_region *r, *prev;
493 unsigned long id;
494 int rc = -EINVAL;
495
496 if (ar->start >= ar->end)
497 return -EINVAL;
498
499 damon_for_each_target(t, c) {
500 id = t->id;
501 if (targetid_is_pid(c))
502 id = (unsigned long)pid_vnr((struct pid *)id);
503 if (id == target_id) {
504 r = damon_new_region(ar->start, ar->end);
505 if (!r)
506 return -ENOMEM;
507 damon_add_region(r, t);
508 if (damon_nr_regions(t) > 1) {
509 prev = damon_prev_region(r);
510 if (prev->ar.end > r->ar.start) {
511 damon_destroy_region(r, t);
512 return -EINVAL;
513 }
514 }
515 rc = 0;
516 }
517 }
518 return rc;
519 }
520
set_init_regions(struct damon_ctx * c,const char * str,ssize_t len)521 static int set_init_regions(struct damon_ctx *c, const char *str, ssize_t len)
522 {
523 struct damon_target *t;
524 struct damon_region *r, *next;
525 int pos = 0, parsed, ret;
526 unsigned long target_id;
527 struct damon_addr_range ar;
528 int err;
529
530 damon_for_each_target(t, c) {
531 damon_for_each_region_safe(r, next, t)
532 damon_destroy_region(r, t);
533 }
534
535 while (pos < len) {
536 ret = sscanf(&str[pos], "%lu %lu %lu%n",
537 &target_id, &ar.start, &ar.end, &parsed);
538 if (ret != 3)
539 break;
540 err = add_init_region(c, target_id, &ar);
541 if (err)
542 goto fail;
543 pos += parsed;
544 }
545
546 return 0;
547
548 fail:
549 damon_for_each_target(t, c) {
550 damon_for_each_region_safe(r, next, t)
551 damon_destroy_region(r, t);
552 }
553 return err;
554 }
555
dbgfs_init_regions_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)556 static ssize_t dbgfs_init_regions_write(struct file *file,
557 const char __user *buf, size_t count,
558 loff_t *ppos)
559 {
560 struct damon_ctx *ctx = file->private_data;
561 char *kbuf;
562 ssize_t ret = count;
563 int err;
564
565 kbuf = user_input_str(buf, count, ppos);
566 if (IS_ERR(kbuf))
567 return PTR_ERR(kbuf);
568
569 mutex_lock(&ctx->kdamond_lock);
570 if (ctx->kdamond) {
571 ret = -EBUSY;
572 goto unlock_out;
573 }
574
575 err = set_init_regions(ctx, kbuf, ret);
576 if (err)
577 ret = err;
578
579 unlock_out:
580 mutex_unlock(&ctx->kdamond_lock);
581 kfree(kbuf);
582 return ret;
583 }
584
dbgfs_kdamond_pid_read(struct file * file,char __user * buf,size_t count,loff_t * ppos)585 static ssize_t dbgfs_kdamond_pid_read(struct file *file,
586 char __user *buf, size_t count, loff_t *ppos)
587 {
588 struct damon_ctx *ctx = file->private_data;
589 char *kbuf;
590 ssize_t len;
591
592 kbuf = kmalloc(count, GFP_KERNEL | __GFP_NOWARN);
593 if (!kbuf)
594 return -ENOMEM;
595
596 mutex_lock(&ctx->kdamond_lock);
597 if (ctx->kdamond)
598 len = scnprintf(kbuf, count, "%d\n", ctx->kdamond->pid);
599 else
600 len = scnprintf(kbuf, count, "none\n");
601 mutex_unlock(&ctx->kdamond_lock);
602 if (!len)
603 goto out;
604 len = simple_read_from_buffer(buf, count, ppos, kbuf, len);
605
606 out:
607 kfree(kbuf);
608 return len;
609 }
610
damon_dbgfs_open(struct inode * inode,struct file * file)611 static int damon_dbgfs_open(struct inode *inode, struct file *file)
612 {
613 file->private_data = inode->i_private;
614
615 return nonseekable_open(inode, file);
616 }
617
618 static const struct file_operations attrs_fops = {
619 .open = damon_dbgfs_open,
620 .read = dbgfs_attrs_read,
621 .write = dbgfs_attrs_write,
622 };
623
624 static const struct file_operations schemes_fops = {
625 .open = damon_dbgfs_open,
626 .read = dbgfs_schemes_read,
627 .write = dbgfs_schemes_write,
628 };
629
630 static const struct file_operations target_ids_fops = {
631 .open = damon_dbgfs_open,
632 .read = dbgfs_target_ids_read,
633 .write = dbgfs_target_ids_write,
634 };
635
636 static const struct file_operations init_regions_fops = {
637 .open = damon_dbgfs_open,
638 .read = dbgfs_init_regions_read,
639 .write = dbgfs_init_regions_write,
640 };
641
642 static const struct file_operations kdamond_pid_fops = {
643 .open = damon_dbgfs_open,
644 .read = dbgfs_kdamond_pid_read,
645 };
646
dbgfs_fill_ctx_dir(struct dentry * dir,struct damon_ctx * ctx)647 static void dbgfs_fill_ctx_dir(struct dentry *dir, struct damon_ctx *ctx)
648 {
649 const char * const file_names[] = {"attrs", "schemes", "target_ids",
650 "init_regions", "kdamond_pid"};
651 const struct file_operations *fops[] = {&attrs_fops, &schemes_fops,
652 &target_ids_fops, &init_regions_fops, &kdamond_pid_fops};
653 int i;
654
655 for (i = 0; i < ARRAY_SIZE(file_names); i++)
656 debugfs_create_file(file_names[i], 0600, dir, ctx, fops[i]);
657 }
658
dbgfs_before_terminate(struct damon_ctx * ctx)659 static void dbgfs_before_terminate(struct damon_ctx *ctx)
660 {
661 struct damon_target *t, *next;
662
663 if (!targetid_is_pid(ctx))
664 return;
665
666 mutex_lock(&ctx->kdamond_lock);
667 damon_for_each_target_safe(t, next, ctx) {
668 put_pid((struct pid *)t->id);
669 damon_destroy_target(t);
670 }
671 mutex_unlock(&ctx->kdamond_lock);
672 }
673
dbgfs_new_ctx(void)674 static struct damon_ctx *dbgfs_new_ctx(void)
675 {
676 struct damon_ctx *ctx;
677
678 ctx = damon_new_ctx();
679 if (!ctx)
680 return NULL;
681
682 damon_va_set_primitives(ctx);
683 ctx->callback.before_terminate = dbgfs_before_terminate;
684 return ctx;
685 }
686
dbgfs_destroy_ctx(struct damon_ctx * ctx)687 static void dbgfs_destroy_ctx(struct damon_ctx *ctx)
688 {
689 damon_destroy_ctx(ctx);
690 }
691
692 /*
693 * Make a context of @name and create a debugfs directory for it.
694 *
695 * This function should be called while holding damon_dbgfs_lock.
696 *
697 * Returns 0 on success, negative error code otherwise.
698 */
dbgfs_mk_context(char * name)699 static int dbgfs_mk_context(char *name)
700 {
701 struct dentry *root, **new_dirs, *new_dir;
702 struct damon_ctx **new_ctxs, *new_ctx;
703
704 if (damon_nr_running_ctxs())
705 return -EBUSY;
706
707 new_ctxs = krealloc(dbgfs_ctxs, sizeof(*dbgfs_ctxs) *
708 (dbgfs_nr_ctxs + 1), GFP_KERNEL);
709 if (!new_ctxs)
710 return -ENOMEM;
711 dbgfs_ctxs = new_ctxs;
712
713 new_dirs = krealloc(dbgfs_dirs, sizeof(*dbgfs_dirs) *
714 (dbgfs_nr_ctxs + 1), GFP_KERNEL);
715 if (!new_dirs)
716 return -ENOMEM;
717 dbgfs_dirs = new_dirs;
718
719 root = dbgfs_dirs[0];
720 if (!root)
721 return -ENOENT;
722
723 new_dir = debugfs_create_dir(name, root);
724 /* Below check is required for a potential duplicated name case */
725 if (IS_ERR(new_dir))
726 return PTR_ERR(new_dir);
727 dbgfs_dirs[dbgfs_nr_ctxs] = new_dir;
728
729 new_ctx = dbgfs_new_ctx();
730 if (!new_ctx) {
731 debugfs_remove(new_dir);
732 dbgfs_dirs[dbgfs_nr_ctxs] = NULL;
733 return -ENOMEM;
734 }
735
736 dbgfs_ctxs[dbgfs_nr_ctxs] = new_ctx;
737 dbgfs_fill_ctx_dir(dbgfs_dirs[dbgfs_nr_ctxs],
738 dbgfs_ctxs[dbgfs_nr_ctxs]);
739 dbgfs_nr_ctxs++;
740
741 return 0;
742 }
743
dbgfs_mk_context_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)744 static ssize_t dbgfs_mk_context_write(struct file *file,
745 const char __user *buf, size_t count, loff_t *ppos)
746 {
747 char *kbuf;
748 char *ctx_name;
749 ssize_t ret;
750
751 kbuf = user_input_str(buf, count, ppos);
752 if (IS_ERR(kbuf))
753 return PTR_ERR(kbuf);
754 ctx_name = kmalloc(count + 1, GFP_KERNEL);
755 if (!ctx_name) {
756 kfree(kbuf);
757 return -ENOMEM;
758 }
759
760 /* Trim white space */
761 if (sscanf(kbuf, "%s", ctx_name) != 1) {
762 ret = -EINVAL;
763 goto out;
764 }
765
766 mutex_lock(&damon_dbgfs_lock);
767 ret = dbgfs_mk_context(ctx_name);
768 if (!ret)
769 ret = count;
770 mutex_unlock(&damon_dbgfs_lock);
771
772 out:
773 kfree(kbuf);
774 kfree(ctx_name);
775 return ret;
776 }
777
778 /*
779 * Remove a context of @name and its debugfs directory.
780 *
781 * This function should be called while holding damon_dbgfs_lock.
782 *
783 * Return 0 on success, negative error code otherwise.
784 */
dbgfs_rm_context(char * name)785 static int dbgfs_rm_context(char *name)
786 {
787 struct dentry *root, *dir, **new_dirs;
788 struct inode *inode;
789 struct damon_ctx **new_ctxs;
790 int i, j;
791 int ret = 0;
792
793 if (damon_nr_running_ctxs())
794 return -EBUSY;
795
796 root = dbgfs_dirs[0];
797 if (!root)
798 return -ENOENT;
799
800 dir = debugfs_lookup(name, root);
801 if (!dir)
802 return -ENOENT;
803
804 inode = d_inode(dir);
805 if (!S_ISDIR(inode->i_mode)) {
806 ret = -EINVAL;
807 goto out_dput;
808 }
809
810 new_dirs = kmalloc_array(dbgfs_nr_ctxs - 1, sizeof(*dbgfs_dirs),
811 GFP_KERNEL);
812 if (!new_dirs) {
813 ret = -ENOMEM;
814 goto out_dput;
815 }
816
817 new_ctxs = kmalloc_array(dbgfs_nr_ctxs - 1, sizeof(*dbgfs_ctxs),
818 GFP_KERNEL);
819 if (!new_ctxs) {
820 ret = -ENOMEM;
821 goto out_new_dirs;
822 }
823
824 for (i = 0, j = 0; i < dbgfs_nr_ctxs; i++) {
825 if (dbgfs_dirs[i] == dir) {
826 debugfs_remove(dbgfs_dirs[i]);
827 dbgfs_destroy_ctx(dbgfs_ctxs[i]);
828 continue;
829 }
830 new_dirs[j] = dbgfs_dirs[i];
831 new_ctxs[j++] = dbgfs_ctxs[i];
832 }
833
834 kfree(dbgfs_dirs);
835 kfree(dbgfs_ctxs);
836
837 dbgfs_dirs = new_dirs;
838 dbgfs_ctxs = new_ctxs;
839 dbgfs_nr_ctxs--;
840
841 goto out_dput;
842
843 out_new_dirs:
844 kfree(new_dirs);
845 out_dput:
846 dput(dir);
847 return ret;
848 }
849
dbgfs_rm_context_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)850 static ssize_t dbgfs_rm_context_write(struct file *file,
851 const char __user *buf, size_t count, loff_t *ppos)
852 {
853 char *kbuf;
854 ssize_t ret;
855 char *ctx_name;
856
857 kbuf = user_input_str(buf, count, ppos);
858 if (IS_ERR(kbuf))
859 return PTR_ERR(kbuf);
860 ctx_name = kmalloc(count + 1, GFP_KERNEL);
861 if (!ctx_name) {
862 kfree(kbuf);
863 return -ENOMEM;
864 }
865
866 /* Trim white space */
867 if (sscanf(kbuf, "%s", ctx_name) != 1) {
868 ret = -EINVAL;
869 goto out;
870 }
871
872 mutex_lock(&damon_dbgfs_lock);
873 ret = dbgfs_rm_context(ctx_name);
874 if (!ret)
875 ret = count;
876 mutex_unlock(&damon_dbgfs_lock);
877
878 out:
879 kfree(kbuf);
880 kfree(ctx_name);
881 return ret;
882 }
883
dbgfs_monitor_on_read(struct file * file,char __user * buf,size_t count,loff_t * ppos)884 static ssize_t dbgfs_monitor_on_read(struct file *file,
885 char __user *buf, size_t count, loff_t *ppos)
886 {
887 char monitor_on_buf[5];
888 bool monitor_on = damon_nr_running_ctxs() != 0;
889 int len;
890
891 len = scnprintf(monitor_on_buf, 5, monitor_on ? "on\n" : "off\n");
892
893 return simple_read_from_buffer(buf, count, ppos, monitor_on_buf, len);
894 }
895
dbgfs_monitor_on_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)896 static ssize_t dbgfs_monitor_on_write(struct file *file,
897 const char __user *buf, size_t count, loff_t *ppos)
898 {
899 ssize_t ret;
900 char *kbuf;
901
902 kbuf = user_input_str(buf, count, ppos);
903 if (IS_ERR(kbuf))
904 return PTR_ERR(kbuf);
905
906 /* Remove white space */
907 if (sscanf(kbuf, "%s", kbuf) != 1) {
908 kfree(kbuf);
909 return -EINVAL;
910 }
911
912 mutex_lock(&damon_dbgfs_lock);
913 if (!strncmp(kbuf, "on", count)) {
914 int i;
915
916 for (i = 0; i < dbgfs_nr_ctxs; i++) {
917 if (damon_targets_empty(dbgfs_ctxs[i])) {
918 kfree(kbuf);
919 mutex_unlock(&damon_dbgfs_lock);
920 return -EINVAL;
921 }
922 }
923 ret = damon_start(dbgfs_ctxs, dbgfs_nr_ctxs);
924 } else if (!strncmp(kbuf, "off", count)) {
925 ret = damon_stop(dbgfs_ctxs, dbgfs_nr_ctxs);
926 } else {
927 ret = -EINVAL;
928 }
929 mutex_unlock(&damon_dbgfs_lock);
930
931 if (!ret)
932 ret = count;
933 kfree(kbuf);
934 return ret;
935 }
936
937 static const struct file_operations mk_contexts_fops = {
938 .write = dbgfs_mk_context_write,
939 };
940
941 static const struct file_operations rm_contexts_fops = {
942 .write = dbgfs_rm_context_write,
943 };
944
945 static const struct file_operations monitor_on_fops = {
946 .read = dbgfs_monitor_on_read,
947 .write = dbgfs_monitor_on_write,
948 };
949
__damon_dbgfs_init(void)950 static int __init __damon_dbgfs_init(void)
951 {
952 struct dentry *dbgfs_root;
953 const char * const file_names[] = {"mk_contexts", "rm_contexts",
954 "monitor_on"};
955 const struct file_operations *fops[] = {&mk_contexts_fops,
956 &rm_contexts_fops, &monitor_on_fops};
957 int i;
958
959 dbgfs_root = debugfs_create_dir("damon", NULL);
960
961 for (i = 0; i < ARRAY_SIZE(file_names); i++)
962 debugfs_create_file(file_names[i], 0600, dbgfs_root, NULL,
963 fops[i]);
964 dbgfs_fill_ctx_dir(dbgfs_root, dbgfs_ctxs[0]);
965
966 dbgfs_dirs = kmalloc_array(1, sizeof(dbgfs_root), GFP_KERNEL);
967 if (!dbgfs_dirs) {
968 debugfs_remove(dbgfs_root);
969 return -ENOMEM;
970 }
971 dbgfs_dirs[0] = dbgfs_root;
972
973 return 0;
974 }
975
976 /*
977 * Functions for the initialization
978 */
979
damon_dbgfs_init(void)980 static int __init damon_dbgfs_init(void)
981 {
982 int rc = -ENOMEM;
983
984 mutex_lock(&damon_dbgfs_lock);
985 dbgfs_ctxs = kmalloc(sizeof(*dbgfs_ctxs), GFP_KERNEL);
986 if (!dbgfs_ctxs)
987 goto out;
988 dbgfs_ctxs[0] = dbgfs_new_ctx();
989 if (!dbgfs_ctxs[0]) {
990 kfree(dbgfs_ctxs);
991 goto out;
992 }
993 dbgfs_nr_ctxs = 1;
994
995 rc = __damon_dbgfs_init();
996 if (rc) {
997 kfree(dbgfs_ctxs[0]);
998 kfree(dbgfs_ctxs);
999 pr_err("%s: dbgfs init failed\n", __func__);
1000 }
1001
1002 out:
1003 mutex_unlock(&damon_dbgfs_lock);
1004 return rc;
1005 }
1006
1007 module_init(damon_dbgfs_init);
1008
1009 #include "dbgfs-test.h"
1010