• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1  /*
2   * Copyright (C) 2009  Matt Fleming
3   *
4   * Based, in part, on kernel/time/clocksource.c.
5   *
6   * This file provides arbitration code for stack unwinders.
7   *
8   * Multiple stack unwinders can be available on a system, usually with
9   * the most accurate unwinder being the currently active one.
10   */
11  #include <linux/errno.h>
12  #include <linux/list.h>
13  #include <linux/spinlock.h>
14  #include <linux/module.h>
15  #include <asm/unwinder.h>
16  #include <linux/atomic.h>
17  
18  /*
19   * This is the most basic stack unwinder an architecture can
20   * provide. For architectures without reliable frame pointers, e.g.
21   * RISC CPUs, it can be implemented by looking through the stack for
22   * addresses that lie within the kernel text section.
23   *
24   * Other CPUs, e.g. x86, can use their frame pointer register to
25   * construct more accurate stack traces.
26   */
27  static struct list_head unwinder_list;
28  static struct unwinder stack_reader = {
29  	.name = "stack-reader",
30  	.dump = stack_reader_dump,
31  	.rating = 50,
32  	.list = {
33  		.next = &unwinder_list,
34  		.prev = &unwinder_list,
35  	},
36  };
37  
38  /*
39   * "curr_unwinder" points to the stack unwinder currently in use. This
40   * is the unwinder with the highest rating.
41   *
42   * "unwinder_list" is a linked-list of all available unwinders, sorted
43   * by rating.
44   *
45   * All modifications of "curr_unwinder" and "unwinder_list" must be
46   * performed whilst holding "unwinder_lock".
47   */
48  static struct unwinder *curr_unwinder = &stack_reader;
49  
50  static struct list_head unwinder_list = {
51  	.next = &stack_reader.list,
52  	.prev = &stack_reader.list,
53  };
54  
55  static DEFINE_SPINLOCK(unwinder_lock);
56  
57  /**
58   * select_unwinder - Select the best registered stack unwinder.
59   *
60   * Private function. Must hold unwinder_lock when called.
61   *
62   * Select the stack unwinder with the best rating. This is useful for
63   * setting up curr_unwinder.
64   */
select_unwinder(void)65  static struct unwinder *select_unwinder(void)
66  {
67  	struct unwinder *best;
68  
69  	if (list_empty(&unwinder_list))
70  		return NULL;
71  
72  	best = list_entry(unwinder_list.next, struct unwinder, list);
73  	if (best == curr_unwinder)
74  		return NULL;
75  
76  	return best;
77  }
78  
79  /*
80   * Enqueue the stack unwinder sorted by rating.
81   */
unwinder_enqueue(struct unwinder * ops)82  static int unwinder_enqueue(struct unwinder *ops)
83  {
84  	struct list_head *tmp, *entry = &unwinder_list;
85  
86  	list_for_each(tmp, &unwinder_list) {
87  		struct unwinder *o;
88  
89  		o = list_entry(tmp, struct unwinder, list);
90  		if (o == ops)
91  			return -EBUSY;
92  		/* Keep track of the place, where to insert */
93  		if (o->rating >= ops->rating)
94  			entry = tmp;
95  	}
96  	list_add(&ops->list, entry);
97  
98  	return 0;
99  }
100  
101  /**
102   * unwinder_register - Used to install new stack unwinder
103   * @u: unwinder to be registered
104   *
105   * Install the new stack unwinder on the unwinder list, which is sorted
106   * by rating.
107   *
108   * Returns -EBUSY if registration fails, zero otherwise.
109   */
unwinder_register(struct unwinder * u)110  int unwinder_register(struct unwinder *u)
111  {
112  	unsigned long flags;
113  	int ret;
114  
115  	spin_lock_irqsave(&unwinder_lock, flags);
116  	ret = unwinder_enqueue(u);
117  	if (!ret)
118  		curr_unwinder = select_unwinder();
119  	spin_unlock_irqrestore(&unwinder_lock, flags);
120  
121  	return ret;
122  }
123  
124  int unwinder_faulted = 0;
125  
126  /*
127   * Unwind the call stack and pass information to the stacktrace_ops
128   * functions. Also handle the case where we need to switch to a new
129   * stack dumper because the current one faulted unexpectedly.
130   */
unwind_stack(struct task_struct * task,struct pt_regs * regs,unsigned long * sp,const struct stacktrace_ops * ops,void * data)131  void unwind_stack(struct task_struct *task, struct pt_regs *regs,
132  		  unsigned long *sp, const struct stacktrace_ops *ops,
133  		  void *data)
134  {
135  	unsigned long flags;
136  
137  	/*
138  	 * The problem with unwinders with high ratings is that they are
139  	 * inherently more complicated than the simple ones with lower
140  	 * ratings. We are therefore more likely to fault in the
141  	 * complicated ones, e.g. hitting BUG()s. If we fault in the
142  	 * code for the current stack unwinder we try to downgrade to
143  	 * one with a lower rating.
144  	 *
145  	 * Hopefully this will give us a semi-reliable stacktrace so we
146  	 * can diagnose why curr_unwinder->dump() faulted.
147  	 */
148  	if (unwinder_faulted) {
149  		spin_lock_irqsave(&unwinder_lock, flags);
150  
151  		/* Make sure no one beat us to changing the unwinder */
152  		if (unwinder_faulted && !list_is_singular(&unwinder_list)) {
153  			list_del(&curr_unwinder->list);
154  			curr_unwinder = select_unwinder();
155  
156  			unwinder_faulted = 0;
157  		}
158  
159  		spin_unlock_irqrestore(&unwinder_lock, flags);
160  	}
161  
162  	curr_unwinder->dump(task, regs, sp, ops, data);
163  }
164  EXPORT_SYMBOL_GPL(unwind_stack);
165