| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 1 | /* | 
 | 2 |  * Copyright (C) 2009  Matt Fleming | 
 | 3 |  * | 
 | 4 |  * Based, in part, on kernel/time/clocksource.c. | 
 | 5 |  * | 
 | 6 |  * This file provides arbitration code for stack unwinders. | 
 | 7 |  * | 
 | 8 |  * Multiple stack unwinders can be available on a system, usually with | 
 | 9 |  * the most accurate unwinder being the currently active one. | 
 | 10 |  */ | 
 | 11 | #include <linux/errno.h> | 
 | 12 | #include <linux/list.h> | 
 | 13 | #include <linux/spinlock.h> | 
| Paul Mundt | 4ab8f24 | 2009-08-22 03:43:15 +0900 | [diff] [blame] | 14 | #include <linux/module.h> | 
| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 15 | #include <asm/unwinder.h> | 
 | 16 | #include <asm/atomic.h> | 
 | 17 |  | 
 | 18 | /* | 
 | 19 |  * This is the most basic stack unwinder an architecture can | 
 | 20 |  * provide. For architectures without reliable frame pointers, e.g. | 
 | 21 |  * RISC CPUs, it can be implemented by looking through the stack for | 
 | 22 |  * addresses that lie within the kernel text section. | 
 | 23 |  * | 
 | 24 |  * Other CPUs, e.g. x86, can use their frame pointer register to | 
 | 25 |  * construct more accurate stack traces. | 
 | 26 |  */ | 
 | 27 | static struct list_head unwinder_list; | 
 | 28 | static struct unwinder stack_reader = { | 
 | 29 | 	.name = "stack-reader", | 
 | 30 | 	.dump = stack_reader_dump, | 
 | 31 | 	.rating = 50, | 
 | 32 | 	.list = { | 
 | 33 | 		.next = &unwinder_list, | 
 | 34 | 		.prev = &unwinder_list, | 
 | 35 | 	}, | 
 | 36 | }; | 
 | 37 |  | 
 | 38 | /* | 
 | 39 |  * "curr_unwinder" points to the stack unwinder currently in use. This | 
 | 40 |  * is the unwinder with the highest rating. | 
 | 41 |  * | 
 | 42 |  * "unwinder_list" is a linked-list of all available unwinders, sorted | 
 | 43 |  * by rating. | 
 | 44 |  * | 
 | 45 |  * All modifications of "curr_unwinder" and "unwinder_list" must be | 
 | 46 |  * performed whilst holding "unwinder_lock". | 
 | 47 |  */ | 
 | 48 | static struct unwinder *curr_unwinder = &stack_reader; | 
 | 49 |  | 
 | 50 | static struct list_head unwinder_list = { | 
 | 51 | 	.next = &stack_reader.list, | 
 | 52 | 	.prev = &stack_reader.list, | 
 | 53 | }; | 
 | 54 |  | 
 | 55 | static DEFINE_SPINLOCK(unwinder_lock); | 
 | 56 |  | 
| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 57 | /** | 
 | 58 |  * select_unwinder - Select the best registered stack unwinder. | 
 | 59 |  * | 
 | 60 |  * Private function. Must hold unwinder_lock when called. | 
 | 61 |  * | 
 | 62 |  * Select the stack unwinder with the best rating. This is useful for | 
 | 63 |  * setting up curr_unwinder. | 
 | 64 |  */ | 
 | 65 | static struct unwinder *select_unwinder(void) | 
 | 66 | { | 
 | 67 | 	struct unwinder *best; | 
 | 68 |  | 
 | 69 | 	if (list_empty(&unwinder_list)) | 
 | 70 | 		return NULL; | 
 | 71 |  | 
 | 72 | 	best = list_entry(unwinder_list.next, struct unwinder, list); | 
 | 73 | 	if (best == curr_unwinder) | 
 | 74 | 		return NULL; | 
 | 75 |  | 
 | 76 | 	return best; | 
 | 77 | } | 
 | 78 |  | 
 | 79 | /* | 
 | 80 |  * Enqueue the stack unwinder sorted by rating. | 
 | 81 |  */ | 
 | 82 | static int unwinder_enqueue(struct unwinder *ops) | 
 | 83 | { | 
 | 84 | 	struct list_head *tmp, *entry = &unwinder_list; | 
 | 85 |  | 
 | 86 | 	list_for_each(tmp, &unwinder_list) { | 
 | 87 | 		struct unwinder *o; | 
 | 88 |  | 
 | 89 | 		o = list_entry(tmp, struct unwinder, list); | 
 | 90 | 		if (o == ops) | 
 | 91 | 			return -EBUSY; | 
 | 92 | 		/* Keep track of the place, where to insert */ | 
 | 93 | 		if (o->rating >= ops->rating) | 
 | 94 | 			entry = tmp; | 
 | 95 | 	} | 
 | 96 | 	list_add(&ops->list, entry); | 
 | 97 |  | 
 | 98 | 	return 0; | 
 | 99 | } | 
 | 100 |  | 
 | 101 | /** | 
 | 102 |  * unwinder_register - Used to install new stack unwinder | 
 | 103 |  * @u: unwinder to be registered | 
 | 104 |  * | 
 | 105 |  * Install the new stack unwinder on the unwinder list, which is sorted | 
 | 106 |  * by rating. | 
 | 107 |  * | 
 | 108 |  * Returns -EBUSY if registration fails, zero otherwise. | 
 | 109 |  */ | 
 | 110 | int unwinder_register(struct unwinder *u) | 
 | 111 | { | 
 | 112 | 	unsigned long flags; | 
 | 113 | 	int ret; | 
 | 114 |  | 
 | 115 | 	spin_lock_irqsave(&unwinder_lock, flags); | 
 | 116 | 	ret = unwinder_enqueue(u); | 
 | 117 | 	if (!ret) | 
 | 118 | 		curr_unwinder = select_unwinder(); | 
 | 119 | 	spin_unlock_irqrestore(&unwinder_lock, flags); | 
 | 120 |  | 
 | 121 | 	return ret; | 
 | 122 | } | 
 | 123 |  | 
| Matt Fleming | b344e24 | 2009-08-16 21:54:48 +0100 | [diff] [blame] | 124 | int unwinder_faulted = 0; | 
 | 125 |  | 
| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 126 | /* | 
 | 127 |  * Unwind the call stack and pass information to the stacktrace_ops | 
 | 128 |  * functions. Also handle the case where we need to switch to a new | 
 | 129 |  * stack dumper because the current one faulted unexpectedly. | 
 | 130 |  */ | 
 | 131 | void unwind_stack(struct task_struct *task, struct pt_regs *regs, | 
 | 132 | 		  unsigned long *sp, const struct stacktrace_ops *ops, | 
 | 133 | 		  void *data) | 
 | 134 | { | 
 | 135 | 	unsigned long flags; | 
 | 136 |  | 
 | 137 | 	/* | 
 | 138 | 	 * The problem with unwinders with high ratings is that they are | 
 | 139 | 	 * inherently more complicated than the simple ones with lower | 
 | 140 | 	 * ratings. We are therefore more likely to fault in the | 
 | 141 | 	 * complicated ones, e.g. hitting BUG()s. If we fault in the | 
 | 142 | 	 * code for the current stack unwinder we try to downgrade to | 
 | 143 | 	 * one with a lower rating. | 
 | 144 | 	 * | 
 | 145 | 	 * Hopefully this will give us a semi-reliable stacktrace so we | 
 | 146 | 	 * can diagnose why curr_unwinder->dump() faulted. | 
 | 147 | 	 */ | 
| Matt Fleming | b344e24 | 2009-08-16 21:54:48 +0100 | [diff] [blame] | 148 | 	if (unwinder_faulted) { | 
| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 149 | 		spin_lock_irqsave(&unwinder_lock, flags); | 
 | 150 |  | 
| Matt Fleming | b344e24 | 2009-08-16 21:54:48 +0100 | [diff] [blame] | 151 | 		/* Make sure no one beat us to changing the unwinder */ | 
 | 152 | 		if (unwinder_faulted && !list_is_singular(&unwinder_list)) { | 
| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 153 | 			list_del(&curr_unwinder->list); | 
 | 154 | 			curr_unwinder = select_unwinder(); | 
| Matt Fleming | b344e24 | 2009-08-16 21:54:48 +0100 | [diff] [blame] | 155 |  | 
 | 156 | 			unwinder_faulted = 0; | 
| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 157 | 		} | 
 | 158 |  | 
 | 159 | 		spin_unlock_irqrestore(&unwinder_lock, flags); | 
| Matt Fleming | bf61ad1 | 2009-08-13 19:49:03 +0900 | [diff] [blame] | 160 | 	} | 
 | 161 |  | 
 | 162 | 	curr_unwinder->dump(task, regs, sp, ops, data); | 
| Matt Fleming | b344e24 | 2009-08-16 21:54:48 +0100 | [diff] [blame] | 163 | } | 
| Paul Mundt | 4ab8f24 | 2009-08-22 03:43:15 +0900 | [diff] [blame] | 164 | EXPORT_SYMBOL_GPL(unwind_stack); |