| /* | 
 |  * mcount and friends -- ftrace stuff | 
 |  * | 
 |  * Copyright (C) 2009-2010 Analog Devices Inc. | 
 |  * Licensed under the GPL-2 or later. | 
 |  */ | 
 |  | 
 | #include <linux/linkage.h> | 
 | #include <asm/ftrace.h> | 
 |  | 
 | .text | 
 |  | 
 | #ifdef CONFIG_DYNAMIC_FTRACE | 
 |  | 
 | /* Simple stub so we can boot the kernel until runtime patching has | 
 |  * disabled all calls to this.  Then it'll be unused. | 
 |  */ | 
 | ENTRY(__mcount) | 
 | # if ANOMALY_05000371 | 
 | 	nop; nop; nop; nop; | 
 | # endif | 
 | 	rts; | 
 | ENDPROC(__mcount) | 
 |  | 
 | /* GCC will have called us before setting up the function prologue, so we | 
 |  * can clobber the normal scratch registers, but we need to make sure to | 
 |  * save/restore the registers used for argument passing (R0-R2) in case | 
 |  * the profiled function is using them.  With data registers, R3 is the | 
 |  * only one we can blow away.  With pointer registers, we have P0-P2. | 
 |  * | 
 |  * Upon entry, the RETS will point to the top of the current profiled | 
 |  * function.  And since GCC pushed the previous RETS for us, the previous | 
 |  * function will be waiting there.  mmmm pie. | 
 |  */ | 
 | ENTRY(_ftrace_caller) | 
 | # ifdef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | 
 | 	/* optional micro optimization: return if stopped */ | 
 | 	p1.l = _function_trace_stop; | 
 | 	p1.h = _function_trace_stop; | 
 | 	r3 = [p1]; | 
 | 	cc = r3 == 0; | 
 | 	if ! cc jump _ftrace_stub (bp); | 
 | # endif | 
 |  | 
 | 	/* save first/second/third function arg and the return register */ | 
 | 	[--sp] = r2; | 
 | 	[--sp] = r0; | 
 | 	[--sp] = r1; | 
 | 	[--sp] = rets; | 
 |  | 
 | 	/* function_trace_call(unsigned long ip, unsigned long parent_ip): | 
 | 	 *  ip: this point was called by ... | 
 | 	 *  parent_ip: ... this function | 
 | 	 * the ip itself will need adjusting for the mcount call | 
 | 	 */ | 
 | 	r0 = rets; | 
 | 	r1 = [sp + 16];	/* skip the 4 local regs on stack */ | 
 | 	r0 += -MCOUNT_INSN_SIZE; | 
 |  | 
 | .globl _ftrace_call | 
 | _ftrace_call: | 
 | 	call _ftrace_stub | 
 |  | 
 | # ifdef CONFIG_FUNCTION_GRAPH_TRACER | 
 | .globl _ftrace_graph_call | 
 | _ftrace_graph_call: | 
 | 	nop;	/* jump _ftrace_graph_caller; */ | 
 | # endif | 
 |  | 
 | 	/* restore state and get out of dodge */ | 
 | .Lfinish_trace: | 
 | 	rets = [sp++]; | 
 | 	r1 = [sp++]; | 
 | 	r0 = [sp++]; | 
 | 	r2 = [sp++]; | 
 |  | 
 | .globl _ftrace_stub | 
 | _ftrace_stub: | 
 | 	rts; | 
 | ENDPROC(_ftrace_caller) | 
 |  | 
 | #else | 
 |  | 
 | /* See documentation for _ftrace_caller */ | 
 | ENTRY(__mcount) | 
 | # ifdef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | 
 | 	/* optional micro optimization: return if stopped */ | 
 | 	p1.l = _function_trace_stop; | 
 | 	p1.h = _function_trace_stop; | 
 | 	r3 = [p1]; | 
 | 	cc = r3 == 0; | 
 | 	if ! cc jump _ftrace_stub (bp); | 
 | # endif | 
 |  | 
 | 	/* save third function arg early so we can do testing below */ | 
 | 	[--sp] = r2; | 
 |  | 
 | 	/* load the function pointer to the tracer */ | 
 | 	p0.l = _ftrace_trace_function; | 
 | 	p0.h = _ftrace_trace_function; | 
 | 	r3 = [p0]; | 
 |  | 
 | 	/* optional micro optimization: don't call the stub tracer */ | 
 | 	r2.l = _ftrace_stub; | 
 | 	r2.h = _ftrace_stub; | 
 | 	cc = r2 == r3; | 
 | 	if ! cc jump .Ldo_trace; | 
 |  | 
 | # ifdef CONFIG_FUNCTION_GRAPH_TRACER | 
 | 	/* if the ftrace_graph_return function pointer is not set to | 
 | 	 * the ftrace_stub entry, call prepare_ftrace_return(). | 
 | 	 */ | 
 | 	p0.l = _ftrace_graph_return; | 
 | 	p0.h = _ftrace_graph_return; | 
 | 	r3 = [p0]; | 
 | 	cc = r2 == r3; | 
 | 	if ! cc jump _ftrace_graph_caller; | 
 |  | 
 | 	/* similarly, if the ftrace_graph_entry function pointer is not | 
 | 	 * set to the ftrace_graph_entry_stub entry, ... | 
 | 	 */ | 
 | 	p0.l = _ftrace_graph_entry; | 
 | 	p0.h = _ftrace_graph_entry; | 
 | 	r2.l = _ftrace_graph_entry_stub; | 
 | 	r2.h = _ftrace_graph_entry_stub; | 
 | 	r3 = [p0]; | 
 | 	cc = r2 == r3; | 
 | 	if ! cc jump _ftrace_graph_caller; | 
 | # endif | 
 |  | 
 | 	r2 = [sp++]; | 
 | 	rts; | 
 |  | 
 | .Ldo_trace: | 
 |  | 
 | 	/* save first/second function arg and the return register */ | 
 | 	[--sp] = r0; | 
 | 	[--sp] = r1; | 
 | 	[--sp] = rets; | 
 |  | 
 | 	/* setup the tracer function */ | 
 | 	p0 = r3; | 
 |  | 
 | 	/* function_trace_call(unsigned long ip, unsigned long parent_ip): | 
 | 	 *  ip: this point was called by ... | 
 | 	 *  parent_ip: ... this function | 
 | 	 * the ip itself will need adjusting for the mcount call | 
 | 	 */ | 
 | 	r0 = rets; | 
 | 	r1 = [sp + 16];	/* skip the 4 local regs on stack */ | 
 | 	r0 += -MCOUNT_INSN_SIZE; | 
 |  | 
 | 	/* call the tracer */ | 
 | 	call (p0); | 
 |  | 
 | 	/* restore state and get out of dodge */ | 
 | .Lfinish_trace: | 
 | 	rets = [sp++]; | 
 | 	r1 = [sp++]; | 
 | 	r0 = [sp++]; | 
 | 	r2 = [sp++]; | 
 |  | 
 | .globl _ftrace_stub | 
 | _ftrace_stub: | 
 | 	rts; | 
 | ENDPROC(__mcount) | 
 |  | 
 | #endif | 
 |  | 
 | #ifdef CONFIG_FUNCTION_GRAPH_TRACER | 
 | /* The prepare_ftrace_return() function is similar to the trace function | 
 |  * except it takes a pointer to the location of the frompc.  This is so | 
 |  * the prepare_ftrace_return() can hijack it temporarily for probing | 
 |  * purposes. | 
 |  */ | 
 | ENTRY(_ftrace_graph_caller) | 
 | # ifndef CONFIG_DYNAMIC_FTRACE | 
 | 	/* save first/second function arg and the return register */ | 
 | 	[--sp] = r0; | 
 | 	[--sp] = r1; | 
 | 	[--sp] = rets; | 
 |  | 
 | 	/* prepare_ftrace_return(parent, self_addr, frame_pointer) */ | 
 | 	r0 = sp;	/* unsigned long *parent */ | 
 | 	r1 = rets;	/* unsigned long self_addr */ | 
 | # else | 
 | 	r0 = sp;	/* unsigned long *parent */ | 
 | 	r1 = [sp];	/* unsigned long self_addr */ | 
 | # endif | 
 | # ifdef CONFIG_HAVE_FUNCTION_GRAPH_FP_TEST | 
 | 	r2 = fp;	/* unsigned long frame_pointer */ | 
 | # endif | 
 | 	r0 += 16;	/* skip the 4 local regs on stack */ | 
 | 	r1 += -MCOUNT_INSN_SIZE; | 
 | 	call _prepare_ftrace_return; | 
 |  | 
 | 	jump .Lfinish_trace; | 
 | ENDPROC(_ftrace_graph_caller) | 
 |  | 
 | /* Undo the rewrite caused by ftrace_graph_caller().  The common function | 
 |  * ftrace_return_to_handler() will return the original rets so we can | 
 |  * restore it and be on our way. | 
 |  */ | 
 | ENTRY(_return_to_handler) | 
 | 	/* make sure original return values are saved */ | 
 | 	[--sp] = p0; | 
 | 	[--sp] = r0; | 
 | 	[--sp] = r1; | 
 |  | 
 | 	/* get original return address */ | 
 | # ifdef CONFIG_HAVE_FUNCTION_GRAPH_FP_TEST | 
 | 	r0 = fp;	/* Blackfin is sane, so omit this */ | 
 | # endif | 
 | 	call _ftrace_return_to_handler; | 
 | 	rets = r0; | 
 |  | 
 | 	/* anomaly 05000371 - make sure we have at least three instructions | 
 | 	 * between rets setting and the return | 
 | 	 */ | 
 | 	r1 = [sp++]; | 
 | 	r0 = [sp++]; | 
 | 	p0 = [sp++]; | 
 | 	rts; | 
 | ENDPROC(_return_to_handler) | 
 | #endif |