| /* |
| * arch/sh/lib/mcount.S |
| * |
| * Copyright (C) 2008, 2009 Paul Mundt |
| * Copyright (C) 2008, 2009 Matt Fleming |
| * |
| * This file is subject to the terms and conditions of the GNU General Public |
| * License. See the file "COPYING" in the main directory of this archive |
| * for more details. |
| */ |
| #include <asm/ftrace.h> |
| #include <asm/thread_info.h> |
| #include <asm/asm-offsets.h> |
| |
| #define MCOUNT_ENTER() \ |
| mov.l r4, @-r15; \ |
| mov.l r5, @-r15; \ |
| mov.l r6, @-r15; \ |
| mov.l r7, @-r15; \ |
| sts.l pr, @-r15; \ |
| \ |
| mov.l @(20,r15),r4; \ |
| sts pr, r5 |
| |
| #define MCOUNT_LEAVE() \ |
| lds.l @r15+, pr; \ |
| mov.l @r15+, r7; \ |
| mov.l @r15+, r6; \ |
| mov.l @r15+, r5; \ |
| rts; \ |
| mov.l @r15+, r4 |
| |
| #ifdef CONFIG_STACK_DEBUG |
| /* |
| * Perform diagnostic checks on the state of the kernel stack. |
| * |
| * Check for stack overflow. If there is less than 1KB free |
| * then it has overflowed. |
| * |
| * Make sure the stack pointer contains a valid address. Valid |
| * addresses for kernel stacks are anywhere after the bss |
| * (after __bss_stop) and anywhere in init_thread_union (init_stack). |
| */ |
| #define STACK_CHECK() \ |
| mov #(THREAD_SIZE >> 10), r0; \ |
| shll8 r0; \ |
| shll2 r0; \ |
| \ |
| /* r1 = sp & (THREAD_SIZE - 1) */ \ |
| mov #-1, r1; \ |
| add r0, r1; \ |
| and r15, r1; \ |
| \ |
| mov #TI_SIZE, r3; \ |
| mov #(STACK_WARN >> 8), r2; \ |
| shll8 r2; \ |
| add r3, r2; \ |
| \ |
| /* Is the stack overflowing? */ \ |
| cmp/hi r2, r1; \ |
| bf stack_panic; \ |
| \ |
| /* If sp > __bss_stop then we're OK. */ \ |
| mov.l .L_ebss, r1; \ |
| cmp/hi r1, r15; \ |
| bt 1f; \ |
| \ |
| /* If sp < init_stack, we're not OK. */ \ |
| mov.l .L_init_thread_union, r1; \ |
| cmp/hs r1, r15; \ |
| bf stack_panic; \ |
| \ |
| /* If sp > init_stack && sp < __bss_stop, not OK. */ \ |
| add r0, r1; \ |
| cmp/hs r1, r15; \ |
| bt stack_panic; \ |
| 1: |
| #else |
| #define STACK_CHECK() |
| #endif /* CONFIG_STACK_DEBUG */ |
| |
| .align 2 |
| .globl _mcount |
| .type _mcount,@function |
| .globl mcount |
| .type mcount,@function |
| _mcount: |
| mcount: |
| STACK_CHECK() |
| |
| #ifndef CONFIG_FUNCTION_TRACER |
| rts |
| nop |
| #else |
| #ifndef CONFIG_DYNAMIC_FTRACE |
| mov.l .Lfunction_trace_stop, r0 |
| mov.l @r0, r0 |
| tst r0, r0 |
| bf ftrace_stub |
| #endif |
| |
| MCOUNT_ENTER() |
| |
| #ifdef CONFIG_DYNAMIC_FTRACE |
| .globl mcount_call |
| mcount_call: |
| mov.l .Lftrace_stub, r6 |
| #else |
| mov.l .Lftrace_trace_function, r6 |
| mov.l ftrace_stub, r7 |
| cmp/eq r6, r7 |
| bt skip_trace |
| mov.l @r6, r6 |
| #endif |
| |
| jsr @r6 |
| nop |
| |
| #ifdef CONFIG_FUNCTION_GRAPH_TRACER |
| mov.l .Lftrace_graph_return, r6 |
| mov.l .Lftrace_stub, r7 |
| cmp/eq r6, r7 |
| bt 1f |
| |
| mov.l .Lftrace_graph_caller, r0 |
| jmp @r0 |
| nop |
| |
| 1: |
| mov.l .Lftrace_graph_entry, r6 |
| mov.l .Lftrace_graph_entry_stub, r7 |
| cmp/eq r6, r7 |
| bt skip_trace |
| |
| mov.l .Lftrace_graph_caller, r0 |
| jmp @r0 |
| nop |
| |
| .align 2 |
| .Lftrace_graph_return: |
| .long ftrace_graph_return |
| .Lftrace_graph_entry: |
| .long ftrace_graph_entry |
| .Lftrace_graph_entry_stub: |
| .long ftrace_graph_entry_stub |
| .Lftrace_graph_caller: |
| .long ftrace_graph_caller |
| #endif /* CONFIG_FUNCTION_GRAPH_TRACER */ |
| |
| .globl skip_trace |
| skip_trace: |
| MCOUNT_LEAVE() |
| |
| .align 2 |
| .Lftrace_trace_function: |
| .long ftrace_trace_function |
| |
| #ifdef CONFIG_DYNAMIC_FTRACE |
| #ifdef CONFIG_FUNCTION_GRAPH_TRACER |
| /* |
| * NOTE: Do not move either ftrace_graph_call or ftrace_caller |
| * as this will affect the calculation of GRAPH_INSN_OFFSET. |
| */ |
| .globl ftrace_graph_call |
| ftrace_graph_call: |
| mov.l .Lskip_trace, r0 |
| jmp @r0 |
| nop |
| |
| .align 2 |
| .Lskip_trace: |
| .long skip_trace |
| #endif /* CONFIG_FUNCTION_GRAPH_TRACER */ |
| |
| .globl ftrace_caller |
| ftrace_caller: |
| mov.l .Lfunction_trace_stop, r0 |
| mov.l @r0, r0 |
| tst r0, r0 |
| bf ftrace_stub |
| |
| MCOUNT_ENTER() |
| |
| .globl ftrace_call |
| ftrace_call: |
| mov.l .Lftrace_stub, r6 |
| jsr @r6 |
| nop |
| |
| #ifdef CONFIG_FUNCTION_GRAPH_TRACER |
| bra ftrace_graph_call |
| nop |
| #else |
| MCOUNT_LEAVE() |
| #endif /* CONFIG_FUNCTION_GRAPH_TRACER */ |
| #endif /* CONFIG_DYNAMIC_FTRACE */ |
| |
| .align 2 |
| .Lfunction_trace_stop: |
| .long function_trace_stop |
| |
| /* |
| * NOTE: From here on the locations of the .Lftrace_stub label and |
| * ftrace_stub itself are fixed. Adding additional data here will skew |
| * the displacement for the memory table and break the block replacement. |
| * Place new labels either after the ftrace_stub body, or before |
| * ftrace_caller. You have been warned. |
| */ |
| .Lftrace_stub: |
| .long ftrace_stub |
| |
| .globl ftrace_stub |
| ftrace_stub: |
| rts |
| nop |
| |
| #ifdef CONFIG_FUNCTION_GRAPH_TRACER |
| .globl ftrace_graph_caller |
| ftrace_graph_caller: |
| mov.l 2f, r0 |
| mov.l @r0, r0 |
| tst r0, r0 |
| bt 1f |
| |
| mov.l 3f, r1 |
| jmp @r1 |
| nop |
| 1: |
| /* |
| * MCOUNT_ENTER() pushed 5 registers onto the stack, so |
| * the stack address containing our return address is |
| * r15 + 20. |
| */ |
| mov #20, r0 |
| add r15, r0 |
| mov r0, r4 |
| |
| mov.l .Lprepare_ftrace_return, r0 |
| jsr @r0 |
| nop |
| |
| MCOUNT_LEAVE() |
| |
| .align 2 |
| 2: .long function_trace_stop |
| 3: .long skip_trace |
| .Lprepare_ftrace_return: |
| .long prepare_ftrace_return |
| |
| .globl return_to_handler |
| return_to_handler: |
| /* |
| * Save the return values. |
| */ |
| mov.l r0, @-r15 |
| mov.l r1, @-r15 |
| |
| mov #0, r4 |
| |
| mov.l .Lftrace_return_to_handler, r0 |
| jsr @r0 |
| nop |
| |
| /* |
| * The return value from ftrace_return_handler has the real |
| * address that we should return to. |
| */ |
| lds r0, pr |
| mov.l @r15+, r1 |
| rts |
| mov.l @r15+, r0 |
| |
| |
| .align 2 |
| .Lftrace_return_to_handler: |
| .long ftrace_return_to_handler |
| #endif /* CONFIG_FUNCTION_GRAPH_TRACER */ |
| #endif /* CONFIG_FUNCTION_TRACER */ |
| |
| #ifdef CONFIG_STACK_DEBUG |
| .globl stack_panic |
| stack_panic: |
| mov.l .Ldump_stack, r0 |
| jsr @r0 |
| nop |
| |
| mov.l .Lpanic, r0 |
| jsr @r0 |
| mov.l .Lpanic_s, r4 |
| |
| rts |
| nop |
| |
| .align 2 |
| .L_init_thread_union: |
| .long init_thread_union |
| .L_ebss: |
| .long __bss_stop |
| .Lpanic: |
| .long panic |
| .Lpanic_s: |
| .long .Lpanic_str |
| .Ldump_stack: |
| .long dump_stack |
| |
| .section .rodata |
| .align 2 |
| .Lpanic_str: |
| .string "Stack error" |
| #endif /* CONFIG_STACK_DEBUG */ |