1 // SPDX-License-Identifier: GPL-2.0
3 * Code for tracing calls in Linux kernel.
6 * based on code for x86 which is:
9 * future possible enhancements:
10 * - add CONFIG_STACK_TRACER
13 #include <linux/init.h>
14 #include <linux/ftrace.h>
15 #include <linux/uaccess.h>
16 #include <linux/kprobes.h>
17 #include <linux/ptrace.h>
19 #include <asm/assembly.h>
20 #include <asm/sections.h>
21 #include <asm/ftrace.h>
22 #include <asm/patch.h>
24 #define __hot __attribute__ ((__section__ (".text.hot")))
26 #ifdef CONFIG_FUNCTION_GRAPH_TRACER
28 * Hook the return address and push it in the stack of return addrs
29 * in current thread info.
31 static void __hot prepare_ftrace_return(unsigned long *parent,
32 unsigned long self_addr)
35 extern int parisc_return_to_handler;
37 if (unlikely(ftrace_graph_is_dead()))
40 if (unlikely(atomic_read(¤t->tracing_graph_pause)))
45 if (!function_graph_enter(old, self_addr, 0, NULL))
46 /* activate parisc_return_to_handler() as return point */
47 *parent = (unsigned long) &parisc_return_to_handler;
49 #endif /* CONFIG_FUNCTION_GRAPH_TRACER */
51 void notrace __hot ftrace_function_trampoline(unsigned long parent,
52 unsigned long self_addr,
53 unsigned long org_sp_gr3,
56 #ifndef CONFIG_DYNAMIC_FTRACE
57 extern ftrace_func_t ftrace_trace_function;
59 extern struct ftrace_ops *function_trace_op;
61 if (function_trace_op->flags & FTRACE_OPS_FL_ENABLED &&
62 ftrace_trace_function != ftrace_stub)
63 ftrace_trace_function(self_addr, parent,
64 function_trace_op, regs);
66 #ifdef CONFIG_FUNCTION_GRAPH_TRACER
67 if (ftrace_graph_return != (trace_func_graph_ret_t) ftrace_stub ||
68 ftrace_graph_entry != ftrace_graph_entry_stub) {
69 unsigned long *parent_rp;
71 /* calculate pointer to %rp in stack */
72 parent_rp = (unsigned long *) (org_sp_gr3 - RP_OFFSET);
73 /* sanity check: parent_rp should hold parent */
74 if (*parent_rp != parent)
77 prepare_ftrace_return(parent_rp, self_addr);
83 #ifdef CONFIG_FUNCTION_GRAPH_TRACER
84 int ftrace_enable_ftrace_graph_caller(void)
89 int ftrace_disable_ftrace_graph_caller(void)
95 #ifdef CONFIG_DYNAMIC_FTRACE
97 int __init ftrace_dyn_arch_init(void)
101 int ftrace_update_ftrace_func(ftrace_func_t func)
106 int ftrace_modify_call(struct dyn_ftrace *rec, unsigned long old_addr,
112 unsigned long ftrace_call_adjust(unsigned long addr)
114 return addr+(FTRACE_PATCHABLE_FUNCTION_SIZE-1)*4;
117 int ftrace_make_call(struct dyn_ftrace *rec, unsigned long addr)
119 u32 insn[FTRACE_PATCHABLE_FUNCTION_SIZE];
125 unsigned long addr2 =
126 (unsigned long)dereference_function_descriptor((void *)addr);
128 u32 ftrace_trampoline[] = {
129 0x73c10208, /* std,ma r1,100(sp) */
130 0x0c2110c1, /* ldd -10(r1),r1 */
131 0xe820d002, /* bve,n (r1) */
134 0xe83f1fd7, /* b,l,n .-14,r1 */
137 u32 ftrace_trampoline_unaligned[] = {
140 0x37de0200, /* ldo 100(sp),sp */
141 0x73c13e01, /* std r1,-100(sp) */
142 0x34213ff9, /* ldo -4(r1),r1 */
143 0x50213fc1, /* ldd -20(r1),r1 */
144 0xe820d002, /* bve,n (r1) */
145 0xe83f1fcf, /* b,l,n .-20,r1 */
148 BUILD_BUG_ON(ARRAY_SIZE(ftrace_trampoline_unaligned) >
149 FTRACE_PATCHABLE_FUNCTION_SIZE);
151 u32 ftrace_trampoline[] = {
153 0x6fc10080, /* stw,ma r1,40(sp) */
154 0x48213fd1, /* ldw -18(r1),r1 */
155 0xe820c002, /* bv,n r0(r1) */
156 0xe83f1fdf, /* b,l,n .-c,r1 */
160 BUILD_BUG_ON(ARRAY_SIZE(ftrace_trampoline) >
161 FTRACE_PATCHABLE_FUNCTION_SIZE);
163 size = sizeof(ftrace_trampoline);
164 tramp = ftrace_trampoline;
168 size = sizeof(ftrace_trampoline_unaligned);
169 tramp = ftrace_trampoline_unaligned;
173 ip = (void *)(rec->ip + 4 - size);
175 ret = probe_kernel_read(insn, ip, size);
179 for (i = 0; i < size / 4; i++) {
180 if (insn[i] != INSN_NOP)
184 __patch_text_multiple(ip, tramp, size);
188 int ftrace_make_nop(struct module *mod, struct dyn_ftrace *rec,
191 u32 insn[FTRACE_PATCHABLE_FUNCTION_SIZE];
194 for (i = 0; i < ARRAY_SIZE(insn); i++)
197 __patch_text((void *)rec->ip, INSN_NOP);
198 __patch_text_multiple((void *)rec->ip + 4 - sizeof(insn),
199 insn, sizeof(insn)-4);
204 #ifdef CONFIG_KPROBES_ON_FTRACE
205 void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip,
206 struct ftrace_ops *ops, struct pt_regs *regs)
208 struct kprobe_ctlblk *kcb;
209 struct kprobe *p = get_kprobe((kprobe_opcode_t *)ip);
211 if (unlikely(!p) || kprobe_disabled(p))
214 if (kprobe_running()) {
215 kprobes_inc_nmissed_count(p);
219 __this_cpu_write(current_kprobe, p);
221 kcb = get_kprobe_ctlblk();
222 kcb->kprobe_status = KPROBE_HIT_ACTIVE;
225 regs->iaoq[1] = ip + 4;
227 if (!p->pre_handler || !p->pre_handler(p, regs)) {
228 regs->iaoq[0] = ip + 4;
229 regs->iaoq[1] = ip + 8;
231 if (unlikely(p->post_handler)) {
232 kcb->kprobe_status = KPROBE_HIT_SSDONE;
233 p->post_handler(p, regs, 0);
236 __this_cpu_write(current_kprobe, NULL);
238 NOKPROBE_SYMBOL(kprobe_ftrace_handler);
240 int arch_prepare_kprobe_ftrace(struct kprobe *p)
242 p->ainsn.insn = NULL;