]> Git Repo - linux.git/blob - arch/riscv/kernel/traps.c
scsi: zfcp: Trace when request remove fails after qdio send fails
[linux.git] / arch / riscv / kernel / traps.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) 2012 Regents of the University of California
4  */
5
6 #include <linux/cpu.h>
7 #include <linux/kernel.h>
8 #include <linux/init.h>
9 #include <linux/sched.h>
10 #include <linux/sched/debug.h>
11 #include <linux/sched/signal.h>
12 #include <linux/signal.h>
13 #include <linux/kdebug.h>
14 #include <linux/uaccess.h>
15 #include <linux/kprobes.h>
16 #include <linux/mm.h>
17 #include <linux/module.h>
18 #include <linux/irq.h>
19 #include <linux/kexec.h>
20
21 #include <asm/asm-prototypes.h>
22 #include <asm/bug.h>
23 #include <asm/csr.h>
24 #include <asm/processor.h>
25 #include <asm/ptrace.h>
26 #include <asm/thread_info.h>
27
28 int show_unhandled_signals = 1;
29
30 static DEFINE_SPINLOCK(die_lock);
31
32 void die(struct pt_regs *regs, const char *str)
33 {
34         static int die_counter;
35         int ret;
36         long cause;
37
38         oops_enter();
39
40         spin_lock_irq(&die_lock);
41         console_verbose();
42         bust_spinlocks(1);
43
44         pr_emerg("%s [#%d]\n", str, ++die_counter);
45         print_modules();
46         if (regs)
47                 show_regs(regs);
48
49         cause = regs ? regs->cause : -1;
50         ret = notify_die(DIE_OOPS, str, regs, 0, cause, SIGSEGV);
51
52         if (kexec_should_crash(current))
53                 crash_kexec(regs);
54
55         bust_spinlocks(0);
56         add_taint(TAINT_DIE, LOCKDEP_NOW_UNRELIABLE);
57         spin_unlock_irq(&die_lock);
58         oops_exit();
59
60         if (in_interrupt())
61                 panic("Fatal exception in interrupt");
62         if (panic_on_oops)
63                 panic("Fatal exception");
64         if (ret != NOTIFY_STOP)
65                 make_task_dead(SIGSEGV);
66 }
67
68 void do_trap(struct pt_regs *regs, int signo, int code, unsigned long addr)
69 {
70         struct task_struct *tsk = current;
71
72         if (show_unhandled_signals && unhandled_signal(tsk, signo)
73             && printk_ratelimit()) {
74                 pr_info("%s[%d]: unhandled signal %d code 0x%x at 0x" REG_FMT,
75                         tsk->comm, task_pid_nr(tsk), signo, code, addr);
76                 print_vma_addr(KERN_CONT " in ", instruction_pointer(regs));
77                 pr_cont("\n");
78                 __show_regs(regs);
79         }
80
81         force_sig_fault(signo, code, (void __user *)addr);
82 }
83
84 static void do_trap_error(struct pt_regs *regs, int signo, int code,
85         unsigned long addr, const char *str)
86 {
87         current->thread.bad_cause = regs->cause;
88
89         if (user_mode(regs)) {
90                 do_trap(regs, signo, code, addr);
91         } else {
92                 if (!fixup_exception(regs))
93                         die(regs, str);
94         }
95 }
96
97 #if defined(CONFIG_XIP_KERNEL) && defined(CONFIG_RISCV_ALTERNATIVE)
98 #define __trap_section          __section(".xip.traps")
99 #else
100 #define __trap_section
101 #endif
102 #define DO_ERROR_INFO(name, signo, code, str)                           \
103 asmlinkage __visible __trap_section void name(struct pt_regs *regs)     \
104 {                                                                       \
105         do_trap_error(regs, signo, code, regs->epc, "Oops - " str);     \
106 }
107
108 DO_ERROR_INFO(do_trap_unknown,
109         SIGILL, ILL_ILLTRP, "unknown exception");
110 DO_ERROR_INFO(do_trap_insn_misaligned,
111         SIGBUS, BUS_ADRALN, "instruction address misaligned");
112 DO_ERROR_INFO(do_trap_insn_fault,
113         SIGSEGV, SEGV_ACCERR, "instruction access fault");
114 DO_ERROR_INFO(do_trap_insn_illegal,
115         SIGILL, ILL_ILLOPC, "illegal instruction");
116 DO_ERROR_INFO(do_trap_load_fault,
117         SIGSEGV, SEGV_ACCERR, "load access fault");
118 #ifndef CONFIG_RISCV_M_MODE
119 DO_ERROR_INFO(do_trap_load_misaligned,
120         SIGBUS, BUS_ADRALN, "Oops - load address misaligned");
121 DO_ERROR_INFO(do_trap_store_misaligned,
122         SIGBUS, BUS_ADRALN, "Oops - store (or AMO) address misaligned");
123 #else
124 int handle_misaligned_load(struct pt_regs *regs);
125 int handle_misaligned_store(struct pt_regs *regs);
126
127 asmlinkage void __trap_section do_trap_load_misaligned(struct pt_regs *regs)
128 {
129         if (!handle_misaligned_load(regs))
130                 return;
131         do_trap_error(regs, SIGBUS, BUS_ADRALN, regs->epc,
132                       "Oops - load address misaligned");
133 }
134
135 asmlinkage void __trap_section do_trap_store_misaligned(struct pt_regs *regs)
136 {
137         if (!handle_misaligned_store(regs))
138                 return;
139         do_trap_error(regs, SIGBUS, BUS_ADRALN, regs->epc,
140                       "Oops - store (or AMO) address misaligned");
141 }
142 #endif
143 DO_ERROR_INFO(do_trap_store_fault,
144         SIGSEGV, SEGV_ACCERR, "store (or AMO) access fault");
145 DO_ERROR_INFO(do_trap_ecall_u,
146         SIGILL, ILL_ILLTRP, "environment call from U-mode");
147 DO_ERROR_INFO(do_trap_ecall_s,
148         SIGILL, ILL_ILLTRP, "environment call from S-mode");
149 DO_ERROR_INFO(do_trap_ecall_m,
150         SIGILL, ILL_ILLTRP, "environment call from M-mode");
151
152 static inline unsigned long get_break_insn_length(unsigned long pc)
153 {
154         bug_insn_t insn;
155
156         if (get_kernel_nofault(insn, (bug_insn_t *)pc))
157                 return 0;
158
159         return GET_INSN_LENGTH(insn);
160 }
161
162 asmlinkage __visible __trap_section void do_trap_break(struct pt_regs *regs)
163 {
164 #ifdef CONFIG_KPROBES
165         if (kprobe_single_step_handler(regs))
166                 return;
167
168         if (kprobe_breakpoint_handler(regs))
169                 return;
170 #endif
171 #ifdef CONFIG_UPROBES
172         if (uprobe_single_step_handler(regs))
173                 return;
174
175         if (uprobe_breakpoint_handler(regs))
176                 return;
177 #endif
178         current->thread.bad_cause = regs->cause;
179
180         if (user_mode(regs))
181                 force_sig_fault(SIGTRAP, TRAP_BRKPT, (void __user *)regs->epc);
182 #ifdef CONFIG_KGDB
183         else if (notify_die(DIE_TRAP, "EBREAK", regs, 0, regs->cause, SIGTRAP)
184                                                                 == NOTIFY_STOP)
185                 return;
186 #endif
187         else if (report_bug(regs->epc, regs) == BUG_TRAP_TYPE_WARN)
188                 regs->epc += get_break_insn_length(regs->epc);
189         else
190                 die(regs, "Kernel BUG");
191 }
192 NOKPROBE_SYMBOL(do_trap_break);
193
194 #ifdef CONFIG_GENERIC_BUG
195 int is_valid_bugaddr(unsigned long pc)
196 {
197         bug_insn_t insn;
198
199         if (pc < VMALLOC_START)
200                 return 0;
201         if (get_kernel_nofault(insn, (bug_insn_t *)pc))
202                 return 0;
203         if ((insn & __INSN_LENGTH_MASK) == __INSN_LENGTH_32)
204                 return (insn == __BUG_INSN_32);
205         else
206                 return ((insn & __COMPRESSED_INSN_MASK) == __BUG_INSN_16);
207 }
208 #endif /* CONFIG_GENERIC_BUG */
209
210 #ifdef CONFIG_VMAP_STACK
211 /*
212  * Extra stack space that allows us to provide panic messages when the kernel
213  * has overflowed its stack.
214  */
215 static DEFINE_PER_CPU(unsigned long [OVERFLOW_STACK_SIZE/sizeof(long)],
216                 overflow_stack)__aligned(16);
217 /*
218  * A temporary stack for use by handle_kernel_stack_overflow.  This is used so
219  * we can call into C code to get the per-hart overflow stack.  Usage of this
220  * stack must be protected by spin_shadow_stack.
221  */
222 long shadow_stack[SHADOW_OVERFLOW_STACK_SIZE/sizeof(long)] __aligned(16);
223
224 /*
225  * A pseudo spinlock to protect the shadow stack from being used by multiple
226  * harts concurrently.  This isn't a real spinlock because the lock side must
227  * be taken without a valid stack and only a single register, it's only taken
228  * while in the process of panicing anyway so the performance and error
229  * checking a proper spinlock gives us doesn't matter.
230  */
231 unsigned long spin_shadow_stack;
232
233 asmlinkage unsigned long get_overflow_stack(void)
234 {
235         return (unsigned long)this_cpu_ptr(overflow_stack) +
236                 OVERFLOW_STACK_SIZE;
237 }
238
239 asmlinkage void handle_bad_stack(struct pt_regs *regs)
240 {
241         unsigned long tsk_stk = (unsigned long)current->stack;
242         unsigned long ovf_stk = (unsigned long)this_cpu_ptr(overflow_stack);
243
244         /*
245          * We're done with the shadow stack by this point, as we're on the
246          * overflow stack.  Tell any other concurrent overflowing harts that
247          * they can proceed with panicing by releasing the pseudo-spinlock.
248          *
249          * This pairs with an amoswap.aq in handle_kernel_stack_overflow.
250          */
251         smp_store_release(&spin_shadow_stack, 0);
252
253         console_verbose();
254
255         pr_emerg("Insufficient stack space to handle exception!\n");
256         pr_emerg("Task stack:     [0x%016lx..0x%016lx]\n",
257                         tsk_stk, tsk_stk + THREAD_SIZE);
258         pr_emerg("Overflow stack: [0x%016lx..0x%016lx]\n",
259                         ovf_stk, ovf_stk + OVERFLOW_STACK_SIZE);
260
261         __show_regs(regs);
262         panic("Kernel stack overflow");
263
264         for (;;)
265                 wait_for_interrupt();
266 }
267 #endif
This page took 0.047244 seconds and 4 git commands to generate.