stacktrace.c (7338B)
1// SPDX-License-Identifier: GPL-2.0-only 2/* 3 * stacktrace.c : stacktracing APIs needed by rest of kernel 4 * (wrappers over ARC dwarf based unwinder) 5 * 6 * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com) 7 * 8 * vineetg: aug 2009 9 * -Implemented CONFIG_STACKTRACE APIs, primarily save_stack_trace_tsk( ) 10 * for displaying task's kernel mode call stack in /proc/<pid>/stack 11 * -Iterator based approach to have single copy of unwinding core and APIs 12 * needing unwinding, implement the logic in iterator regarding: 13 * = which frame onwards to start capture 14 * = which frame to stop capturing (wchan) 15 * = specifics of data structs where trace is saved(CONFIG_STACKTRACE etc) 16 * 17 * vineetg: March 2009 18 * -Implemented correct versions of thread_saved_pc() and __get_wchan() 19 * 20 * rajeshwarr: 2008 21 * -Initial implementation 22 */ 23 24#include <linux/ptrace.h> 25#include <linux/export.h> 26#include <linux/stacktrace.h> 27#include <linux/kallsyms.h> 28#include <linux/sched/debug.h> 29 30#include <asm/arcregs.h> 31#include <asm/unwind.h> 32#include <asm/switch_to.h> 33 34/*------------------------------------------------------------------------- 35 * Unwinder Iterator 36 *------------------------------------------------------------------------- 37 */ 38 39#ifdef CONFIG_ARC_DW2_UNWIND 40 41static int 42seed_unwind_frame_info(struct task_struct *tsk, struct pt_regs *regs, 43 struct unwind_frame_info *frame_info) 44{ 45 if (regs) { 46 /* 47 * Asynchronous unwinding of intr/exception 48 * - Just uses the pt_regs passed 49 */ 50 frame_info->task = tsk; 51 52 frame_info->regs.r27 = regs->fp; 53 frame_info->regs.r28 = regs->sp; 54 frame_info->regs.r31 = regs->blink; 55 frame_info->regs.r63 = regs->ret; 56 frame_info->call_frame = 0; 57 } else if (tsk == NULL || tsk == current) { 58 /* 59 * synchronous unwinding (e.g. dump_stack) 60 * - uses current values of SP and friends 61 */ 62 unsigned long fp, sp, blink, ret; 63 frame_info->task = current; 64 65 __asm__ __volatile__( 66 "mov %0,r27\n\t" 67 "mov %1,r28\n\t" 68 "mov %2,r31\n\t" 69 "mov %3,r63\n\t" 70 : "=r"(fp), "=r"(sp), "=r"(blink), "=r"(ret) 71 ); 72 73 frame_info->regs.r27 = fp; 74 frame_info->regs.r28 = sp; 75 frame_info->regs.r31 = blink; 76 frame_info->regs.r63 = ret; 77 frame_info->call_frame = 0; 78 } else { 79 /* 80 * Asynchronous unwinding of a likely sleeping task 81 * - first ensure it is actually sleeping 82 * - if so, it will be in __switch_to, kernel mode SP of task 83 * is safe-kept and BLINK at a well known location in there 84 */ 85 86 if (task_is_running(tsk)) 87 return -1; 88 89 frame_info->task = tsk; 90 91 frame_info->regs.r27 = TSK_K_FP(tsk); 92 frame_info->regs.r28 = TSK_K_ESP(tsk); 93 frame_info->regs.r31 = TSK_K_BLINK(tsk); 94 frame_info->regs.r63 = (unsigned int)__switch_to; 95 96 /* In the prologue of __switch_to, first FP is saved on stack 97 * and then SP is copied to FP. Dwarf assumes cfa as FP based 98 * but we didn't save FP. The value retrieved above is FP's 99 * state in previous frame. 100 * As a work around for this, we unwind from __switch_to start 101 * and adjust SP accordingly. The other limitation is that 102 * __switch_to macro is dwarf rules are not generated for inline 103 * assembly code 104 */ 105 frame_info->regs.r27 = 0; 106 frame_info->regs.r28 += 60; 107 frame_info->call_frame = 0; 108 109 } 110 return 0; 111} 112 113#endif 114 115notrace noinline unsigned int 116arc_unwind_core(struct task_struct *tsk, struct pt_regs *regs, 117 int (*consumer_fn) (unsigned int, void *), void *arg) 118{ 119#ifdef CONFIG_ARC_DW2_UNWIND 120 int ret = 0, cnt = 0; 121 unsigned int address; 122 struct unwind_frame_info frame_info; 123 124 if (seed_unwind_frame_info(tsk, regs, &frame_info)) 125 return 0; 126 127 while (1) { 128 address = UNW_PC(&frame_info); 129 130 if (!address || !__kernel_text_address(address)) 131 break; 132 133 if (consumer_fn(address, arg) == -1) 134 break; 135 136 ret = arc_unwind(&frame_info); 137 if (ret) 138 break; 139 140 frame_info.regs.r63 = frame_info.regs.r31; 141 142 if (cnt++ > 128) { 143 printk("unwinder looping too long, aborting !\n"); 144 return 0; 145 } 146 } 147 148 return address; /* return the last address it saw */ 149#else 150 /* On ARC, only Dward based unwinder works. fp based backtracing is 151 * not possible (-fno-omit-frame-pointer) because of the way function 152 * prologue is setup (callee regs saved and then fp set and not other 153 * way around 154 */ 155 pr_warn_once("CONFIG_ARC_DW2_UNWIND needs to be enabled\n"); 156 return 0; 157 158#endif 159} 160 161/*------------------------------------------------------------------------- 162 * callbacks called by unwinder iterator to implement kernel APIs 163 * 164 * The callback can return -1 to force the iterator to stop, which by default 165 * keeps going till the bottom-most frame. 166 *------------------------------------------------------------------------- 167 */ 168 169/* Call-back which plugs into unwinding core to dump the stack in 170 * case of panic/OOPs/BUG etc 171 */ 172static int __print_sym(unsigned int address, void *arg) 173{ 174 const char *loglvl = arg; 175 176 printk("%s %pS\n", loglvl, (void *)address); 177 return 0; 178} 179 180#ifdef CONFIG_STACKTRACE 181 182/* Call-back which plugs into unwinding core to capture the 183 * traces needed by kernel on /proc/<pid>/stack 184 */ 185static int __collect_all(unsigned int address, void *arg) 186{ 187 struct stack_trace *trace = arg; 188 189 if (trace->skip > 0) 190 trace->skip--; 191 else 192 trace->entries[trace->nr_entries++] = address; 193 194 if (trace->nr_entries >= trace->max_entries) 195 return -1; 196 197 return 0; 198} 199 200static int __collect_all_but_sched(unsigned int address, void *arg) 201{ 202 struct stack_trace *trace = arg; 203 204 if (in_sched_functions(address)) 205 return 0; 206 207 if (trace->skip > 0) 208 trace->skip--; 209 else 210 trace->entries[trace->nr_entries++] = address; 211 212 if (trace->nr_entries >= trace->max_entries) 213 return -1; 214 215 return 0; 216} 217 218#endif 219 220static int __get_first_nonsched(unsigned int address, void *unused) 221{ 222 if (in_sched_functions(address)) 223 return 0; 224 225 return -1; 226} 227 228/*------------------------------------------------------------------------- 229 * APIs expected by various kernel sub-systems 230 *------------------------------------------------------------------------- 231 */ 232 233noinline void show_stacktrace(struct task_struct *tsk, struct pt_regs *regs, 234 const char *loglvl) 235{ 236 printk("%s\nStack Trace:\n", loglvl); 237 arc_unwind_core(tsk, regs, __print_sym, (void *)loglvl); 238} 239EXPORT_SYMBOL(show_stacktrace); 240 241/* Expected by sched Code */ 242void show_stack(struct task_struct *tsk, unsigned long *sp, const char *loglvl) 243{ 244 show_stacktrace(tsk, NULL, loglvl); 245} 246 247/* Another API expected by schedular, shows up in "ps" as Wait Channel 248 * Of course just returning schedule( ) would be pointless so unwind until 249 * the function is not in schedular code 250 */ 251unsigned int __get_wchan(struct task_struct *tsk) 252{ 253 return arc_unwind_core(tsk, NULL, __get_first_nonsched, NULL); 254} 255 256#ifdef CONFIG_STACKTRACE 257 258/* 259 * API required by CONFIG_STACKTRACE, CONFIG_LATENCYTOP. 260 * A typical use is when /proc/<pid>/stack is queried by userland 261 */ 262void save_stack_trace_tsk(struct task_struct *tsk, struct stack_trace *trace) 263{ 264 /* Assumes @tsk is sleeping so unwinds from __switch_to */ 265 arc_unwind_core(tsk, NULL, __collect_all_but_sched, trace); 266} 267 268void save_stack_trace(struct stack_trace *trace) 269{ 270 /* Pass NULL for task so it unwinds the current call frame */ 271 arc_unwind_core(NULL, NULL, __collect_all, trace); 272} 273EXPORT_SYMBOL_GPL(save_stack_trace); 274#endif