xref: /linux/kernel/debug/debug_core.c (revision e7d759f31ca295d589f7420719c311870bb3166f)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Kernel Debug Core
4  *
5  * Maintainer: Jason Wessel <jason.wessel@windriver.com>
6  *
7  * Copyright (C) 2000-2001 VERITAS Software Corporation.
8  * Copyright (C) 2002-2004 Timesys Corporation
9  * Copyright (C) 2003-2004 Amit S. Kale <amitkale@linsyssoft.com>
10  * Copyright (C) 2004 Pavel Machek <pavel@ucw.cz>
11  * Copyright (C) 2004-2006 Tom Rini <trini@kernel.crashing.org>
12  * Copyright (C) 2004-2006 LinSysSoft Technologies Pvt. Ltd.
13  * Copyright (C) 2005-2009 Wind River Systems, Inc.
14  * Copyright (C) 2007 MontaVista Software, Inc.
15  * Copyright (C) 2008 Red Hat, Inc., Ingo Molnar <mingo@redhat.com>
16  *
17  * Contributors at various stages not listed above:
18  *  Jason Wessel ( jason.wessel@windriver.com )
19  *  George Anzinger <george@mvista.com>
20  *  Anurekh Saxena (anurekh.saxena@timesys.com)
21  *  Lake Stevens Instrument Division (Glenn Engel)
22  *  Jim Kingdon, Cygnus Support.
23  *
24  * Original KGDB stub: David Grothe <dave@gcom.com>,
25  * Tigran Aivazian <tigran@sco.com>
26  */
27 
28 #define pr_fmt(fmt) "KGDB: " fmt
29 
30 #include <linux/pid_namespace.h>
31 #include <linux/clocksource.h>
32 #include <linux/serial_core.h>
33 #include <linux/interrupt.h>
34 #include <linux/spinlock.h>
35 #include <linux/console.h>
36 #include <linux/threads.h>
37 #include <linux/uaccess.h>
38 #include <linux/kernel.h>
39 #include <linux/module.h>
40 #include <linux/ptrace.h>
41 #include <linux/string.h>
42 #include <linux/delay.h>
43 #include <linux/sched.h>
44 #include <linux/sysrq.h>
45 #include <linux/reboot.h>
46 #include <linux/init.h>
47 #include <linux/kgdb.h>
48 #include <linux/kdb.h>
49 #include <linux/nmi.h>
50 #include <linux/pid.h>
51 #include <linux/smp.h>
52 #include <linux/mm.h>
53 #include <linux/rcupdate.h>
54 #include <linux/irq.h>
55 #include <linux/security.h>
56 
57 #include <asm/cacheflush.h>
58 #include <asm/byteorder.h>
59 #include <linux/atomic.h>
60 
61 #include "debug_core.h"
62 
63 static int kgdb_break_asap;
64 
65 struct debuggerinfo_struct kgdb_info[NR_CPUS];
66 
67 /* kgdb_connected - Is a host GDB connected to us? */
68 int				kgdb_connected;
69 EXPORT_SYMBOL_GPL(kgdb_connected);
70 
71 /* All the KGDB handlers are installed */
72 int			kgdb_io_module_registered;
73 
74 /* Guard for recursive entry */
75 static int			exception_level;
76 
77 struct kgdb_io		*dbg_io_ops;
78 static DEFINE_SPINLOCK(kgdb_registration_lock);
79 
80 /* Action for the reboot notifier, a global allow kdb to change it */
81 static int kgdbreboot;
82 /* kgdb console driver is loaded */
83 static int kgdb_con_registered;
84 /* determine if kgdb console output should be used */
85 static int kgdb_use_con;
86 /* Flag for alternate operations for early debugging */
87 bool dbg_is_early = true;
88 /* Next cpu to become the master debug core */
89 int dbg_switch_cpu;
90 
91 /* Use kdb or gdbserver mode */
92 int dbg_kdb_mode = 1;
93 
94 module_param(kgdb_use_con, int, 0644);
95 module_param(kgdbreboot, int, 0644);
96 
97 /*
98  * Holds information about breakpoints in a kernel. These breakpoints are
99  * added and removed by gdb.
100  */
101 static struct kgdb_bkpt		kgdb_break[KGDB_MAX_BREAKPOINTS] = {
102 	[0 ... KGDB_MAX_BREAKPOINTS-1] = { .state = BP_UNDEFINED }
103 };
104 
105 /*
106  * The CPU# of the active CPU, or -1 if none:
107  */
108 atomic_t			kgdb_active = ATOMIC_INIT(-1);
109 EXPORT_SYMBOL_GPL(kgdb_active);
110 static DEFINE_RAW_SPINLOCK(dbg_master_lock);
111 static DEFINE_RAW_SPINLOCK(dbg_slave_lock);
112 
113 /*
114  * We use NR_CPUs not PERCPU, in case kgdb is used to debug early
115  * bootup code (which might not have percpu set up yet):
116  */
117 static atomic_t			masters_in_kgdb;
118 static atomic_t			slaves_in_kgdb;
119 atomic_t			kgdb_setting_breakpoint;
120 
121 struct task_struct		*kgdb_usethread;
122 struct task_struct		*kgdb_contthread;
123 
124 int				kgdb_single_step;
125 static pid_t			kgdb_sstep_pid;
126 
127 /* to keep track of the CPU which is doing the single stepping*/
128 atomic_t			kgdb_cpu_doing_single_step = ATOMIC_INIT(-1);
129 
130 /*
131  * If you are debugging a problem where roundup (the collection of
132  * all other CPUs) is a problem [this should be extremely rare],
133  * then use the nokgdbroundup option to avoid roundup. In that case
134  * the other CPUs might interfere with your debugging context, so
135  * use this with care:
136  */
137 static int kgdb_do_roundup = 1;
138 
139 static int __init opt_nokgdbroundup(char *str)
140 {
141 	kgdb_do_roundup = 0;
142 
143 	return 0;
144 }
145 
146 early_param("nokgdbroundup", opt_nokgdbroundup);
147 
148 /*
149  * Finally, some KGDB code :-)
150  */
151 
152 /*
153  * Weak aliases for breakpoint management,
154  * can be overridden by architectures when needed:
155  */
156 int __weak kgdb_arch_set_breakpoint(struct kgdb_bkpt *bpt)
157 {
158 	int err;
159 
160 	err = copy_from_kernel_nofault(bpt->saved_instr, (char *)bpt->bpt_addr,
161 				BREAK_INSTR_SIZE);
162 	if (err)
163 		return err;
164 	err = copy_to_kernel_nofault((char *)bpt->bpt_addr,
165 				 arch_kgdb_ops.gdb_bpt_instr, BREAK_INSTR_SIZE);
166 	return err;
167 }
168 NOKPROBE_SYMBOL(kgdb_arch_set_breakpoint);
169 
170 int __weak kgdb_arch_remove_breakpoint(struct kgdb_bkpt *bpt)
171 {
172 	return copy_to_kernel_nofault((char *)bpt->bpt_addr,
173 				  (char *)bpt->saved_instr, BREAK_INSTR_SIZE);
174 }
175 NOKPROBE_SYMBOL(kgdb_arch_remove_breakpoint);
176 
177 int __weak kgdb_validate_break_address(unsigned long addr)
178 {
179 	struct kgdb_bkpt tmp;
180 	int err;
181 
182 	if (kgdb_within_blocklist(addr))
183 		return -EINVAL;
184 
185 	/* Validate setting the breakpoint and then removing it.  If the
186 	 * remove fails, the kernel needs to emit a bad message because we
187 	 * are deep trouble not being able to put things back the way we
188 	 * found them.
189 	 */
190 	tmp.bpt_addr = addr;
191 	err = kgdb_arch_set_breakpoint(&tmp);
192 	if (err)
193 		return err;
194 	err = kgdb_arch_remove_breakpoint(&tmp);
195 	if (err)
196 		pr_err("Critical breakpoint error, kernel memory destroyed at: %lx\n",
197 		       addr);
198 	return err;
199 }
200 
201 unsigned long __weak kgdb_arch_pc(int exception, struct pt_regs *regs)
202 {
203 	return instruction_pointer(regs);
204 }
205 NOKPROBE_SYMBOL(kgdb_arch_pc);
206 
207 int __weak kgdb_arch_init(void)
208 {
209 	return 0;
210 }
211 
212 int __weak kgdb_skipexception(int exception, struct pt_regs *regs)
213 {
214 	return 0;
215 }
216 NOKPROBE_SYMBOL(kgdb_skipexception);
217 
218 #ifdef CONFIG_SMP
219 
220 /*
221  * Default (weak) implementation for kgdb_roundup_cpus
222  */
223 
224 void __weak kgdb_call_nmi_hook(void *ignored)
225 {
226 	/*
227 	 * NOTE: get_irq_regs() is supposed to get the registers from
228 	 * before the IPI interrupt happened and so is supposed to
229 	 * show where the processor was.  In some situations it's
230 	 * possible we might be called without an IPI, so it might be
231 	 * safer to figure out how to make kgdb_breakpoint() work
232 	 * properly here.
233 	 */
234 	kgdb_nmicallback(raw_smp_processor_id(), get_irq_regs());
235 }
236 NOKPROBE_SYMBOL(kgdb_call_nmi_hook);
237 
238 static DEFINE_PER_CPU(call_single_data_t, kgdb_roundup_csd) =
239 	CSD_INIT(kgdb_call_nmi_hook, NULL);
240 
241 void __weak kgdb_roundup_cpus(void)
242 {
243 	call_single_data_t *csd;
244 	int this_cpu = raw_smp_processor_id();
245 	int cpu;
246 	int ret;
247 
248 	for_each_online_cpu(cpu) {
249 		/* No need to roundup ourselves */
250 		if (cpu == this_cpu)
251 			continue;
252 
253 		csd = &per_cpu(kgdb_roundup_csd, cpu);
254 
255 		/*
256 		 * If it didn't round up last time, don't try again
257 		 * since smp_call_function_single_async() will block.
258 		 *
259 		 * If rounding_up is false then we know that the
260 		 * previous call must have at least started and that
261 		 * means smp_call_function_single_async() won't block.
262 		 */
263 		if (kgdb_info[cpu].rounding_up)
264 			continue;
265 		kgdb_info[cpu].rounding_up = true;
266 
267 		ret = smp_call_function_single_async(cpu, csd);
268 		if (ret)
269 			kgdb_info[cpu].rounding_up = false;
270 	}
271 }
272 NOKPROBE_SYMBOL(kgdb_roundup_cpus);
273 
274 #endif
275 
276 /*
277  * Some architectures need cache flushes when we set/clear a
278  * breakpoint:
279  */
280 static void kgdb_flush_swbreak_addr(unsigned long addr)
281 {
282 	if (!CACHE_FLUSH_IS_SAFE)
283 		return;
284 
285 	/* Force flush instruction cache if it was outside the mm */
286 	flush_icache_range(addr, addr + BREAK_INSTR_SIZE);
287 }
288 NOKPROBE_SYMBOL(kgdb_flush_swbreak_addr);
289 
290 /*
291  * SW breakpoint management:
292  */
293 int dbg_activate_sw_breakpoints(void)
294 {
295 	int error;
296 	int ret = 0;
297 	int i;
298 
299 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
300 		if (kgdb_break[i].state != BP_SET)
301 			continue;
302 
303 		error = kgdb_arch_set_breakpoint(&kgdb_break[i]);
304 		if (error) {
305 			ret = error;
306 			pr_info("BP install failed: %lx\n",
307 				kgdb_break[i].bpt_addr);
308 			continue;
309 		}
310 
311 		kgdb_flush_swbreak_addr(kgdb_break[i].bpt_addr);
312 		kgdb_break[i].state = BP_ACTIVE;
313 	}
314 	return ret;
315 }
316 NOKPROBE_SYMBOL(dbg_activate_sw_breakpoints);
317 
318 int dbg_set_sw_break(unsigned long addr)
319 {
320 	int err = kgdb_validate_break_address(addr);
321 	int breakno = -1;
322 	int i;
323 
324 	if (err)
325 		return err;
326 
327 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
328 		if ((kgdb_break[i].state == BP_SET) &&
329 					(kgdb_break[i].bpt_addr == addr))
330 			return -EEXIST;
331 	}
332 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
333 		if (kgdb_break[i].state == BP_REMOVED &&
334 					kgdb_break[i].bpt_addr == addr) {
335 			breakno = i;
336 			break;
337 		}
338 	}
339 
340 	if (breakno == -1) {
341 		for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
342 			if (kgdb_break[i].state == BP_UNDEFINED) {
343 				breakno = i;
344 				break;
345 			}
346 		}
347 	}
348 
349 	if (breakno == -1)
350 		return -E2BIG;
351 
352 	kgdb_break[breakno].state = BP_SET;
353 	kgdb_break[breakno].type = BP_BREAKPOINT;
354 	kgdb_break[breakno].bpt_addr = addr;
355 
356 	return 0;
357 }
358 
359 int dbg_deactivate_sw_breakpoints(void)
360 {
361 	int error;
362 	int ret = 0;
363 	int i;
364 
365 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
366 		if (kgdb_break[i].state != BP_ACTIVE)
367 			continue;
368 		error = kgdb_arch_remove_breakpoint(&kgdb_break[i]);
369 		if (error) {
370 			pr_info("BP remove failed: %lx\n",
371 				kgdb_break[i].bpt_addr);
372 			ret = error;
373 		}
374 
375 		kgdb_flush_swbreak_addr(kgdb_break[i].bpt_addr);
376 		kgdb_break[i].state = BP_SET;
377 	}
378 	return ret;
379 }
380 NOKPROBE_SYMBOL(dbg_deactivate_sw_breakpoints);
381 
382 int dbg_remove_sw_break(unsigned long addr)
383 {
384 	int i;
385 
386 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
387 		if ((kgdb_break[i].state == BP_SET) &&
388 				(kgdb_break[i].bpt_addr == addr)) {
389 			kgdb_break[i].state = BP_REMOVED;
390 			return 0;
391 		}
392 	}
393 	return -ENOENT;
394 }
395 
396 int kgdb_isremovedbreak(unsigned long addr)
397 {
398 	int i;
399 
400 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
401 		if ((kgdb_break[i].state == BP_REMOVED) &&
402 					(kgdb_break[i].bpt_addr == addr))
403 			return 1;
404 	}
405 	return 0;
406 }
407 
408 int kgdb_has_hit_break(unsigned long addr)
409 {
410 	int i;
411 
412 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
413 		if (kgdb_break[i].state == BP_ACTIVE &&
414 		    kgdb_break[i].bpt_addr == addr)
415 			return 1;
416 	}
417 	return 0;
418 }
419 
420 int dbg_remove_all_break(void)
421 {
422 	int error;
423 	int i;
424 
425 	/* Clear memory breakpoints. */
426 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
427 		if (kgdb_break[i].state != BP_ACTIVE)
428 			goto setundefined;
429 		error = kgdb_arch_remove_breakpoint(&kgdb_break[i]);
430 		if (error)
431 			pr_err("breakpoint remove failed: %lx\n",
432 			       kgdb_break[i].bpt_addr);
433 setundefined:
434 		kgdb_break[i].state = BP_UNDEFINED;
435 	}
436 
437 	/* Clear hardware breakpoints. */
438 	if (arch_kgdb_ops.remove_all_hw_break)
439 		arch_kgdb_ops.remove_all_hw_break();
440 
441 	return 0;
442 }
443 
444 void kgdb_free_init_mem(void)
445 {
446 	int i;
447 
448 	/* Clear init memory breakpoints. */
449 	for (i = 0; i < KGDB_MAX_BREAKPOINTS; i++) {
450 		if (init_section_contains((void *)kgdb_break[i].bpt_addr, 0))
451 			kgdb_break[i].state = BP_UNDEFINED;
452 	}
453 }
454 
455 #ifdef CONFIG_KGDB_KDB
456 void kdb_dump_stack_on_cpu(int cpu)
457 {
458 	if (cpu == raw_smp_processor_id() || !IS_ENABLED(CONFIG_SMP)) {
459 		dump_stack();
460 		return;
461 	}
462 
463 	if (!(kgdb_info[cpu].exception_state & DCPU_IS_SLAVE)) {
464 		kdb_printf("ERROR: Task on cpu %d didn't stop in the debugger\n",
465 			   cpu);
466 		return;
467 	}
468 
469 	/*
470 	 * In general, architectures don't support dumping the stack of a
471 	 * "running" process that's not the current one.  From the point of
472 	 * view of the Linux, kernel processes that are looping in the kgdb
473 	 * slave loop are still "running".  There's also no API (that actually
474 	 * works across all architectures) that can do a stack crawl based
475 	 * on registers passed as a parameter.
476 	 *
477 	 * Solve this conundrum by asking slave CPUs to do the backtrace
478 	 * themselves.
479 	 */
480 	kgdb_info[cpu].exception_state |= DCPU_WANT_BT;
481 	while (kgdb_info[cpu].exception_state & DCPU_WANT_BT)
482 		cpu_relax();
483 }
484 #endif
485 
486 /*
487  * Return true if there is a valid kgdb I/O module.  Also if no
488  * debugger is attached a message can be printed to the console about
489  * waiting for the debugger to attach.
490  *
491  * The print_wait argument is only to be true when called from inside
492  * the core kgdb_handle_exception, because it will wait for the
493  * debugger to attach.
494  */
495 static int kgdb_io_ready(int print_wait)
496 {
497 	if (!dbg_io_ops)
498 		return 0;
499 	if (kgdb_connected)
500 		return 1;
501 	if (atomic_read(&kgdb_setting_breakpoint))
502 		return 1;
503 	if (print_wait) {
504 #ifdef CONFIG_KGDB_KDB
505 		if (!dbg_kdb_mode)
506 			pr_crit("waiting... or $3#33 for KDB\n");
507 #else
508 		pr_crit("Waiting for remote debugger\n");
509 #endif
510 	}
511 	return 1;
512 }
513 NOKPROBE_SYMBOL(kgdb_io_ready);
514 
515 static int kgdb_reenter_check(struct kgdb_state *ks)
516 {
517 	unsigned long addr;
518 
519 	if (atomic_read(&kgdb_active) != raw_smp_processor_id())
520 		return 0;
521 
522 	/* Panic on recursive debugger calls: */
523 	exception_level++;
524 	addr = kgdb_arch_pc(ks->ex_vector, ks->linux_regs);
525 	dbg_deactivate_sw_breakpoints();
526 
527 	/*
528 	 * If the break point removed ok at the place exception
529 	 * occurred, try to recover and print a warning to the end
530 	 * user because the user planted a breakpoint in a place that
531 	 * KGDB needs in order to function.
532 	 */
533 	if (dbg_remove_sw_break(addr) == 0) {
534 		exception_level = 0;
535 		kgdb_skipexception(ks->ex_vector, ks->linux_regs);
536 		dbg_activate_sw_breakpoints();
537 		pr_crit("re-enter error: breakpoint removed %lx\n", addr);
538 		WARN_ON_ONCE(1);
539 
540 		return 1;
541 	}
542 	dbg_remove_all_break();
543 	kgdb_skipexception(ks->ex_vector, ks->linux_regs);
544 
545 	if (exception_level > 1) {
546 		dump_stack();
547 		kgdb_io_module_registered = false;
548 		panic("Recursive entry to debugger");
549 	}
550 
551 	pr_crit("re-enter exception: ALL breakpoints killed\n");
552 #ifdef CONFIG_KGDB_KDB
553 	/* Allow kdb to debug itself one level */
554 	return 0;
555 #endif
556 	dump_stack();
557 	panic("Recursive entry to debugger");
558 
559 	return 1;
560 }
561 NOKPROBE_SYMBOL(kgdb_reenter_check);
562 
563 static void dbg_touch_watchdogs(void)
564 {
565 	touch_softlockup_watchdog_sync();
566 	clocksource_touch_watchdog();
567 	rcu_cpu_stall_reset();
568 }
569 NOKPROBE_SYMBOL(dbg_touch_watchdogs);
570 
571 static int kgdb_cpu_enter(struct kgdb_state *ks, struct pt_regs *regs,
572 		int exception_state)
573 {
574 	unsigned long flags;
575 	int sstep_tries = 100;
576 	int error;
577 	int cpu;
578 	int trace_on = 0;
579 	int online_cpus = num_online_cpus();
580 	u64 time_left;
581 
582 	kgdb_info[ks->cpu].enter_kgdb++;
583 	kgdb_info[ks->cpu].exception_state |= exception_state;
584 
585 	if (exception_state == DCPU_WANT_MASTER)
586 		atomic_inc(&masters_in_kgdb);
587 	else
588 		atomic_inc(&slaves_in_kgdb);
589 
590 	if (arch_kgdb_ops.disable_hw_break)
591 		arch_kgdb_ops.disable_hw_break(regs);
592 
593 acquirelock:
594 	rcu_read_lock();
595 	/*
596 	 * Interrupts will be restored by the 'trap return' code, except when
597 	 * single stepping.
598 	 */
599 	local_irq_save(flags);
600 
601 	cpu = ks->cpu;
602 	kgdb_info[cpu].debuggerinfo = regs;
603 	kgdb_info[cpu].task = current;
604 	kgdb_info[cpu].ret_state = 0;
605 	kgdb_info[cpu].irq_depth = hardirq_count() >> HARDIRQ_SHIFT;
606 
607 	/* Make sure the above info reaches the primary CPU */
608 	smp_mb();
609 
610 	if (exception_level == 1) {
611 		if (raw_spin_trylock(&dbg_master_lock))
612 			atomic_xchg(&kgdb_active, cpu);
613 		goto cpu_master_loop;
614 	}
615 
616 	/*
617 	 * CPU will loop if it is a slave or request to become a kgdb
618 	 * master cpu and acquire the kgdb_active lock:
619 	 */
620 	while (1) {
621 cpu_loop:
622 		if (kgdb_info[cpu].exception_state & DCPU_NEXT_MASTER) {
623 			kgdb_info[cpu].exception_state &= ~DCPU_NEXT_MASTER;
624 			goto cpu_master_loop;
625 		} else if (kgdb_info[cpu].exception_state & DCPU_WANT_MASTER) {
626 			if (raw_spin_trylock(&dbg_master_lock)) {
627 				atomic_xchg(&kgdb_active, cpu);
628 				break;
629 			}
630 		} else if (kgdb_info[cpu].exception_state & DCPU_WANT_BT) {
631 			dump_stack();
632 			kgdb_info[cpu].exception_state &= ~DCPU_WANT_BT;
633 		} else if (kgdb_info[cpu].exception_state & DCPU_IS_SLAVE) {
634 			if (!raw_spin_is_locked(&dbg_slave_lock))
635 				goto return_normal;
636 		} else {
637 return_normal:
638 			/* Return to normal operation by executing any
639 			 * hw breakpoint fixup.
640 			 */
641 			if (arch_kgdb_ops.correct_hw_break)
642 				arch_kgdb_ops.correct_hw_break();
643 			if (trace_on)
644 				tracing_on();
645 			kgdb_info[cpu].debuggerinfo = NULL;
646 			kgdb_info[cpu].task = NULL;
647 			kgdb_info[cpu].exception_state &=
648 				~(DCPU_WANT_MASTER | DCPU_IS_SLAVE);
649 			kgdb_info[cpu].enter_kgdb--;
650 			smp_mb__before_atomic();
651 			atomic_dec(&slaves_in_kgdb);
652 			dbg_touch_watchdogs();
653 			local_irq_restore(flags);
654 			rcu_read_unlock();
655 			return 0;
656 		}
657 		cpu_relax();
658 	}
659 
660 	/*
661 	 * For single stepping, try to only enter on the processor
662 	 * that was single stepping.  To guard against a deadlock, the
663 	 * kernel will only try for the value of sstep_tries before
664 	 * giving up and continuing on.
665 	 */
666 	if (atomic_read(&kgdb_cpu_doing_single_step) != -1 &&
667 	    (kgdb_info[cpu].task &&
668 	     kgdb_info[cpu].task->pid != kgdb_sstep_pid) && --sstep_tries) {
669 		atomic_set(&kgdb_active, -1);
670 		raw_spin_unlock(&dbg_master_lock);
671 		dbg_touch_watchdogs();
672 		local_irq_restore(flags);
673 		rcu_read_unlock();
674 
675 		goto acquirelock;
676 	}
677 
678 	if (!kgdb_io_ready(1)) {
679 		kgdb_info[cpu].ret_state = 1;
680 		goto kgdb_restore; /* No I/O connection, resume the system */
681 	}
682 
683 	/*
684 	 * Don't enter if we have hit a removed breakpoint.
685 	 */
686 	if (kgdb_skipexception(ks->ex_vector, ks->linux_regs))
687 		goto kgdb_restore;
688 
689 	atomic_inc(&ignore_console_lock_warning);
690 
691 	/* Call the I/O driver's pre_exception routine */
692 	if (dbg_io_ops->pre_exception)
693 		dbg_io_ops->pre_exception();
694 
695 	/*
696 	 * Get the passive CPU lock which will hold all the non-primary
697 	 * CPU in a spin state while the debugger is active
698 	 */
699 	if (!kgdb_single_step)
700 		raw_spin_lock(&dbg_slave_lock);
701 
702 #ifdef CONFIG_SMP
703 	/* If send_ready set, slaves are already waiting */
704 	if (ks->send_ready)
705 		atomic_set(ks->send_ready, 1);
706 
707 	/* Signal the other CPUs to enter kgdb_wait() */
708 	else if ((!kgdb_single_step) && kgdb_do_roundup)
709 		kgdb_roundup_cpus();
710 #endif
711 
712 	/*
713 	 * Wait for the other CPUs to be notified and be waiting for us:
714 	 */
715 	time_left = MSEC_PER_SEC;
716 	while (kgdb_do_roundup && --time_left &&
717 	       (atomic_read(&masters_in_kgdb) + atomic_read(&slaves_in_kgdb)) !=
718 		   online_cpus)
719 		udelay(1000);
720 	if (!time_left)
721 		pr_crit("Timed out waiting for secondary CPUs.\n");
722 
723 	/*
724 	 * At this point the primary processor is completely
725 	 * in the debugger and all secondary CPUs are quiescent
726 	 */
727 	dbg_deactivate_sw_breakpoints();
728 	kgdb_single_step = 0;
729 	kgdb_contthread = current;
730 	exception_level = 0;
731 	trace_on = tracing_is_on();
732 	if (trace_on)
733 		tracing_off();
734 
735 	while (1) {
736 cpu_master_loop:
737 		if (dbg_kdb_mode) {
738 			kgdb_connected = 1;
739 			error = kdb_stub(ks);
740 			if (error == -1)
741 				continue;
742 			kgdb_connected = 0;
743 		} else {
744 			/*
745 			 * This is a brutal way to interfere with the debugger
746 			 * and prevent gdb being used to poke at kernel memory.
747 			 * This could cause trouble if lockdown is applied when
748 			 * there is already an active gdb session. For now the
749 			 * answer is simply "don't do that". Typically lockdown
750 			 * *will* be applied before the debug core gets started
751 			 * so only developers using kgdb for fairly advanced
752 			 * early kernel debug can be biten by this. Hopefully
753 			 * they are sophisticated enough to take care of
754 			 * themselves, especially with help from the lockdown
755 			 * message printed on the console!
756 			 */
757 			if (security_locked_down(LOCKDOWN_DBG_WRITE_KERNEL)) {
758 				if (IS_ENABLED(CONFIG_KGDB_KDB)) {
759 					/* Switch back to kdb if possible... */
760 					dbg_kdb_mode = 1;
761 					continue;
762 				} else {
763 					/* ... otherwise just bail */
764 					break;
765 				}
766 			}
767 			error = gdb_serial_stub(ks);
768 		}
769 
770 		if (error == DBG_PASS_EVENT) {
771 			dbg_kdb_mode = !dbg_kdb_mode;
772 		} else if (error == DBG_SWITCH_CPU_EVENT) {
773 			kgdb_info[dbg_switch_cpu].exception_state |=
774 				DCPU_NEXT_MASTER;
775 			goto cpu_loop;
776 		} else {
777 			kgdb_info[cpu].ret_state = error;
778 			break;
779 		}
780 	}
781 
782 	dbg_activate_sw_breakpoints();
783 
784 	/* Call the I/O driver's post_exception routine */
785 	if (dbg_io_ops->post_exception)
786 		dbg_io_ops->post_exception();
787 
788 	atomic_dec(&ignore_console_lock_warning);
789 
790 	if (!kgdb_single_step) {
791 		raw_spin_unlock(&dbg_slave_lock);
792 		/* Wait till all the CPUs have quit from the debugger. */
793 		while (kgdb_do_roundup && atomic_read(&slaves_in_kgdb))
794 			cpu_relax();
795 	}
796 
797 kgdb_restore:
798 	if (atomic_read(&kgdb_cpu_doing_single_step) != -1) {
799 		int sstep_cpu = atomic_read(&kgdb_cpu_doing_single_step);
800 		if (kgdb_info[sstep_cpu].task)
801 			kgdb_sstep_pid = kgdb_info[sstep_cpu].task->pid;
802 		else
803 			kgdb_sstep_pid = 0;
804 	}
805 	if (arch_kgdb_ops.correct_hw_break)
806 		arch_kgdb_ops.correct_hw_break();
807 	if (trace_on)
808 		tracing_on();
809 
810 	kgdb_info[cpu].debuggerinfo = NULL;
811 	kgdb_info[cpu].task = NULL;
812 	kgdb_info[cpu].exception_state &=
813 		~(DCPU_WANT_MASTER | DCPU_IS_SLAVE);
814 	kgdb_info[cpu].enter_kgdb--;
815 	smp_mb__before_atomic();
816 	atomic_dec(&masters_in_kgdb);
817 	/* Free kgdb_active */
818 	atomic_set(&kgdb_active, -1);
819 	raw_spin_unlock(&dbg_master_lock);
820 	dbg_touch_watchdogs();
821 	local_irq_restore(flags);
822 	rcu_read_unlock();
823 
824 	return kgdb_info[cpu].ret_state;
825 }
826 NOKPROBE_SYMBOL(kgdb_cpu_enter);
827 
828 /*
829  * kgdb_handle_exception() - main entry point from a kernel exception
830  *
831  * Locking hierarchy:
832  *	interface locks, if any (begin_session)
833  *	kgdb lock (kgdb_active)
834  */
835 int
836 kgdb_handle_exception(int evector, int signo, int ecode, struct pt_regs *regs)
837 {
838 	struct kgdb_state kgdb_var;
839 	struct kgdb_state *ks = &kgdb_var;
840 	int ret = 0;
841 
842 	if (arch_kgdb_ops.enable_nmi)
843 		arch_kgdb_ops.enable_nmi(0);
844 	/*
845 	 * Avoid entering the debugger if we were triggered due to an oops
846 	 * but panic_timeout indicates the system should automatically
847 	 * reboot on panic. We don't want to get stuck waiting for input
848 	 * on such systems, especially if its "just" an oops.
849 	 */
850 	if (signo != SIGTRAP && panic_timeout)
851 		return 1;
852 
853 	memset(ks, 0, sizeof(struct kgdb_state));
854 	ks->cpu			= raw_smp_processor_id();
855 	ks->ex_vector		= evector;
856 	ks->signo		= signo;
857 	ks->err_code		= ecode;
858 	ks->linux_regs		= regs;
859 
860 	if (kgdb_reenter_check(ks))
861 		goto out; /* Ouch, double exception ! */
862 	if (kgdb_info[ks->cpu].enter_kgdb != 0)
863 		goto out;
864 
865 	ret = kgdb_cpu_enter(ks, regs, DCPU_WANT_MASTER);
866 out:
867 	if (arch_kgdb_ops.enable_nmi)
868 		arch_kgdb_ops.enable_nmi(1);
869 	return ret;
870 }
871 NOKPROBE_SYMBOL(kgdb_handle_exception);
872 
873 /*
874  * GDB places a breakpoint at this function to know dynamically loaded objects.
875  */
876 static int module_event(struct notifier_block *self, unsigned long val,
877 	void *data)
878 {
879 	return 0;
880 }
881 
882 static struct notifier_block dbg_module_load_nb = {
883 	.notifier_call	= module_event,
884 };
885 
886 int kgdb_nmicallback(int cpu, void *regs)
887 {
888 #ifdef CONFIG_SMP
889 	struct kgdb_state kgdb_var;
890 	struct kgdb_state *ks = &kgdb_var;
891 
892 	kgdb_info[cpu].rounding_up = false;
893 
894 	memset(ks, 0, sizeof(struct kgdb_state));
895 	ks->cpu			= cpu;
896 	ks->linux_regs		= regs;
897 
898 	if (kgdb_info[ks->cpu].enter_kgdb == 0 &&
899 			raw_spin_is_locked(&dbg_master_lock)) {
900 		kgdb_cpu_enter(ks, regs, DCPU_IS_SLAVE);
901 		return 0;
902 	}
903 #endif
904 	return 1;
905 }
906 NOKPROBE_SYMBOL(kgdb_nmicallback);
907 
908 int kgdb_nmicallin(int cpu, int trapnr, void *regs, int err_code,
909 							atomic_t *send_ready)
910 {
911 #ifdef CONFIG_SMP
912 	if (!kgdb_io_ready(0) || !send_ready)
913 		return 1;
914 
915 	if (kgdb_info[cpu].enter_kgdb == 0) {
916 		struct kgdb_state kgdb_var;
917 		struct kgdb_state *ks = &kgdb_var;
918 
919 		memset(ks, 0, sizeof(struct kgdb_state));
920 		ks->cpu			= cpu;
921 		ks->ex_vector		= trapnr;
922 		ks->signo		= SIGTRAP;
923 		ks->err_code		= err_code;
924 		ks->linux_regs		= regs;
925 		ks->send_ready		= send_ready;
926 		kgdb_cpu_enter(ks, regs, DCPU_WANT_MASTER);
927 		return 0;
928 	}
929 #endif
930 	return 1;
931 }
932 NOKPROBE_SYMBOL(kgdb_nmicallin);
933 
934 static void kgdb_console_write(struct console *co, const char *s,
935    unsigned count)
936 {
937 	unsigned long flags;
938 
939 	/* If we're debugging, or KGDB has not connected, don't try
940 	 * and print. */
941 	if (!kgdb_connected || atomic_read(&kgdb_active) != -1 || dbg_kdb_mode)
942 		return;
943 
944 	local_irq_save(flags);
945 	gdbstub_msg_write(s, count);
946 	local_irq_restore(flags);
947 }
948 
949 static struct console kgdbcons = {
950 	.name		= "kgdb",
951 	.write		= kgdb_console_write,
952 	.flags		= CON_PRINTBUFFER | CON_ENABLED,
953 	.index		= -1,
954 };
955 
956 static int __init opt_kgdb_con(char *str)
957 {
958 	kgdb_use_con = 1;
959 
960 	if (kgdb_io_module_registered && !kgdb_con_registered) {
961 		register_console(&kgdbcons);
962 		kgdb_con_registered = 1;
963 	}
964 
965 	return 0;
966 }
967 
968 early_param("kgdbcon", opt_kgdb_con);
969 
970 #ifdef CONFIG_MAGIC_SYSRQ
971 static void sysrq_handle_dbg(u8 key)
972 {
973 	if (!dbg_io_ops) {
974 		pr_crit("ERROR: No KGDB I/O module available\n");
975 		return;
976 	}
977 	if (!kgdb_connected) {
978 #ifdef CONFIG_KGDB_KDB
979 		if (!dbg_kdb_mode)
980 			pr_crit("KGDB or $3#33 for KDB\n");
981 #else
982 		pr_crit("Entering KGDB\n");
983 #endif
984 	}
985 
986 	kgdb_breakpoint();
987 }
988 
989 static const struct sysrq_key_op sysrq_dbg_op = {
990 	.handler	= sysrq_handle_dbg,
991 	.help_msg	= "debug(g)",
992 	.action_msg	= "DEBUG",
993 };
994 #endif
995 
996 void kgdb_panic(const char *msg)
997 {
998 	if (!kgdb_io_module_registered)
999 		return;
1000 
1001 	/*
1002 	 * We don't want to get stuck waiting for input from user if
1003 	 * "panic_timeout" indicates the system should automatically
1004 	 * reboot on panic.
1005 	 */
1006 	if (panic_timeout)
1007 		return;
1008 
1009 	debug_locks_off();
1010 	console_flush_on_panic(CONSOLE_FLUSH_PENDING);
1011 
1012 	if (dbg_kdb_mode)
1013 		kdb_printf("PANIC: %s\n", msg);
1014 
1015 	kgdb_breakpoint();
1016 }
1017 
1018 static void kgdb_initial_breakpoint(void)
1019 {
1020 	kgdb_break_asap = 0;
1021 
1022 	pr_crit("Waiting for connection from remote gdb...\n");
1023 	kgdb_breakpoint();
1024 }
1025 
1026 void __weak kgdb_arch_late(void)
1027 {
1028 }
1029 
1030 void __init dbg_late_init(void)
1031 {
1032 	dbg_is_early = false;
1033 	if (kgdb_io_module_registered)
1034 		kgdb_arch_late();
1035 	kdb_init(KDB_INIT_FULL);
1036 
1037 	if (kgdb_io_module_registered && kgdb_break_asap)
1038 		kgdb_initial_breakpoint();
1039 }
1040 
1041 static int
1042 dbg_notify_reboot(struct notifier_block *this, unsigned long code, void *x)
1043 {
1044 	/*
1045 	 * Take the following action on reboot notify depending on value:
1046 	 *    1 == Enter debugger
1047 	 *    0 == [the default] detach debug client
1048 	 *   -1 == Do nothing... and use this until the board resets
1049 	 */
1050 	switch (kgdbreboot) {
1051 	case 1:
1052 		kgdb_breakpoint();
1053 		goto done;
1054 	case -1:
1055 		goto done;
1056 	}
1057 	if (!dbg_kdb_mode)
1058 		gdbstub_exit(code);
1059 done:
1060 	return NOTIFY_DONE;
1061 }
1062 
1063 static struct notifier_block dbg_reboot_notifier = {
1064 	.notifier_call		= dbg_notify_reboot,
1065 	.next			= NULL,
1066 	.priority		= INT_MAX,
1067 };
1068 
1069 static void kgdb_register_callbacks(void)
1070 {
1071 	if (!kgdb_io_module_registered) {
1072 		kgdb_io_module_registered = 1;
1073 		kgdb_arch_init();
1074 		if (!dbg_is_early)
1075 			kgdb_arch_late();
1076 		register_module_notifier(&dbg_module_load_nb);
1077 		register_reboot_notifier(&dbg_reboot_notifier);
1078 #ifdef CONFIG_MAGIC_SYSRQ
1079 		register_sysrq_key('g', &sysrq_dbg_op);
1080 #endif
1081 		if (kgdb_use_con && !kgdb_con_registered) {
1082 			register_console(&kgdbcons);
1083 			kgdb_con_registered = 1;
1084 		}
1085 	}
1086 }
1087 
1088 static void kgdb_unregister_callbacks(void)
1089 {
1090 	/*
1091 	 * When this routine is called KGDB should unregister from
1092 	 * handlers and clean up, making sure it is not handling any
1093 	 * break exceptions at the time.
1094 	 */
1095 	if (kgdb_io_module_registered) {
1096 		kgdb_io_module_registered = 0;
1097 		unregister_reboot_notifier(&dbg_reboot_notifier);
1098 		unregister_module_notifier(&dbg_module_load_nb);
1099 		kgdb_arch_exit();
1100 #ifdef CONFIG_MAGIC_SYSRQ
1101 		unregister_sysrq_key('g', &sysrq_dbg_op);
1102 #endif
1103 		if (kgdb_con_registered) {
1104 			unregister_console(&kgdbcons);
1105 			kgdb_con_registered = 0;
1106 		}
1107 	}
1108 }
1109 
1110 /**
1111  *	kgdb_register_io_module - register KGDB IO module
1112  *	@new_dbg_io_ops: the io ops vector
1113  *
1114  *	Register it with the KGDB core.
1115  */
1116 int kgdb_register_io_module(struct kgdb_io *new_dbg_io_ops)
1117 {
1118 	struct kgdb_io *old_dbg_io_ops;
1119 	int err;
1120 
1121 	spin_lock(&kgdb_registration_lock);
1122 
1123 	old_dbg_io_ops = dbg_io_ops;
1124 	if (old_dbg_io_ops) {
1125 		if (!old_dbg_io_ops->deinit) {
1126 			spin_unlock(&kgdb_registration_lock);
1127 
1128 			pr_err("KGDB I/O driver %s can't replace %s.\n",
1129 				new_dbg_io_ops->name, old_dbg_io_ops->name);
1130 			return -EBUSY;
1131 		}
1132 		pr_info("Replacing I/O driver %s with %s\n",
1133 			old_dbg_io_ops->name, new_dbg_io_ops->name);
1134 	}
1135 
1136 	if (new_dbg_io_ops->init) {
1137 		err = new_dbg_io_ops->init();
1138 		if (err) {
1139 			spin_unlock(&kgdb_registration_lock);
1140 			return err;
1141 		}
1142 	}
1143 
1144 	dbg_io_ops = new_dbg_io_ops;
1145 
1146 	spin_unlock(&kgdb_registration_lock);
1147 
1148 	if (old_dbg_io_ops) {
1149 		old_dbg_io_ops->deinit();
1150 		return 0;
1151 	}
1152 
1153 	pr_info("Registered I/O driver %s\n", new_dbg_io_ops->name);
1154 
1155 	/* Arm KGDB now. */
1156 	kgdb_register_callbacks();
1157 
1158 	if (kgdb_break_asap &&
1159 	    (!dbg_is_early || IS_ENABLED(CONFIG_ARCH_HAS_EARLY_DEBUG)))
1160 		kgdb_initial_breakpoint();
1161 
1162 	return 0;
1163 }
1164 EXPORT_SYMBOL_GPL(kgdb_register_io_module);
1165 
1166 /**
1167  *	kgdb_unregister_io_module - unregister KGDB IO module
1168  *	@old_dbg_io_ops: the io ops vector
1169  *
1170  *	Unregister it with the KGDB core.
1171  */
1172 void kgdb_unregister_io_module(struct kgdb_io *old_dbg_io_ops)
1173 {
1174 	BUG_ON(kgdb_connected);
1175 
1176 	/*
1177 	 * KGDB is no longer able to communicate out, so
1178 	 * unregister our callbacks and reset state.
1179 	 */
1180 	kgdb_unregister_callbacks();
1181 
1182 	spin_lock(&kgdb_registration_lock);
1183 
1184 	WARN_ON_ONCE(dbg_io_ops != old_dbg_io_ops);
1185 	dbg_io_ops = NULL;
1186 
1187 	spin_unlock(&kgdb_registration_lock);
1188 
1189 	if (old_dbg_io_ops->deinit)
1190 		old_dbg_io_ops->deinit();
1191 
1192 	pr_info("Unregistered I/O driver %s, debugger disabled\n",
1193 		old_dbg_io_ops->name);
1194 }
1195 EXPORT_SYMBOL_GPL(kgdb_unregister_io_module);
1196 
1197 int dbg_io_get_char(void)
1198 {
1199 	int ret = dbg_io_ops->read_char();
1200 	if (ret == NO_POLL_CHAR)
1201 		return -1;
1202 	if (!dbg_kdb_mode)
1203 		return ret;
1204 	if (ret == 127)
1205 		return 8;
1206 	return ret;
1207 }
1208 
1209 /**
1210  * kgdb_breakpoint - generate breakpoint exception
1211  *
1212  * This function will generate a breakpoint exception.  It is used at the
1213  * beginning of a program to sync up with a debugger and can be used
1214  * otherwise as a quick means to stop program execution and "break" into
1215  * the debugger.
1216  */
1217 noinline void kgdb_breakpoint(void)
1218 {
1219 	atomic_inc(&kgdb_setting_breakpoint);
1220 	wmb(); /* Sync point before breakpoint */
1221 	arch_kgdb_breakpoint();
1222 	wmb(); /* Sync point after breakpoint */
1223 	atomic_dec(&kgdb_setting_breakpoint);
1224 }
1225 EXPORT_SYMBOL_GPL(kgdb_breakpoint);
1226 
1227 static int __init opt_kgdb_wait(char *str)
1228 {
1229 	kgdb_break_asap = 1;
1230 
1231 	kdb_init(KDB_INIT_EARLY);
1232 	if (kgdb_io_module_registered &&
1233 	    IS_ENABLED(CONFIG_ARCH_HAS_EARLY_DEBUG))
1234 		kgdb_initial_breakpoint();
1235 
1236 	return 0;
1237 }
1238 
1239 early_param("kgdbwait", opt_kgdb_wait);
1240