bb81a09e55
When a spinlock lockup occurs, arrange for the NMI code to emit an all-cpu backtrace, so we get to see which CPU is holding the lock, and where. Cc: Andi Kleen <ak@muc.de> Cc: Ingo Molnar <mingo@elte.hu> Cc: Badari Pulavarty <pbadari@us.ibm.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Andi Kleen <ak@suse.de>
53 lines
1.4 KiB
C
53 lines
1.4 KiB
C
/*
|
|
* linux/include/asm-i386/nmi.h
|
|
*/
|
|
#ifndef ASM_NMI_H
|
|
#define ASM_NMI_H
|
|
|
|
#include <linux/pm.h>
|
|
#include <asm/irq.h>
|
|
|
|
#ifdef ARCH_HAS_NMI_WATCHDOG
|
|
|
|
/**
|
|
* do_nmi_callback
|
|
*
|
|
* Check to see if a callback exists and execute it. Return 1
|
|
* if the handler exists and was handled successfully.
|
|
*/
|
|
int do_nmi_callback(struct pt_regs *regs, int cpu);
|
|
|
|
extern int nmi_watchdog_enabled;
|
|
extern int avail_to_resrv_perfctr_nmi_bit(unsigned int);
|
|
extern int avail_to_resrv_perfctr_nmi(unsigned int);
|
|
extern int reserve_perfctr_nmi(unsigned int);
|
|
extern void release_perfctr_nmi(unsigned int);
|
|
extern int reserve_evntsel_nmi(unsigned int);
|
|
extern void release_evntsel_nmi(unsigned int);
|
|
|
|
extern void setup_apic_nmi_watchdog (void *);
|
|
extern void stop_apic_nmi_watchdog (void *);
|
|
extern void disable_timer_nmi_watchdog(void);
|
|
extern void enable_timer_nmi_watchdog(void);
|
|
extern int nmi_watchdog_tick (struct pt_regs * regs, unsigned reason);
|
|
|
|
extern atomic_t nmi_active;
|
|
extern unsigned int nmi_watchdog;
|
|
#define NMI_DEFAULT -1
|
|
#define NMI_NONE 0
|
|
#define NMI_IO_APIC 1
|
|
#define NMI_LOCAL_APIC 2
|
|
#define NMI_INVALID 3
|
|
|
|
struct ctl_table;
|
|
struct file;
|
|
extern int proc_nmi_enabled(struct ctl_table *, int , struct file *,
|
|
void __user *, size_t *, loff_t *);
|
|
extern int unknown_nmi_panic;
|
|
|
|
void __trigger_all_cpu_backtrace(void);
|
|
#define trigger_all_cpu_backtrace() __trigger_all_cpu_backtrace()
|
|
|
|
#endif
|
|
|
|
#endif /* ASM_NMI_H */
|