nmi.h 5.1 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6
/*
 *  linux/include/linux/nmi.h
 */
#ifndef LINUX_NMI_H
#define LINUX_NMI_H

7
#include <linux/sched.h>
L
Linus Torvalds 已提交
8
#include <asm/irq.h>
9 10 11
#if defined(CONFIG_HAVE_NMI_WATCHDOG)
#include <asm/nmi.h>
#endif
L
Linus Torvalds 已提交
12

13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46
#ifdef CONFIG_LOCKUP_DETECTOR
extern void touch_softlockup_watchdog_sched(void);
extern void touch_softlockup_watchdog(void);
extern void touch_softlockup_watchdog_sync(void);
extern void touch_all_softlockup_watchdogs(void);
extern unsigned int  softlockup_panic;
extern unsigned int  hardlockup_panic;
void lockup_detector_init(void);
#else
static inline void touch_softlockup_watchdog_sched(void)
{
}
static inline void touch_softlockup_watchdog(void)
{
}
static inline void touch_softlockup_watchdog_sync(void)
{
}
static inline void touch_all_softlockup_watchdogs(void)
{
}
static inline void lockup_detector_init(void)
{
}
#endif

#ifdef CONFIG_DETECT_HUNG_TASK
void reset_hung_task_detector(void);
#else
static inline void reset_hung_task_detector(void)
{
}
#endif

47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63
/*
 * The run state of the lockup detectors is controlled by the content of the
 * 'watchdog_enabled' variable. Each lockup detector has its dedicated bit -
 * bit 0 for the hard lockup detector and bit 1 for the soft lockup detector.
 *
 * 'watchdog_user_enabled', 'nmi_watchdog_enabled' and 'soft_watchdog_enabled'
 * are variables that are only used as an 'interface' between the parameters
 * in /proc/sys/kernel and the internal state bits in 'watchdog_enabled'. The
 * 'watchdog_thresh' variable is handled differently because its value is not
 * boolean, and the lockup detectors are 'suspended' while 'watchdog_thresh'
 * is equal zero.
 */
#define NMI_WATCHDOG_ENABLED_BIT   0
#define SOFT_WATCHDOG_ENABLED_BIT  1
#define NMI_WATCHDOG_ENABLED      (1 << NMI_WATCHDOG_ENABLED_BIT)
#define SOFT_WATCHDOG_ENABLED     (1 << SOFT_WATCHDOG_ENABLED_BIT)

64 65 66 67 68 69 70 71 72 73 74 75
#if defined(CONFIG_HARDLOCKUP_DETECTOR)
extern void hardlockup_detector_disable(void);
#else
static inline void hardlockup_detector_disable(void) {}
#endif

#if defined(CONFIG_HARDLOCKUP_DETECTOR) || defined(CONFIG_HAVE_NMI_WATCHDOG)
extern void arch_touch_nmi_watchdog(void);
#else
static inline void arch_touch_nmi_watchdog(void) {}
#endif

L
Linus Torvalds 已提交
76 77 78 79 80 81 82
/**
 * touch_nmi_watchdog - restart NMI watchdog timeout.
 * 
 * If the architecture supports the NMI watchdog, touch_nmi_watchdog()
 * may be used to reset the timeout - for code which intentionally
 * disables interrupts for a long time. This call is stateless.
 */
83 84
static inline void touch_nmi_watchdog(void)
{
85
	arch_touch_nmi_watchdog();
86 87
	touch_softlockup_watchdog();
}
88

89 90 91 92 93
/*
 * Create trigger_all_cpu_backtrace() out of the arch-provided
 * base function. Return whether such support was available,
 * to allow calling code to fall back to some other mechanism:
 */
94
#ifdef arch_trigger_cpumask_backtrace
95 96
static inline bool trigger_all_cpu_backtrace(void)
{
97
	arch_trigger_cpumask_backtrace(cpu_online_mask, false);
98 99
	return true;
}
100

101 102
static inline bool trigger_allbutself_cpu_backtrace(void)
{
103 104 105 106 107 108 109 110 111 112 113 114 115
	arch_trigger_cpumask_backtrace(cpu_online_mask, true);
	return true;
}

static inline bool trigger_cpumask_backtrace(struct cpumask *mask)
{
	arch_trigger_cpumask_backtrace(mask, false);
	return true;
}

static inline bool trigger_single_cpu_backtrace(int cpu)
{
	arch_trigger_cpumask_backtrace(cpumask_of(cpu), false);
116 117
	return true;
}
118 119

/* generic implementation */
120 121
void nmi_trigger_cpumask_backtrace(const cpumask_t *mask,
				   bool exclude_self,
122 123 124
				   void (*raise)(cpumask_t *mask));
bool nmi_cpu_backtrace(struct pt_regs *regs);

125 126 127 128 129
#else
static inline bool trigger_all_cpu_backtrace(void)
{
	return false;
}
130 131 132 133
static inline bool trigger_allbutself_cpu_backtrace(void)
{
	return false;
}
134 135 136 137 138 139 140 141
static inline bool trigger_cpumask_backtrace(struct cpumask *mask)
{
	return false;
}
static inline bool trigger_single_cpu_backtrace(int cpu)
{
	return false;
}
A
Andrew Morton 已提交
142 143
#endif

144
#ifdef CONFIG_LOCKUP_DETECTOR
145
u64 hw_nmi_get_sample_period(int watchdog_thresh);
146 147
extern int nmi_watchdog_enabled;
extern int soft_watchdog_enabled;
148
extern int watchdog_user_enabled;
149
extern int watchdog_thresh;
150
extern unsigned long watchdog_enabled;
151
extern unsigned long *watchdog_cpumask_bits;
152
extern atomic_t watchdog_park_in_progress;
153
#ifdef CONFIG_SMP
154
extern int sysctl_softlockup_all_cpu_backtrace;
155
extern int sysctl_hardlockup_all_cpu_backtrace;
156 157 158 159 160
#else
#define sysctl_softlockup_all_cpu_backtrace 0
#define sysctl_hardlockup_all_cpu_backtrace 0
#endif
extern bool is_hardlockup(void);
161
struct ctl_table;
162 163 164 165 166 167 168 169
extern int proc_watchdog(struct ctl_table *, int ,
			 void __user *, size_t *, loff_t *);
extern int proc_nmi_watchdog(struct ctl_table *, int ,
			     void __user *, size_t *, loff_t *);
extern int proc_soft_watchdog(struct ctl_table *, int ,
			      void __user *, size_t *, loff_t *);
extern int proc_watchdog_thresh(struct ctl_table *, int ,
				void __user *, size_t *, loff_t *);
170 171
extern int proc_watchdog_cpumask(struct ctl_table *, int,
				 void __user *, size_t *, loff_t *);
172 173
extern int lockup_detector_suspend(void);
extern void lockup_detector_resume(void);
174
#else
175
static inline int lockup_detector_suspend(void)
176 177 178 179
{
	return 0;
}

180
static inline void lockup_detector_resume(void)
181 182
{
}
183 184
#endif

185 186 187 188
#ifdef CONFIG_HAVE_ACPI_APEI_NMI
#include <asm/nmi.h>
#endif

L
Linus Torvalds 已提交
189
#endif