ipmi_si_intf.c 62.7 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0+
L
Linus Torvalds 已提交
2 3 4 5 6 7 8 9 10 11 12
/*
 * ipmi_si.c
 *
 * The interface to the IPMI driver for the system interfaces (KCS, SMIC,
 * BT).
 *
 * Author: MontaVista Software, Inc.
 *         Corey Minyard <minyard@mvista.com>
 *         source@mvista.com
 *
 * Copyright 2002 MontaVista Software Inc.
13
 * Copyright 2006 IBM Corp., Christian Krafft <krafft@de.ibm.com>
L
Linus Torvalds 已提交
14 15 16 17 18 19 20 21 22 23 24
 */

/*
 * This file holds the "policy" for the interface to the SMI state
 * machine.  It does the configuration, handles timers and interrupts,
 * and drives the real SMI state machine.
 */

#include <linux/module.h>
#include <linux/moduleparam.h>
#include <linux/sched.h>
25
#include <linux/seq_file.h>
L
Linus Torvalds 已提交
26 27 28 29 30 31
#include <linux/timer.h>
#include <linux/errno.h>
#include <linux/spinlock.h>
#include <linux/slab.h>
#include <linux/delay.h>
#include <linux/list.h>
32
#include <linux/notifier.h>
33
#include <linux/mutex.h>
M
Matt Domsch 已提交
34
#include <linux/kthread.h>
L
Linus Torvalds 已提交
35 36 37
#include <asm/irq.h>
#include <linux/interrupt.h>
#include <linux/rcupdate.h>
38
#include <linux/ipmi.h>
L
Linus Torvalds 已提交
39
#include <linux/ipmi_smi.h>
40
#include "ipmi_si.h"
41 42
#include <linux/string.h>
#include <linux/ctype.h>
43

44
#define PFX "ipmi_si: "
L
Linus Torvalds 已提交
45 46 47 48 49 50 51 52 53

/* Measure times between events in the driver. */
#undef DEBUG_TIMING

/* Call every 10 ms. */
#define SI_TIMEOUT_TIME_USEC	10000
#define SI_USEC_PER_JIFFY	(1000000/HZ)
#define SI_TIMEOUT_JIFFIES	(SI_TIMEOUT_TIME_USEC/SI_USEC_PER_JIFFY)
#define SI_SHORT_TIMEOUT_USEC  250 /* .25ms when the SM request a
54
				      short timeout */
L
Linus Torvalds 已提交
55 56 57 58 59 60 61

enum si_intf_state {
	SI_NORMAL,
	SI_GETTING_FLAGS,
	SI_GETTING_EVENTS,
	SI_CLEARING_FLAGS,
	SI_GETTING_MESSAGES,
62 63
	SI_CHECKING_ENABLES,
	SI_SETTING_ENABLES
L
Linus Torvalds 已提交
64 65 66
	/* FIXME - add watchdog stuff. */
};

67 68 69 70 71
/* Some BT-specific defines we need here. */
#define IPMI_BT_INTMASK_REG		2
#define IPMI_BT_INTMASK_CLEAR_IRQ_BIT	2
#define IPMI_BT_INTMASK_ENABLE_IRQ_BIT	1

72
static const char * const si_to_str[] = { "invalid", "kcs", "smic", "bt" };
L
Linus Torvalds 已提交
73

74 75
static int initialized;

76 77 78
/*
 * Indexes into stats[] in smi_info below.
 */
79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99
enum si_stat_indexes {
	/*
	 * Number of times the driver requested a timer while an operation
	 * was in progress.
	 */
	SI_STAT_short_timeouts = 0,

	/*
	 * Number of times the driver requested a timer while nothing was in
	 * progress.
	 */
	SI_STAT_long_timeouts,

	/* Number of times the interface was idle while being polled. */
	SI_STAT_idles,

	/* Number of interrupts the driver handled. */
	SI_STAT_interrupts,

	/* Number of time the driver got an ATTN from the hardware. */
	SI_STAT_attentions,
100

101 102 103 104 105 106 107 108 109 110 111 112 113 114 115
	/* Number of times the driver requested flags from the hardware. */
	SI_STAT_flag_fetches,

	/* Number of times the hardware didn't follow the state machine. */
	SI_STAT_hosed_count,

	/* Number of completed messages. */
	SI_STAT_complete_transactions,

	/* Number of IPMI events received from the hardware. */
	SI_STAT_events,

	/* Number of watchdog pretimeouts. */
	SI_STAT_watchdog_pretimeouts,

116
	/* Number of asynchronous messages received. */
117 118 119 120 121 122
	SI_STAT_incoming_messages,


	/* This *must* remain last, add new values above this. */
	SI_NUM_STATS
};
123

124
struct smi_info {
C
Corey Minyard 已提交
125
	int                    intf_num;
126
	struct ipmi_smi        *intf;
L
Linus Torvalds 已提交
127
	struct si_sm_data      *si_sm;
128
	const struct si_sm_handlers *handlers;
L
Linus Torvalds 已提交
129
	spinlock_t             si_lock;
130
	struct ipmi_smi_msg    *waiting_msg;
L
Linus Torvalds 已提交
131 132 133
	struct ipmi_smi_msg    *curr_msg;
	enum si_intf_state     si_state;

134 135 136 137
	/*
	 * Used to handle the various types of I/O that can occur with
	 * IPMI
	 */
L
Linus Torvalds 已提交
138 139
	struct si_sm_io io;

140 141 142 143 144
	/*
	 * Per-OEM handler, called from handle_flags().  Returns 1
	 * when handle_flags() needs to be re-run or 0 indicating it
	 * set si_state itself.
	 */
145 146
	int (*oem_data_avail_handler)(struct smi_info *smi_info);

147 148 149 150 151
	/*
	 * Flags from the last GET_MSG_FLAGS command, used when an ATTN
	 * is set to hold the flags until we are done handling everything
	 * from the flags.
	 */
L
Linus Torvalds 已提交
152 153 154
#define RECEIVE_MSG_AVAIL	0x01
#define EVENT_MSG_BUFFER_FULL	0x02
#define WDT_PRE_TIMEOUT_INT	0x08
155 156 157 158
#define OEM0_DATA_AVAIL     0x20
#define OEM1_DATA_AVAIL     0x40
#define OEM2_DATA_AVAIL     0x80
#define OEM_DATA_AVAIL      (OEM0_DATA_AVAIL | \
159 160
			     OEM1_DATA_AVAIL | \
			     OEM2_DATA_AVAIL)
L
Linus Torvalds 已提交
161 162
	unsigned char       msg_flags;

163
	/* Does the BMC have an event buffer? */
C
Corey Minyard 已提交
164
	bool		    has_event_buffer;
165

166 167 168 169
	/*
	 * If set to true, this will request events the next time the
	 * state machine is idle.
	 */
L
Linus Torvalds 已提交
170 171
	atomic_t            req_events;

172 173 174 175 176
	/*
	 * If true, run the state machine to completion on every send
	 * call.  Generally used after a panic to make sure stuff goes
	 * out.
	 */
C
Corey Minyard 已提交
177
	bool                run_to_completion;
L
Linus Torvalds 已提交
178 179 180 181

	/* The timer for this si. */
	struct timer_list   si_timer;

182 183 184
	/* This flag is set, if the timer can be set */
	bool		    timer_can_start;

185 186 187
	/* This flag is set, if the timer is running (timer_pending() isn't enough) */
	bool		    timer_running;

L
Linus Torvalds 已提交
188 189 190
	/* The time (in jiffies) the last timeout occurred at. */
	unsigned long       last_timeout_jiffies;

191 192 193
	/* Are we waiting for the events, pretimeouts, received msgs? */
	atomic_t            need_watch;

194 195 196 197 198 199
	/*
	 * The driver will disable interrupts when it gets into a
	 * situation where it cannot handle messages due to lack of
	 * memory.  Once that situation clears up, it will re-enable
	 * interrupts.
	 */
C
Corey Minyard 已提交
200
	bool interrupt_disabled;
L
Linus Torvalds 已提交
201

202 203 204 205 206
	/*
	 * Does the BMC support events?
	 */
	bool supports_event_msg_buff;

207
	/*
208 209 210 211 212 213 214
	 * Can we disable interrupts the global enables receive irq
	 * bit?  There are currently two forms of brokenness, some
	 * systems cannot disable the bit (which is technically within
	 * the spec but a bad idea) and some systems have the bit
	 * forced to zero even though interrupts work (which is
	 * clearly outside the spec).  The next bool tells which form
	 * of brokenness is present.
215
	 */
216 217 218 219 220 221 222
	bool cannot_disable_irq;

	/*
	 * Some systems are broken and cannot set the irq enable
	 * bit, even if they support interrupts.
	 */
	bool irq_enable_broken;
223

224 225 226 227 228
	/*
	 * Did we get an attention that we did not handle?
	 */
	bool got_attn;

229
	/* From the get device id response... */
230
	struct ipmi_device_id device_id;
L
Linus Torvalds 已提交
231

232
	/* Default driver model device. */
233 234
	struct platform_device *pdev;

235 236 237
	/* Have we added the device group to the device? */
	bool dev_group_added;

C
Corey Minyard 已提交
238 239 240
	/* Have we added the platform device? */
	bool pdev_registered;

L
Linus Torvalds 已提交
241
	/* Counters and things for the proc filesystem. */
242
	atomic_t stats[SI_NUM_STATS];
C
Corey Minyard 已提交
243

244
	struct task_struct *thread;
245 246

	struct list_head link;
L
Linus Torvalds 已提交
247 248
};

249 250 251 252 253
#define smi_inc_stat(smi, stat) \
	atomic_inc(&(smi)->stats[SI_STAT_ ## stat])
#define smi_get_stat(smi, stat) \
	((unsigned int) atomic_read(&(smi)->stats[SI_STAT_ ## stat]))

254 255
#define IPMI_MAX_INTFS 4
static int force_kipmid[IPMI_MAX_INTFS];
256 257
static int num_force_kipmid;

258
static unsigned int kipmid_max_busy_us[IPMI_MAX_INTFS];
259 260
static int num_max_busy_us;

C
Corey Minyard 已提交
261
static bool unload_when_empty = true;
262

263
static int try_smi_init(struct smi_info *smi);
C
Corey Minyard 已提交
264 265
static void shutdown_one_si(struct smi_info *smi_info);
static void cleanup_one_si(struct smi_info *smi_info);
266
static void cleanup_ipmi_si(void);
267

268 269 270
#ifdef DEBUG_TIMING
void debug_timestamp(char *msg)
{
271
	struct timespec64 t;
272

273 274
	getnstimeofday64(&t);
	pr_debug("**%s: %lld.%9.9ld\n", msg, (long long) t.tv_sec, t.tv_nsec);
275 276 277 278 279
}
#else
#define debug_timestamp(x)
#endif

280
static ATOMIC_NOTIFIER_HEAD(xaction_notifier_list);
281
static int register_xaction_notifier(struct notifier_block *nb)
282
{
283
	return atomic_notifier_chain_register(&xaction_notifier_list, nb);
284 285
}

L
Linus Torvalds 已提交
286 287 288
static void deliver_recv_msg(struct smi_info *smi_info,
			     struct ipmi_smi_msg *msg)
{
289
	/* Deliver the message to the upper layer. */
C
Corey Minyard 已提交
290 291 292 293
	if (smi_info->intf)
		ipmi_smi_msg_received(smi_info->intf, msg);
	else
		ipmi_free_smi_msg(msg);
L
Linus Torvalds 已提交
294 295
}

C
Corey Minyard 已提交
296
static void return_hosed_msg(struct smi_info *smi_info, int cCode)
L
Linus Torvalds 已提交
297 298 299
{
	struct ipmi_smi_msg *msg = smi_info->curr_msg;

C
Corey Minyard 已提交
300 301 302 303
	if (cCode < 0 || cCode > IPMI_ERR_UNSPECIFIED)
		cCode = IPMI_ERR_UNSPECIFIED;
	/* else use it as is */

L
Lucas De Marchi 已提交
304
	/* Make it a response */
L
Linus Torvalds 已提交
305 306
	msg->rsp[0] = msg->data[0] | 4;
	msg->rsp[1] = msg->data[1];
C
Corey Minyard 已提交
307
	msg->rsp[2] = cCode;
L
Linus Torvalds 已提交
308 309 310 311 312 313 314 315 316 317
	msg->rsp_size = 3;

	smi_info->curr_msg = NULL;
	deliver_recv_msg(smi_info, msg);
}

static enum si_sm_result start_next_msg(struct smi_info *smi_info)
{
	int              rv;

318
	if (!smi_info->waiting_msg) {
L
Linus Torvalds 已提交
319 320 321 322 323
		smi_info->curr_msg = NULL;
		rv = SI_SM_IDLE;
	} else {
		int err;

324 325
		smi_info->curr_msg = smi_info->waiting_msg;
		smi_info->waiting_msg = NULL;
326
		debug_timestamp("Start2");
327 328
		err = atomic_notifier_call_chain(&xaction_notifier_list,
				0, smi_info);
329 330 331 332
		if (err & NOTIFY_STOP_MASK) {
			rv = SI_SM_CALL_WITHOUT_DELAY;
			goto out;
		}
L
Linus Torvalds 已提交
333 334 335 336
		err = smi_info->handlers->start_transaction(
			smi_info->si_sm,
			smi_info->curr_msg->data,
			smi_info->curr_msg->data_size);
337
		if (err)
C
Corey Minyard 已提交
338
			return_hosed_msg(smi_info, err);
L
Linus Torvalds 已提交
339 340 341

		rv = SI_SM_CALL_WITHOUT_DELAY;
	}
342
out:
L
Linus Torvalds 已提交
343 344 345
	return rv;
}

346 347
static void smi_mod_timer(struct smi_info *smi_info, unsigned long new_val)
{
348 349
	if (!smi_info->timer_can_start)
		return;
350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368
	smi_info->last_timeout_jiffies = jiffies;
	mod_timer(&smi_info->si_timer, new_val);
	smi_info->timer_running = true;
}

/*
 * Start a new message and (re)start the timer and thread.
 */
static void start_new_msg(struct smi_info *smi_info, unsigned char *msg,
			  unsigned int size)
{
	smi_mod_timer(smi_info, jiffies + SI_TIMEOUT_JIFFIES);

	if (smi_info->thread)
		wake_up_process(smi_info->thread);

	smi_info->handlers->start_transaction(smi_info->si_sm, msg, size);
}

369
static void start_check_enables(struct smi_info *smi_info)
C
Corey Minyard 已提交
370 371 372 373 374 375
{
	unsigned char msg[2];

	msg[0] = (IPMI_NETFN_APP_REQUEST << 2);
	msg[1] = IPMI_GET_BMC_GLOBAL_ENABLES_CMD;

376
	start_new_msg(smi_info, msg, 2);
377
	smi_info->si_state = SI_CHECKING_ENABLES;
C
Corey Minyard 已提交
378 379
}

380
static void start_clear_flags(struct smi_info *smi_info)
L
Linus Torvalds 已提交
381 382 383 384 385 386 387 388
{
	unsigned char msg[3];

	/* Make sure the watchdog pre-timeout flag is not set at startup. */
	msg[0] = (IPMI_NETFN_APP_REQUEST << 2);
	msg[1] = IPMI_CLEAR_MSG_FLAGS_CMD;
	msg[2] = WDT_PRE_TIMEOUT_INT;

389
	start_new_msg(smi_info, msg, 3);
L
Linus Torvalds 已提交
390 391 392
	smi_info->si_state = SI_CLEARING_FLAGS;
}

C
Corey Minyard 已提交
393 394 395 396 397 398
static void start_getting_msg_queue(struct smi_info *smi_info)
{
	smi_info->curr_msg->data[0] = (IPMI_NETFN_APP_REQUEST << 2);
	smi_info->curr_msg->data[1] = IPMI_GET_MSG_CMD;
	smi_info->curr_msg->data_size = 2;

399 400
	start_new_msg(smi_info, smi_info->curr_msg->data,
		      smi_info->curr_msg->data_size);
C
Corey Minyard 已提交
401 402 403 404 405 406 407 408 409
	smi_info->si_state = SI_GETTING_MESSAGES;
}

static void start_getting_events(struct smi_info *smi_info)
{
	smi_info->curr_msg->data[0] = (IPMI_NETFN_APP_REQUEST << 2);
	smi_info->curr_msg->data[1] = IPMI_READ_EVENT_MSG_BUFFER_CMD;
	smi_info->curr_msg->data_size = 2;

410 411
	start_new_msg(smi_info, smi_info->curr_msg->data,
		      smi_info->curr_msg->data_size);
C
Corey Minyard 已提交
412 413 414
	smi_info->si_state = SI_GETTING_EVENTS;
}

415 416 417 418 419
/*
 * When we have a situtaion where we run out of memory and cannot
 * allocate messages, we just leave them in the BMC and run the system
 * polled until we can allocate some memory.  Once we have some
 * memory, we will re-enable the interrupt.
420 421 422
 *
 * Note that we cannot just use disable_irq(), since the interrupt may
 * be shared.
423
 */
424
static inline bool disable_si_irq(struct smi_info *smi_info)
L
Linus Torvalds 已提交
425
{
426
	if ((smi_info->io.irq) && (!smi_info->interrupt_disabled)) {
C
Corey Minyard 已提交
427
		smi_info->interrupt_disabled = true;
428
		start_check_enables(smi_info);
C
Corey Minyard 已提交
429
		return true;
L
Linus Torvalds 已提交
430
	}
C
Corey Minyard 已提交
431
	return false;
L
Linus Torvalds 已提交
432 433
}

C
Corey Minyard 已提交
434
static inline bool enable_si_irq(struct smi_info *smi_info)
L
Linus Torvalds 已提交
435
{
436
	if ((smi_info->io.irq) && (smi_info->interrupt_disabled)) {
C
Corey Minyard 已提交
437
		smi_info->interrupt_disabled = false;
438
		start_check_enables(smi_info);
C
Corey Minyard 已提交
439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455
		return true;
	}
	return false;
}

/*
 * Allocate a message.  If unable to allocate, start the interrupt
 * disable process and return NULL.  If able to allocate but
 * interrupts are disabled, free the message and return NULL after
 * starting the interrupt enable process.
 */
static struct ipmi_smi_msg *alloc_msg_handle_irq(struct smi_info *smi_info)
{
	struct ipmi_smi_msg *msg;

	msg = ipmi_alloc_smi_msg();
	if (!msg) {
456
		if (!disable_si_irq(smi_info))
C
Corey Minyard 已提交
457 458 459 460
			smi_info->si_state = SI_NORMAL;
	} else if (enable_si_irq(smi_info)) {
		ipmi_free_smi_msg(msg);
		msg = NULL;
L
Linus Torvalds 已提交
461
	}
C
Corey Minyard 已提交
462
	return msg;
L
Linus Torvalds 已提交
463 464 465 466
}

static void handle_flags(struct smi_info *smi_info)
{
467
retry:
L
Linus Torvalds 已提交
468 469
	if (smi_info->msg_flags & WDT_PRE_TIMEOUT_INT) {
		/* Watchdog pre-timeout */
470
		smi_inc_stat(smi_info, watchdog_pretimeouts);
L
Linus Torvalds 已提交
471

472
		start_clear_flags(smi_info);
L
Linus Torvalds 已提交
473
		smi_info->msg_flags &= ~WDT_PRE_TIMEOUT_INT;
C
Corey Minyard 已提交
474 475
		if (smi_info->intf)
			ipmi_smi_watchdog_pretimeout(smi_info->intf);
L
Linus Torvalds 已提交
476 477
	} else if (smi_info->msg_flags & RECEIVE_MSG_AVAIL) {
		/* Messages available. */
C
Corey Minyard 已提交
478 479
		smi_info->curr_msg = alloc_msg_handle_irq(smi_info);
		if (!smi_info->curr_msg)
L
Linus Torvalds 已提交
480 481
			return;

C
Corey Minyard 已提交
482
		start_getting_msg_queue(smi_info);
L
Linus Torvalds 已提交
483 484
	} else if (smi_info->msg_flags & EVENT_MSG_BUFFER_FULL) {
		/* Events available. */
C
Corey Minyard 已提交
485 486
		smi_info->curr_msg = alloc_msg_handle_irq(smi_info);
		if (!smi_info->curr_msg)
L
Linus Torvalds 已提交
487 488
			return;

C
Corey Minyard 已提交
489
		start_getting_events(smi_info);
490
	} else if (smi_info->msg_flags & OEM_DATA_AVAIL &&
491
		   smi_info->oem_data_avail_handler) {
492 493
		if (smi_info->oem_data_avail_handler(smi_info))
			goto retry;
494
	} else
L
Linus Torvalds 已提交
495 496 497
		smi_info->si_state = SI_NORMAL;
}

498 499 500 501 502 503
/*
 * Global enables we care about.
 */
#define GLOBAL_ENABLES_MASK (IPMI_BMC_EVT_MSG_BUFF | IPMI_BMC_RCV_MSG_INTR | \
			     IPMI_BMC_EVT_MSG_INTR)

504 505
static u8 current_global_enables(struct smi_info *smi_info, u8 base,
				 bool *irq_on)
506 507 508 509 510 511
{
	u8 enables = 0;

	if (smi_info->supports_event_msg_buff)
		enables |= IPMI_BMC_EVT_MSG_BUFF;

512
	if (((smi_info->io.irq && !smi_info->interrupt_disabled) ||
513 514
	     smi_info->cannot_disable_irq) &&
	    !smi_info->irq_enable_broken)
515 516 517
		enables |= IPMI_BMC_RCV_MSG_INTR;

	if (smi_info->supports_event_msg_buff &&
518
	    smi_info->io.irq && !smi_info->interrupt_disabled &&
519
	    !smi_info->irq_enable_broken)
520 521
		enables |= IPMI_BMC_EVT_MSG_INTR;

522 523
	*irq_on = enables & (IPMI_BMC_EVT_MSG_INTR | IPMI_BMC_RCV_MSG_INTR);

524 525 526
	return enables;
}

527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542
static void check_bt_irq(struct smi_info *smi_info, bool irq_on)
{
	u8 irqstate = smi_info->io.inputb(&smi_info->io, IPMI_BT_INTMASK_REG);

	irqstate &= IPMI_BT_INTMASK_ENABLE_IRQ_BIT;

	if ((bool)irqstate == irq_on)
		return;

	if (irq_on)
		smi_info->io.outputb(&smi_info->io, IPMI_BT_INTMASK_REG,
				     IPMI_BT_INTMASK_ENABLE_IRQ_BIT);
	else
		smi_info->io.outputb(&smi_info->io, IPMI_BT_INTMASK_REG, 0);
}

L
Linus Torvalds 已提交
543 544 545 546
static void handle_transaction_done(struct smi_info *smi_info)
{
	struct ipmi_smi_msg *msg;

547
	debug_timestamp("Done");
L
Linus Torvalds 已提交
548 549
	switch (smi_info->si_state) {
	case SI_NORMAL:
550
		if (!smi_info->curr_msg)
L
Linus Torvalds 已提交
551 552 553 554 555 556 557 558
			break;

		smi_info->curr_msg->rsp_size
			= smi_info->handlers->get_result(
				smi_info->si_sm,
				smi_info->curr_msg->rsp,
				IPMI_MAX_MSG_LENGTH);

559 560 561 562 563
		/*
		 * Do this here becase deliver_recv_msg() releases the
		 * lock, and a new message can be put in during the
		 * time the lock is released.
		 */
L
Linus Torvalds 已提交
564 565 566 567 568 569 570 571 572 573 574 575 576
		msg = smi_info->curr_msg;
		smi_info->curr_msg = NULL;
		deliver_recv_msg(smi_info, msg);
		break;

	case SI_GETTING_FLAGS:
	{
		unsigned char msg[4];
		unsigned int  len;

		/* We got the flags from the SMI, now handle them. */
		len = smi_info->handlers->get_result(smi_info->si_sm, msg, 4);
		if (msg[2] != 0) {
577
			/* Error fetching flags, just give up for now. */
L
Linus Torvalds 已提交
578 579
			smi_info->si_state = SI_NORMAL;
		} else if (len < 4) {
580 581 582 583
			/*
			 * Hmm, no flags.  That's technically illegal, but
			 * don't use uninitialized data.
			 */
L
Linus Torvalds 已提交
584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599
			smi_info->si_state = SI_NORMAL;
		} else {
			smi_info->msg_flags = msg[3];
			handle_flags(smi_info);
		}
		break;
	}

	case SI_CLEARING_FLAGS:
	{
		unsigned char msg[3];

		/* We cleared the flags. */
		smi_info->handlers->get_result(smi_info->si_sm, msg, 3);
		if (msg[2] != 0) {
			/* Error clearing flags */
600
			dev_warn(smi_info->io.dev,
601
				 "Error clearing flags: %2.2x\n", msg[2]);
L
Linus Torvalds 已提交
602
		}
603
		smi_info->si_state = SI_NORMAL;
L
Linus Torvalds 已提交
604 605 606 607 608 609 610 611 612 613 614
		break;
	}

	case SI_GETTING_EVENTS:
	{
		smi_info->curr_msg->rsp_size
			= smi_info->handlers->get_result(
				smi_info->si_sm,
				smi_info->curr_msg->rsp,
				IPMI_MAX_MSG_LENGTH);

615 616 617 618 619
		/*
		 * Do this here becase deliver_recv_msg() releases the
		 * lock, and a new message can be put in during the
		 * time the lock is released.
		 */
L
Linus Torvalds 已提交
620 621 622 623 624 625 626 627 628 629
		msg = smi_info->curr_msg;
		smi_info->curr_msg = NULL;
		if (msg->rsp[2] != 0) {
			/* Error getting event, probably done. */
			msg->done(msg);

			/* Take off the event flag. */
			smi_info->msg_flags &= ~EVENT_MSG_BUFFER_FULL;
			handle_flags(smi_info);
		} else {
630
			smi_inc_stat(smi_info, events);
L
Linus Torvalds 已提交
631

632 633 634 635 636 637
			/*
			 * Do this before we deliver the message
			 * because delivering the message releases the
			 * lock and something else can mess with the
			 * state.
			 */
L
Linus Torvalds 已提交
638 639 640 641 642 643 644 645 646 647 648 649 650 651 652
			handle_flags(smi_info);

			deliver_recv_msg(smi_info, msg);
		}
		break;
	}

	case SI_GETTING_MESSAGES:
	{
		smi_info->curr_msg->rsp_size
			= smi_info->handlers->get_result(
				smi_info->si_sm,
				smi_info->curr_msg->rsp,
				IPMI_MAX_MSG_LENGTH);

653 654 655 656 657
		/*
		 * Do this here becase deliver_recv_msg() releases the
		 * lock, and a new message can be put in during the
		 * time the lock is released.
		 */
L
Linus Torvalds 已提交
658 659 660 661 662 663 664 665 666 667
		msg = smi_info->curr_msg;
		smi_info->curr_msg = NULL;
		if (msg->rsp[2] != 0) {
			/* Error getting event, probably done. */
			msg->done(msg);

			/* Take off the msg flag. */
			smi_info->msg_flags &= ~RECEIVE_MSG_AVAIL;
			handle_flags(smi_info);
		} else {
668
			smi_inc_stat(smi_info, incoming_messages);
L
Linus Torvalds 已提交
669

670 671 672 673 674 675
			/*
			 * Do this before we deliver the message
			 * because delivering the message releases the
			 * lock and something else can mess with the
			 * state.
			 */
L
Linus Torvalds 已提交
676 677 678 679 680 681 682
			handle_flags(smi_info);

			deliver_recv_msg(smi_info, msg);
		}
		break;
	}

683
	case SI_CHECKING_ENABLES:
L
Linus Torvalds 已提交
684 685
	{
		unsigned char msg[4];
686
		u8 enables;
687
		bool irq_on;
L
Linus Torvalds 已提交
688 689 690 691

		/* We got the flags from the SMI, now handle them. */
		smi_info->handlers->get_result(smi_info->si_sm, msg, 4);
		if (msg[2] != 0) {
692
			dev_warn(smi_info->io.dev,
693
				 "Couldn't get irq info: %x.\n", msg[2]);
694
			dev_warn(smi_info->io.dev,
695
				 "Maybe ok, but ipmi might run very slowly.\n");
L
Linus Torvalds 已提交
696
			smi_info->si_state = SI_NORMAL;
697 698
			break;
		}
699
		enables = current_global_enables(smi_info, 0, &irq_on);
700
		if (smi_info->io.si_type == SI_BT)
701 702
			/* BT has its own interrupt enable bit. */
			check_bt_irq(smi_info, irq_on);
703 704
		if (enables != (msg[3] & GLOBAL_ENABLES_MASK)) {
			/* Enables are not correct, fix them. */
L
Linus Torvalds 已提交
705 706
			msg[0] = (IPMI_NETFN_APP_REQUEST << 2);
			msg[1] = IPMI_SET_BMC_GLOBAL_ENABLES_CMD;
707
			msg[2] = enables | (msg[3] & ~GLOBAL_ENABLES_MASK);
L
Linus Torvalds 已提交
708 709
			smi_info->handlers->start_transaction(
				smi_info->si_sm, msg, 3);
710 711 712 713 714 715 716
			smi_info->si_state = SI_SETTING_ENABLES;
		} else if (smi_info->supports_event_msg_buff) {
			smi_info->curr_msg = ipmi_alloc_smi_msg();
			if (!smi_info->curr_msg) {
				smi_info->si_state = SI_NORMAL;
				break;
			}
717
			start_getting_events(smi_info);
718 719
		} else {
			smi_info->si_state = SI_NORMAL;
L
Linus Torvalds 已提交
720 721 722 723
		}
		break;
	}

724
	case SI_SETTING_ENABLES:
L
Linus Torvalds 已提交
725 726 727 728
	{
		unsigned char msg[4];

		smi_info->handlers->get_result(smi_info->si_sm, msg, 4);
729
		if (msg[2] != 0)
730
			dev_warn(smi_info->io.dev,
731 732 733 734 735 736 737 738 739
				 "Could not set the global enables: 0x%x.\n",
				 msg[2]);

		if (smi_info->supports_event_msg_buff) {
			smi_info->curr_msg = ipmi_alloc_smi_msg();
			if (!smi_info->curr_msg) {
				smi_info->si_state = SI_NORMAL;
				break;
			}
740
			start_getting_events(smi_info);
C
Corey Minyard 已提交
741
		} else {
742
			smi_info->si_state = SI_NORMAL;
C
Corey Minyard 已提交
743 744 745
		}
		break;
	}
L
Linus Torvalds 已提交
746 747 748
	}
}

749 750 751 752 753
/*
 * Called on timeouts and events.  Timeouts should pass the elapsed
 * time, interrupts should pass in zero.  Must be called with
 * si_lock held and interrupts disabled.
 */
L
Linus Torvalds 已提交
754 755 756 757 758
static enum si_sm_result smi_event_handler(struct smi_info *smi_info,
					   int time)
{
	enum si_sm_result si_sm_result;

759
restart:
760 761 762 763 764 765 766 767
	/*
	 * There used to be a loop here that waited a little while
	 * (around 25us) before giving up.  That turned out to be
	 * pointless, the minimum delays I was seeing were in the 300us
	 * range, which is far too long to wait in an interrupt.  So
	 * we just run until the state machine tells us something
	 * happened or it needs a delay.
	 */
L
Linus Torvalds 已提交
768 769 770 771 772
	si_sm_result = smi_info->handlers->event(smi_info->si_sm, time);
	time = 0;
	while (si_sm_result == SI_SM_CALL_WITHOUT_DELAY)
		si_sm_result = smi_info->handlers->event(smi_info->si_sm, 0);

773
	if (si_sm_result == SI_SM_TRANSACTION_COMPLETE) {
774
		smi_inc_stat(smi_info, complete_transactions);
L
Linus Torvalds 已提交
775 776

		handle_transaction_done(smi_info);
777
		goto restart;
778
	} else if (si_sm_result == SI_SM_HOSED) {
779
		smi_inc_stat(smi_info, hosed_count);
L
Linus Torvalds 已提交
780

781 782 783 784
		/*
		 * Do the before return_hosed_msg, because that
		 * releases the lock.
		 */
L
Linus Torvalds 已提交
785 786
		smi_info->si_state = SI_NORMAL;
		if (smi_info->curr_msg != NULL) {
787 788 789 790 791
			/*
			 * If we were handling a user message, format
			 * a response to send to the upper layer to
			 * tell it about the error.
			 */
C
Corey Minyard 已提交
792
			return_hosed_msg(smi_info, IPMI_ERR_UNSPECIFIED);
L
Linus Torvalds 已提交
793
		}
794
		goto restart;
L
Linus Torvalds 已提交
795 796
	}

797 798 799 800
	/*
	 * We prefer handling attn over new messages.  But don't do
	 * this if there is not yet an upper layer to handle anything.
	 */
801 802
	if (likely(smi_info->intf) &&
	    (si_sm_result == SI_SM_ATTN || smi_info->got_attn)) {
L
Linus Torvalds 已提交
803 804
		unsigned char msg[2];

805 806 807 808 809 810 811 812 813
		if (smi_info->si_state != SI_NORMAL) {
			/*
			 * We got an ATTN, but we are doing something else.
			 * Handle the ATTN later.
			 */
			smi_info->got_attn = true;
		} else {
			smi_info->got_attn = false;
			smi_inc_stat(smi_info, attentions);
L
Linus Torvalds 已提交
814

815 816 817 818 819 820 821 822 823
			/*
			 * Got a attn, send down a get message flags to see
			 * what's causing it.  It would be better to handle
			 * this in the upper layer, but due to the way
			 * interrupts work with the SMI, that's not really
			 * possible.
			 */
			msg[0] = (IPMI_NETFN_APP_REQUEST << 2);
			msg[1] = IPMI_GET_MSG_FLAGS_CMD;
L
Linus Torvalds 已提交
824

825
			start_new_msg(smi_info, msg, 2);
826 827 828
			smi_info->si_state = SI_GETTING_FLAGS;
			goto restart;
		}
L
Linus Torvalds 已提交
829 830 831 832
	}

	/* If we are currently idle, try to start the next message. */
	if (si_sm_result == SI_SM_IDLE) {
833
		smi_inc_stat(smi_info, idles);
L
Linus Torvalds 已提交
834 835 836 837

		si_sm_result = start_next_msg(smi_info);
		if (si_sm_result != SI_SM_IDLE)
			goto restart;
838
	}
L
Linus Torvalds 已提交
839 840

	if ((si_sm_result == SI_SM_IDLE)
841 842 843 844 845
	    && (atomic_read(&smi_info->req_events))) {
		/*
		 * We are idle and the upper layer requested that I fetch
		 * events, so do so.
		 */
C
Corey Minyard 已提交
846
		atomic_set(&smi_info->req_events, 0);
L
Linus Torvalds 已提交
847

848 849 850 851 852 853
		/*
		 * Take this opportunity to check the interrupt and
		 * message enable state for the BMC.  The BMC can be
		 * asynchronously reset, and may thus get interrupts
		 * disable and messages disabled.
		 */
854
		if (smi_info->supports_event_msg_buff || smi_info->io.irq) {
855
			start_check_enables(smi_info);
856 857 858 859
		} else {
			smi_info->curr_msg = alloc_msg_handle_irq(smi_info);
			if (!smi_info->curr_msg)
				goto out;
L
Linus Torvalds 已提交
860

861 862
			start_getting_events(smi_info);
		}
L
Linus Torvalds 已提交
863 864
		goto restart;
	}
865 866 867 868 869 870 871

	if (si_sm_result == SI_SM_IDLE && smi_info->timer_running) {
		/* Ok it if fails, the timer will just go off. */
		if (del_timer(&smi_info->si_timer))
			smi_info->timer_running = false;
	}

872
out:
L
Linus Torvalds 已提交
873 874 875
	return si_sm_result;
}

876 877 878 879 880 881 882 883 884 885 886 887 888
static void check_start_timer_thread(struct smi_info *smi_info)
{
	if (smi_info->si_state == SI_NORMAL && smi_info->curr_msg == NULL) {
		smi_mod_timer(smi_info, jiffies + SI_TIMEOUT_JIFFIES);

		if (smi_info->thread)
			wake_up_process(smi_info->thread);

		start_next_msg(smi_info);
		smi_event_handler(smi_info, 0);
	}
}

889
static void flush_messages(void *send_info)
890
{
891
	struct smi_info *smi_info = send_info;
892 893 894 895 896 897 898 899 900 901 902 903 904
	enum si_sm_result result;

	/*
	 * Currently, this function is called only in run-to-completion
	 * mode.  This means we are single-threaded, no need for locks.
	 */
	result = smi_event_handler(smi_info, 0);
	while (result != SI_SM_IDLE) {
		udelay(SI_SHORT_TIMEOUT_USEC);
		result = smi_event_handler(smi_info, SI_SHORT_TIMEOUT_USEC);
	}
}

L
Linus Torvalds 已提交
905
static void sender(void                *send_info,
906
		   struct ipmi_smi_msg *msg)
L
Linus Torvalds 已提交
907 908 909 910
{
	struct smi_info   *smi_info = send_info;
	unsigned long     flags;

911
	debug_timestamp("Enqueue");
L
Linus Torvalds 已提交
912 913

	if (smi_info->run_to_completion) {
C
Corey Minyard 已提交
914
		/*
915 916
		 * If we are running to completion, start it.  Upper
		 * layer will call flush_messages to clear it out.
C
Corey Minyard 已提交
917
		 */
918
		smi_info->waiting_msg = msg;
L
Linus Torvalds 已提交
919 920 921
		return;
	}

C
Corey Minyard 已提交
922
	spin_lock_irqsave(&smi_info->si_lock, flags);
923 924 925 926 927 928 929 930 931
	/*
	 * The following two lines don't need to be under the lock for
	 * the lock's sake, but they do need SMP memory barriers to
	 * avoid getting things out of order.  We are already claiming
	 * the lock, anyway, so just do it under the lock to avoid the
	 * ordering problem.
	 */
	BUG_ON(smi_info->waiting_msg);
	smi_info->waiting_msg = msg;
932
	check_start_timer_thread(smi_info);
C
Corey Minyard 已提交
933
	spin_unlock_irqrestore(&smi_info->si_lock, flags);
L
Linus Torvalds 已提交
934 935
}

C
Corey Minyard 已提交
936
static void set_run_to_completion(void *send_info, bool i_run_to_completion)
L
Linus Torvalds 已提交
937 938 939 940
{
	struct smi_info   *smi_info = send_info;

	smi_info->run_to_completion = i_run_to_completion;
941 942
	if (i_run_to_completion)
		flush_messages(smi_info);
L
Linus Torvalds 已提交
943 944
}

945 946 947 948 949
/*
 * Use -1 in the nsec value of the busy waiting timespec to tell that
 * we are spinning in kipmid looking for something and not delaying
 * between checks
 */
950
static inline void ipmi_si_set_not_busy(struct timespec64 *ts)
951 952 953
{
	ts->tv_nsec = -1;
}
954
static inline int ipmi_si_is_busy(struct timespec64 *ts)
955 956 957 958
{
	return ts->tv_nsec != -1;
}

959 960
static inline int ipmi_thread_busy_wait(enum si_sm_result smi_result,
					const struct smi_info *smi_info,
961
					struct timespec64 *busy_until)
962 963 964 965 966 967 968 969
{
	unsigned int max_busy_us = 0;

	if (smi_info->intf_num < num_max_busy_us)
		max_busy_us = kipmid_max_busy_us[smi_info->intf_num];
	if (max_busy_us == 0 || smi_result != SI_SM_CALL_WITH_DELAY)
		ipmi_si_set_not_busy(busy_until);
	else if (!ipmi_si_is_busy(busy_until)) {
970 971
		getnstimeofday64(busy_until);
		timespec64_add_ns(busy_until, max_busy_us*NSEC_PER_USEC);
972
	} else {
973 974 975 976
		struct timespec64 now;

		getnstimeofday64(&now);
		if (unlikely(timespec64_compare(&now, busy_until) > 0)) {
977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993
			ipmi_si_set_not_busy(busy_until);
			return 0;
		}
	}
	return 1;
}


/*
 * A busy-waiting loop for speeding up IPMI operation.
 *
 * Lousy hardware makes this hard.  This is only enabled for systems
 * that are not BT and do not have interrupts.  It starts spinning
 * when an operation is complete or until max_busy tells it to stop
 * (if that is enabled).  See the paragraph on kimid_max_busy_us in
 * Documentation/IPMI.txt for details.
 */
C
Corey Minyard 已提交
994 995 996
static int ipmi_thread(void *data)
{
	struct smi_info *smi_info = data;
M
Matt Domsch 已提交
997
	unsigned long flags;
C
Corey Minyard 已提交
998
	enum si_sm_result smi_result;
999
	struct timespec64 busy_until;
C
Corey Minyard 已提交
1000

1001
	ipmi_si_set_not_busy(&busy_until);
1002
	set_user_nice(current, MAX_NICE);
M
Matt Domsch 已提交
1003
	while (!kthread_should_stop()) {
1004 1005
		int busy_wait;

C
Corey Minyard 已提交
1006
		spin_lock_irqsave(&(smi_info->si_lock), flags);
1007
		smi_result = smi_event_handler(smi_info, 0);
1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018

		/*
		 * If the driver is doing something, there is a possible
		 * race with the timer.  If the timer handler see idle,
		 * and the thread here sees something else, the timer
		 * handler won't restart the timer even though it is
		 * required.  So start it here if necessary.
		 */
		if (smi_result != SI_SM_IDLE && !smi_info->timer_running)
			smi_mod_timer(smi_info, jiffies + SI_TIMEOUT_JIFFIES);

C
Corey Minyard 已提交
1019
		spin_unlock_irqrestore(&(smi_info->si_lock), flags);
1020 1021
		busy_wait = ipmi_thread_busy_wait(smi_result, smi_info,
						  &busy_until);
1022 1023
		if (smi_result == SI_SM_CALL_WITHOUT_DELAY)
			; /* do nothing */
1024
		else if (smi_result == SI_SM_CALL_WITH_DELAY && busy_wait)
1025
			schedule();
1026 1027 1028 1029 1030 1031 1032 1033 1034
		else if (smi_result == SI_SM_IDLE) {
			if (atomic_read(&smi_info->need_watch)) {
				schedule_timeout_interruptible(100);
			} else {
				/* Wait to be woken up when we are needed. */
				__set_current_state(TASK_INTERRUPTIBLE);
				schedule();
			}
		} else
1035
			schedule_timeout_interruptible(1);
C
Corey Minyard 已提交
1036 1037 1038 1039 1040
	}
	return 0;
}


L
Linus Torvalds 已提交
1041 1042 1043
static void poll(void *send_info)
{
	struct smi_info *smi_info = send_info;
C
Corey Minyard 已提交
1044
	unsigned long flags = 0;
C
Corey Minyard 已提交
1045
	bool run_to_completion = smi_info->run_to_completion;
L
Linus Torvalds 已提交
1046

C
Corey Minyard 已提交
1047 1048 1049 1050 1051
	/*
	 * Make sure there is some delay in the poll loop so we can
	 * drive time forward and timeout things.
	 */
	udelay(10);
C
Corey Minyard 已提交
1052 1053
	if (!run_to_completion)
		spin_lock_irqsave(&smi_info->si_lock, flags);
C
Corey Minyard 已提交
1054
	smi_event_handler(smi_info, 10);
C
Corey Minyard 已提交
1055 1056
	if (!run_to_completion)
		spin_unlock_irqrestore(&smi_info->si_lock, flags);
L
Linus Torvalds 已提交
1057 1058 1059 1060 1061 1062
}

static void request_events(void *send_info)
{
	struct smi_info *smi_info = send_info;

1063
	if (!smi_info->has_event_buffer)
1064 1065
		return;

L
Linus Torvalds 已提交
1066 1067 1068
	atomic_set(&smi_info->req_events, 1);
}

C
Corey Minyard 已提交
1069
static void set_need_watch(void *send_info, bool enable)
1070 1071 1072 1073 1074 1075 1076 1077 1078 1079
{
	struct smi_info *smi_info = send_info;
	unsigned long flags;

	atomic_set(&smi_info->need_watch, enable);
	spin_lock_irqsave(&smi_info->si_lock, flags);
	check_start_timer_thread(smi_info);
	spin_unlock_irqrestore(&smi_info->si_lock, flags);
}

1080
static void smi_timeout(struct timer_list *t)
L
Linus Torvalds 已提交
1081
{
1082
	struct smi_info   *smi_info = from_timer(smi_info, t, si_timer);
L
Linus Torvalds 已提交
1083 1084 1085
	enum si_sm_result smi_result;
	unsigned long     flags;
	unsigned long     jiffies_now;
C
Corey Minyard 已提交
1086
	long              time_diff;
M
Matthew Garrett 已提交
1087
	long		  timeout;
L
Linus Torvalds 已提交
1088 1089

	spin_lock_irqsave(&(smi_info->si_lock), flags);
1090 1091
	debug_timestamp("Timer");

L
Linus Torvalds 已提交
1092
	jiffies_now = jiffies;
C
Corey Minyard 已提交
1093
	time_diff = (((long)jiffies_now - (long)smi_info->last_timeout_jiffies)
L
Linus Torvalds 已提交
1094 1095 1096
		     * SI_USEC_PER_JIFFY);
	smi_result = smi_event_handler(smi_info, time_diff);

1097
	if ((smi_info->io.irq) && (!smi_info->interrupt_disabled)) {
L
Linus Torvalds 已提交
1098
		/* Running with interrupts, only do long timeouts. */
M
Matthew Garrett 已提交
1099
		timeout = jiffies + SI_TIMEOUT_JIFFIES;
1100
		smi_inc_stat(smi_info, long_timeouts);
M
Matthew Garrett 已提交
1101
		goto do_mod_timer;
L
Linus Torvalds 已提交
1102 1103
	}

1104 1105 1106 1107
	/*
	 * If the state machine asks for a short delay, then shorten
	 * the timer timeout.
	 */
L
Linus Torvalds 已提交
1108
	if (smi_result == SI_SM_CALL_WITH_DELAY) {
1109
		smi_inc_stat(smi_info, short_timeouts);
M
Matthew Garrett 已提交
1110
		timeout = jiffies + 1;
L
Linus Torvalds 已提交
1111
	} else {
1112
		smi_inc_stat(smi_info, long_timeouts);
M
Matthew Garrett 已提交
1113
		timeout = jiffies + SI_TIMEOUT_JIFFIES;
L
Linus Torvalds 已提交
1114 1115
	}

1116
do_mod_timer:
M
Matthew Garrett 已提交
1117
	if (smi_result != SI_SM_IDLE)
1118 1119 1120 1121
		smi_mod_timer(smi_info, timeout);
	else
		smi_info->timer_running = false;
	spin_unlock_irqrestore(&(smi_info->si_lock), flags);
L
Linus Torvalds 已提交
1122 1123
}

1124
irqreturn_t ipmi_si_irq_handler(int irq, void *data)
L
Linus Torvalds 已提交
1125 1126 1127 1128
{
	struct smi_info *smi_info = data;
	unsigned long   flags;

1129 1130 1131 1132 1133 1134
	if (smi_info->io.si_type == SI_BT)
		/* We need to clear the IRQ flag for the BT interface. */
		smi_info->io.outputb(&smi_info->io, IPMI_BT_INTMASK_REG,
				     IPMI_BT_INTMASK_CLEAR_IRQ_BIT
				     | IPMI_BT_INTMASK_ENABLE_IRQ_BIT);

L
Linus Torvalds 已提交
1135 1136
	spin_lock_irqsave(&(smi_info->si_lock), flags);

1137
	smi_inc_stat(smi_info, interrupts);
L
Linus Torvalds 已提交
1138

1139 1140
	debug_timestamp("Interrupt");

L
Linus Torvalds 已提交
1141 1142 1143 1144 1145
	smi_event_handler(smi_info, 0);
	spin_unlock_irqrestore(&(smi_info->si_lock), flags);
	return IRQ_HANDLED;
}

1146 1147
static int smi_start_processing(void            *send_info,
				struct ipmi_smi *intf)
1148 1149
{
	struct smi_info *new_smi = send_info;
1150
	int             enable = 0;
1151 1152 1153 1154

	new_smi->intf = intf;

	/* Set up the timer that drives the interface. */
1155
	timer_setup(&new_smi->si_timer, smi_timeout, 0);
1156
	new_smi->timer_can_start = true;
1157
	smi_mod_timer(new_smi, jiffies + SI_TIMEOUT_JIFFIES);
1158

1159
	/* Try to claim any interrupts. */
1160 1161 1162 1163
	if (new_smi->io.irq_setup) {
		new_smi->io.irq_handler_data = new_smi;
		new_smi->io.irq_setup(&new_smi->io);
	}
1164

1165 1166 1167 1168 1169
	/*
	 * Check if the user forcefully enabled the daemon.
	 */
	if (new_smi->intf_num < num_force_kipmid)
		enable = force_kipmid[new_smi->intf_num];
1170 1171 1172 1173
	/*
	 * The BT interface is efficient enough to not need a thread,
	 * and there is no need for a thread if we have interrupts.
	 */
1174
	else if ((new_smi->io.si_type != SI_BT) && (!new_smi->io.irq))
1175 1176 1177
		enable = 1;

	if (enable) {
1178 1179 1180
		new_smi->thread = kthread_run(ipmi_thread, new_smi,
					      "kipmi%d", new_smi->intf_num);
		if (IS_ERR(new_smi->thread)) {
1181
			dev_notice(new_smi->io.dev, "Could not start"
1182 1183 1184
				   " kernel thread due to error %ld, only using"
				   " timers to drive the interface\n",
				   PTR_ERR(new_smi->thread));
1185 1186 1187 1188 1189 1190
			new_smi->thread = NULL;
		}
	}

	return 0;
}
1191

1192 1193 1194 1195
static int get_smi_info(void *send_info, struct ipmi_smi_info *data)
{
	struct smi_info *smi = send_info;

1196 1197
	data->addr_src = smi->io.addr_source;
	data->dev = smi->io.dev;
1198
	data->addr_info = smi->io.addr_info;
1199
	get_device(smi->io.dev);
1200 1201 1202 1203

	return 0;
}

C
Corey Minyard 已提交
1204
static void set_maintenance_mode(void *send_info, bool enable)
C
Corey Minyard 已提交
1205 1206 1207 1208 1209 1210 1211
{
	struct smi_info   *smi_info = send_info;

	if (!enable)
		atomic_set(&smi_info->req_events, 0);
}

1212
static void shutdown_smi(void *send_info);
1213
static const struct ipmi_smi_handlers handlers = {
L
Linus Torvalds 已提交
1214
	.owner                  = THIS_MODULE,
1215
	.start_processing       = smi_start_processing,
1216
	.shutdown               = shutdown_smi,
1217
	.get_smi_info		= get_smi_info,
L
Linus Torvalds 已提交
1218 1219
	.sender			= sender,
	.request_events		= request_events,
1220
	.set_need_watch		= set_need_watch,
C
Corey Minyard 已提交
1221
	.set_maintenance_mode   = set_maintenance_mode,
L
Linus Torvalds 已提交
1222
	.set_run_to_completion  = set_run_to_completion,
1223
	.flush_messages		= flush_messages,
L
Linus Torvalds 已提交
1224 1225 1226
	.poll			= poll,
};

1227
static LIST_HEAD(smi_infos);
1228
static DEFINE_MUTEX(smi_infos_lock);
1229
static int smi_num; /* Used to sequence the SMIs */
L
Linus Torvalds 已提交
1230

1231
static const char * const addr_space_to_str[] = { "i/o", "mem" };
1232

1233 1234 1235 1236
module_param_array(force_kipmid, int, &num_force_kipmid, 0);
MODULE_PARM_DESC(force_kipmid, "Force the kipmi daemon to be enabled (1) or"
		 " disabled(0).  Normally the IPMI driver auto-detects"
		 " this, but the value may be overridden by this parm.");
C
Corey Minyard 已提交
1237
module_param(unload_when_empty, bool, 0);
1238 1239 1240
MODULE_PARM_DESC(unload_when_empty, "Unload the module if no interfaces are"
		 " specified or found, default is 1.  Setting to 0"
		 " is useful for hot add of devices using hotmod.");
1241 1242 1243 1244 1245
module_param_array(kipmid_max_busy_us, uint, &num_max_busy_us, 0644);
MODULE_PARM_DESC(kipmid_max_busy_us,
		 "Max time (in microseconds) to busy-wait for IPMI data before"
		 " sleeping. 0 (default) means to wait forever. Set to 100-500"
		 " if kipmid is using up a lot of CPU time.");
L
Linus Torvalds 已提交
1246

1247 1248 1249 1250 1251 1252 1253
void ipmi_irq_finish_setup(struct si_sm_io *io)
{
	if (io->si_type == SI_BT)
		/* Enable the interrupt in the BT interface. */
		io->outputb(io, IPMI_BT_INTMASK_REG,
			    IPMI_BT_INTMASK_ENABLE_IRQ_BIT);
}
L
Linus Torvalds 已提交
1254

1255
void ipmi_irq_start_cleanup(struct si_sm_io *io)
L
Linus Torvalds 已提交
1256
{
1257
	if (io->si_type == SI_BT)
1258
		/* Disable the interrupt in the BT interface. */
1259 1260 1261 1262 1263 1264 1265
		io->outputb(io, IPMI_BT_INTMASK_REG, 0);
}

static void std_irq_cleanup(struct si_sm_io *io)
{
	ipmi_irq_start_cleanup(io);
	free_irq(io->irq, io->irq_handler_data);
L
Linus Torvalds 已提交
1266 1267
}

1268
int ipmi_std_irq_setup(struct si_sm_io *io)
L
Linus Torvalds 已提交
1269 1270 1271
{
	int rv;

1272
	if (!io->irq)
L
Linus Torvalds 已提交
1273 1274
		return 0;

1275 1276 1277 1278 1279
	rv = request_irq(io->irq,
			 ipmi_si_irq_handler,
			 IRQF_SHARED,
			 DEVICE_NAME,
			 io->irq_handler_data);
L
Linus Torvalds 已提交
1280
	if (rv) {
1281
		dev_warn(io->dev, "%s unable to claim interrupt %d,"
1282
			 " running polled\n",
1283 1284
			 DEVICE_NAME, io->irq);
		io->irq = 0;
L
Linus Torvalds 已提交
1285
	} else {
1286 1287 1288
		io->irq_cleanup = std_irq_cleanup;
		ipmi_irq_finish_setup(io);
		dev_info(io->dev, "Using irq %d\n", io->irq);
L
Linus Torvalds 已提交
1289 1290 1291 1292 1293
	}

	return rv;
}

1294
static int wait_for_msg_done(struct smi_info *smi_info)
L
Linus Torvalds 已提交
1295
{
1296
	enum si_sm_result     smi_result;
L
Linus Torvalds 已提交
1297 1298

	smi_result = smi_info->handlers->event(smi_info->si_sm, 0);
1299
	for (;;) {
C
Corey Minyard 已提交
1300 1301
		if (smi_result == SI_SM_CALL_WITH_DELAY ||
		    smi_result == SI_SM_CALL_WITH_TICK_DELAY) {
1302
			schedule_timeout_uninterruptible(1);
L
Linus Torvalds 已提交
1303
			smi_result = smi_info->handlers->event(
1304
				smi_info->si_sm, jiffies_to_usecs(1));
1305
		} else if (smi_result == SI_SM_CALL_WITHOUT_DELAY) {
L
Linus Torvalds 已提交
1306 1307
			smi_result = smi_info->handlers->event(
				smi_info->si_sm, 0);
1308
		} else
L
Linus Torvalds 已提交
1309 1310
			break;
	}
1311
	if (smi_result == SI_SM_HOSED)
1312 1313 1314 1315
		/*
		 * We couldn't get the state machine to run, so whatever's at
		 * the port is probably not an IPMI SMI interface.
		 */
1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341
		return -ENODEV;

	return 0;
}

static int try_get_dev_id(struct smi_info *smi_info)
{
	unsigned char         msg[2];
	unsigned char         *resp;
	unsigned long         resp_len;
	int                   rv = 0;

	resp = kmalloc(IPMI_MAX_MSG_LENGTH, GFP_KERNEL);
	if (!resp)
		return -ENOMEM;

	/*
	 * Do a Get Device ID command, since it comes back with some
	 * useful info.
	 */
	msg[0] = IPMI_NETFN_APP_REQUEST << 2;
	msg[1] = IPMI_GET_DEVICE_ID_CMD;
	smi_info->handlers->start_transaction(smi_info->si_sm, msg, 2);

	rv = wait_for_msg_done(smi_info);
	if (rv)
L
Linus Torvalds 已提交
1342 1343 1344 1345 1346
		goto out;

	resp_len = smi_info->handlers->get_result(smi_info->si_sm,
						  resp, IPMI_MAX_MSG_LENGTH);

C
Corey Minyard 已提交
1347
	/* Check and record info from the get device id, in case we need it. */
1348 1349
	rv = ipmi_demangle_device_id(resp[0] >> 2, resp[1],
			resp + 2, resp_len - 2, &smi_info->device_id);
L
Linus Torvalds 已提交
1350

1351
out:
L
Linus Torvalds 已提交
1352 1353 1354 1355
	kfree(resp);
	return rv;
}

1356
static int get_global_enables(struct smi_info *smi_info, u8 *enables)
1357 1358 1359 1360 1361 1362 1363
{
	unsigned char         msg[3];
	unsigned char         *resp;
	unsigned long         resp_len;
	int                   rv;

	resp = kmalloc(IPMI_MAX_MSG_LENGTH, GFP_KERNEL);
1364 1365
	if (!resp)
		return -ENOMEM;
1366 1367 1368 1369 1370 1371 1372

	msg[0] = IPMI_NETFN_APP_REQUEST << 2;
	msg[1] = IPMI_GET_BMC_GLOBAL_ENABLES_CMD;
	smi_info->handlers->start_transaction(smi_info->si_sm, msg, 2);

	rv = wait_for_msg_done(smi_info);
	if (rv) {
1373
		dev_warn(smi_info->io.dev,
1374 1375
			 "Error getting response from get global enables command: %d\n",
			 rv);
1376 1377 1378 1379 1380 1381 1382 1383 1384 1385
		goto out;
	}

	resp_len = smi_info->handlers->get_result(smi_info->si_sm,
						  resp, IPMI_MAX_MSG_LENGTH);

	if (resp_len < 4 ||
			resp[0] != (IPMI_NETFN_APP_REQUEST | 1) << 2 ||
			resp[1] != IPMI_GET_BMC_GLOBAL_ENABLES_CMD   ||
			resp[2] != 0) {
1386
		dev_warn(smi_info->io.dev,
1387 1388
			 "Invalid return from get global enables command: %ld %x %x %x\n",
			 resp_len, resp[0], resp[1], resp[2]);
1389 1390
		rv = -EINVAL;
		goto out;
1391 1392
	} else {
		*enables = resp[3];
1393 1394
	}

1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412
out:
	kfree(resp);
	return rv;
}

/*
 * Returns 1 if it gets an error from the command.
 */
static int set_global_enables(struct smi_info *smi_info, u8 enables)
{
	unsigned char         msg[3];
	unsigned char         *resp;
	unsigned long         resp_len;
	int                   rv;

	resp = kmalloc(IPMI_MAX_MSG_LENGTH, GFP_KERNEL);
	if (!resp)
		return -ENOMEM;
1413 1414 1415

	msg[0] = IPMI_NETFN_APP_REQUEST << 2;
	msg[1] = IPMI_SET_BMC_GLOBAL_ENABLES_CMD;
1416
	msg[2] = enables;
1417 1418 1419 1420
	smi_info->handlers->start_transaction(smi_info->si_sm, msg, 3);

	rv = wait_for_msg_done(smi_info);
	if (rv) {
1421
		dev_warn(smi_info->io.dev,
1422 1423
			 "Error getting response from set global enables command: %d\n",
			 rv);
1424 1425 1426 1427 1428 1429 1430 1431 1432
		goto out;
	}

	resp_len = smi_info->handlers->get_result(smi_info->si_sm,
						  resp, IPMI_MAX_MSG_LENGTH);

	if (resp_len < 3 ||
			resp[0] != (IPMI_NETFN_APP_REQUEST | 1) << 2 ||
			resp[1] != IPMI_SET_BMC_GLOBAL_ENABLES_CMD) {
1433
		dev_warn(smi_info->io.dev,
1434 1435
			 "Invalid return from set global enables command: %ld %x %x\n",
			 resp_len, resp[0], resp[1]);
1436 1437 1438 1439
		rv = -EINVAL;
		goto out;
	}

1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468
	if (resp[2] != 0)
		rv = 1;

out:
	kfree(resp);
	return rv;
}

/*
 * Some BMCs do not support clearing the receive irq bit in the global
 * enables (even if they don't support interrupts on the BMC).  Check
 * for this and handle it properly.
 */
static void check_clr_rcv_irq(struct smi_info *smi_info)
{
	u8 enables = 0;
	int rv;

	rv = get_global_enables(smi_info, &enables);
	if (!rv) {
		if ((enables & IPMI_BMC_RCV_MSG_INTR) == 0)
			/* Already clear, should work ok. */
			return;

		enables &= ~IPMI_BMC_RCV_MSG_INTR;
		rv = set_global_enables(smi_info, enables);
	}

	if (rv < 0) {
1469
		dev_err(smi_info->io.dev,
1470 1471 1472 1473 1474
			"Cannot check clearing the rcv irq: %d\n", rv);
		return;
	}

	if (rv) {
1475 1476 1477 1478
		/*
		 * An error when setting the event buffer bit means
		 * clearing the bit is not supported.
		 */
1479
		dev_warn(smi_info->io.dev,
1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494
			 "The BMC does not support clearing the recv irq bit, compensating, but the BMC needs to be fixed.\n");
		smi_info->cannot_disable_irq = true;
	}
}

/*
 * Some BMCs do not support setting the interrupt bits in the global
 * enables even if they support interrupts.  Clearly bad, but we can
 * compensate.
 */
static void check_set_rcv_irq(struct smi_info *smi_info)
{
	u8 enables = 0;
	int rv;

1495
	if (!smi_info->io.irq)
1496 1497 1498 1499 1500 1501 1502 1503 1504
		return;

	rv = get_global_enables(smi_info, &enables);
	if (!rv) {
		enables |= IPMI_BMC_RCV_MSG_INTR;
		rv = set_global_enables(smi_info, enables);
	}

	if (rv < 0) {
1505
		dev_err(smi_info->io.dev,
1506 1507 1508 1509 1510 1511 1512 1513 1514
			"Cannot check setting the rcv irq: %d\n", rv);
		return;
	}

	if (rv) {
		/*
		 * An error when setting the event buffer bit means
		 * setting the bit is not supported.
		 */
1515
		dev_warn(smi_info->io.dev,
1516 1517 1518
			 "The BMC does not support setting the recv irq bit, compensating, but the BMC needs to be fixed.\n");
		smi_info->cannot_disable_irq = true;
		smi_info->irq_enable_broken = true;
1519 1520 1521
	}
}

1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538
static int try_enable_event_buffer(struct smi_info *smi_info)
{
	unsigned char         msg[3];
	unsigned char         *resp;
	unsigned long         resp_len;
	int                   rv = 0;

	resp = kmalloc(IPMI_MAX_MSG_LENGTH, GFP_KERNEL);
	if (!resp)
		return -ENOMEM;

	msg[0] = IPMI_NETFN_APP_REQUEST << 2;
	msg[1] = IPMI_GET_BMC_GLOBAL_ENABLES_CMD;
	smi_info->handlers->start_transaction(smi_info->si_sm, msg, 2);

	rv = wait_for_msg_done(smi_info);
	if (rv) {
C
Corey Minyard 已提交
1539
		pr_warn(PFX "Error getting response from get global enables command, the event buffer is not enabled.\n");
1540 1541 1542 1543 1544 1545 1546 1547 1548 1549
		goto out;
	}

	resp_len = smi_info->handlers->get_result(smi_info->si_sm,
						  resp, IPMI_MAX_MSG_LENGTH);

	if (resp_len < 4 ||
			resp[0] != (IPMI_NETFN_APP_REQUEST | 1) << 2 ||
			resp[1] != IPMI_GET_BMC_GLOBAL_ENABLES_CMD   ||
			resp[2] != 0) {
C
Corey Minyard 已提交
1550
		pr_warn(PFX "Invalid return from get global enables command, cannot enable the event buffer.\n");
1551 1552 1553 1554
		rv = -EINVAL;
		goto out;
	}

1555
	if (resp[3] & IPMI_BMC_EVT_MSG_BUFF) {
1556
		/* buffer is already enabled, nothing to do. */
1557
		smi_info->supports_event_msg_buff = true;
1558
		goto out;
1559
	}
1560 1561 1562 1563 1564 1565 1566 1567

	msg[0] = IPMI_NETFN_APP_REQUEST << 2;
	msg[1] = IPMI_SET_BMC_GLOBAL_ENABLES_CMD;
	msg[2] = resp[3] | IPMI_BMC_EVT_MSG_BUFF;
	smi_info->handlers->start_transaction(smi_info->si_sm, msg, 3);

	rv = wait_for_msg_done(smi_info);
	if (rv) {
C
Corey Minyard 已提交
1568
		pr_warn(PFX "Error getting response from set global, enables command, the event buffer is not enabled.\n");
1569 1570 1571 1572 1573 1574 1575 1576 1577
		goto out;
	}

	resp_len = smi_info->handlers->get_result(smi_info->si_sm,
						  resp, IPMI_MAX_MSG_LENGTH);

	if (resp_len < 3 ||
			resp[0] != (IPMI_NETFN_APP_REQUEST | 1) << 2 ||
			resp[1] != IPMI_SET_BMC_GLOBAL_ENABLES_CMD) {
C
Corey Minyard 已提交
1578
		pr_warn(PFX "Invalid return from get global, enables command, not enable the event buffer.\n");
1579 1580 1581 1582 1583 1584 1585 1586 1587 1588
		rv = -EINVAL;
		goto out;
	}

	if (resp[2] != 0)
		/*
		 * An error when setting the event buffer bit means
		 * that the event buffer is not supported.
		 */
		rv = -ENOENT;
1589 1590 1591
	else
		smi_info->supports_event_msg_buff = true;

1592
out:
1593 1594 1595 1596
	kfree(resp);
	return rv;
}

1597
#ifdef CONFIG_IPMI_PROC_INTERFACE
1598
static int smi_type_proc_show(struct seq_file *m, void *v)
L
Linus Torvalds 已提交
1599
{
1600
	struct smi_info *smi = m->private;
L
Linus Torvalds 已提交
1601

1602
	seq_printf(m, "%s\n", si_to_str[smi->io.si_type]);
1603

1604
	return 0;
L
Linus Torvalds 已提交
1605 1606
}

1607
static int smi_type_proc_open(struct inode *inode, struct file *file)
L
Linus Torvalds 已提交
1608
{
A
Al Viro 已提交
1609
	return single_open(file, smi_type_proc_show, PDE_DATA(inode));
1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621
}

static const struct file_operations smi_type_proc_ops = {
	.open		= smi_type_proc_open,
	.read		= seq_read,
	.llseek		= seq_lseek,
	.release	= single_release,
};

static int smi_si_stats_proc_show(struct seq_file *m, void *v)
{
	struct smi_info *smi = m->private;
L
Linus Torvalds 已提交
1622

1623
	seq_printf(m, "interrupts_enabled:    %d\n",
1624
		       smi->io.irq && !smi->interrupt_disabled);
1625
	seq_printf(m, "short_timeouts:        %u\n",
1626
		       smi_get_stat(smi, short_timeouts));
1627
	seq_printf(m, "long_timeouts:         %u\n",
1628
		       smi_get_stat(smi, long_timeouts));
1629
	seq_printf(m, "idles:                 %u\n",
1630
		       smi_get_stat(smi, idles));
1631
	seq_printf(m, "interrupts:            %u\n",
1632
		       smi_get_stat(smi, interrupts));
1633
	seq_printf(m, "attentions:            %u\n",
1634
		       smi_get_stat(smi, attentions));
1635
	seq_printf(m, "flag_fetches:          %u\n",
1636
		       smi_get_stat(smi, flag_fetches));
1637
	seq_printf(m, "hosed_count:           %u\n",
1638
		       smi_get_stat(smi, hosed_count));
1639
	seq_printf(m, "complete_transactions: %u\n",
1640
		       smi_get_stat(smi, complete_transactions));
1641
	seq_printf(m, "events:                %u\n",
1642
		       smi_get_stat(smi, events));
1643
	seq_printf(m, "watchdog_pretimeouts:  %u\n",
1644
		       smi_get_stat(smi, watchdog_pretimeouts));
1645
	seq_printf(m, "incoming_messages:     %u\n",
1646
		       smi_get_stat(smi, incoming_messages));
1647 1648
	return 0;
}
L
Linus Torvalds 已提交
1649

1650 1651
static int smi_si_stats_proc_open(struct inode *inode, struct file *file)
{
A
Al Viro 已提交
1652
	return single_open(file, smi_si_stats_proc_show, PDE_DATA(inode));
1653 1654
}

1655 1656 1657 1658 1659 1660 1661 1662
static const struct file_operations smi_si_stats_proc_ops = {
	.open		= smi_si_stats_proc_open,
	.read		= seq_read,
	.llseek		= seq_lseek,
	.release	= single_release,
};

static int smi_params_proc_show(struct seq_file *m, void *v)
1663
{
1664
	struct smi_info *smi = m->private;
1665

1666 1667
	seq_printf(m,
		   "%s,%s,0x%lx,rsp=%d,rsi=%d,rsh=%d,irq=%d,ipmb=%d\n",
1668
		   si_to_str[smi->io.si_type],
1669 1670 1671 1672 1673
		   addr_space_to_str[smi->io.addr_type],
		   smi->io.addr_data,
		   smi->io.regspacing,
		   smi->io.regsize,
		   smi->io.regshift,
1674 1675
		   smi->io.irq,
		   smi->io.slave_addr);
1676

1677
	return 0;
L
Linus Torvalds 已提交
1678 1679
}

1680 1681
static int smi_params_proc_open(struct inode *inode, struct file *file)
{
A
Al Viro 已提交
1682
	return single_open(file, smi_params_proc_show, PDE_DATA(inode));
1683 1684 1685 1686 1687 1688 1689 1690
}

static const struct file_operations smi_params_proc_ops = {
	.open		= smi_params_proc_open,
	.read		= seq_read,
	.llseek		= seq_lseek,
	.release	= single_release,
};
1691
#endif
1692

1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778
#define IPMI_SI_ATTR(name) \
static ssize_t ipmi_##name##_show(struct device *dev,			\
				  struct device_attribute *attr,	\
				  char *buf)				\
{									\
	struct smi_info *smi_info = dev_get_drvdata(dev);		\
									\
	return snprintf(buf, 10, "%u\n", smi_get_stat(smi_info, name));	\
}									\
static DEVICE_ATTR(name, S_IRUGO, ipmi_##name##_show, NULL)

static ssize_t ipmi_type_show(struct device *dev,
			      struct device_attribute *attr,
			      char *buf)
{
	struct smi_info *smi_info = dev_get_drvdata(dev);

	return snprintf(buf, 10, "%s\n", si_to_str[smi_info->io.si_type]);
}
static DEVICE_ATTR(type, S_IRUGO, ipmi_type_show, NULL);

static ssize_t ipmi_interrupts_enabled_show(struct device *dev,
					    struct device_attribute *attr,
					    char *buf)
{
	struct smi_info *smi_info = dev_get_drvdata(dev);
	int enabled = smi_info->io.irq && !smi_info->interrupt_disabled;

	return snprintf(buf, 10, "%d\n", enabled);
}
static DEVICE_ATTR(interrupts_enabled, S_IRUGO,
		   ipmi_interrupts_enabled_show, NULL);

IPMI_SI_ATTR(short_timeouts);
IPMI_SI_ATTR(long_timeouts);
IPMI_SI_ATTR(idles);
IPMI_SI_ATTR(interrupts);
IPMI_SI_ATTR(attentions);
IPMI_SI_ATTR(flag_fetches);
IPMI_SI_ATTR(hosed_count);
IPMI_SI_ATTR(complete_transactions);
IPMI_SI_ATTR(events);
IPMI_SI_ATTR(watchdog_pretimeouts);
IPMI_SI_ATTR(incoming_messages);

static ssize_t ipmi_params_show(struct device *dev,
				struct device_attribute *attr,
				char *buf)
{
	struct smi_info *smi_info = dev_get_drvdata(dev);

	return snprintf(buf, 200,
			"%s,%s,0x%lx,rsp=%d,rsi=%d,rsh=%d,irq=%d,ipmb=%d\n",
			si_to_str[smi_info->io.si_type],
			addr_space_to_str[smi_info->io.addr_type],
			smi_info->io.addr_data,
			smi_info->io.regspacing,
			smi_info->io.regsize,
			smi_info->io.regshift,
			smi_info->io.irq,
			smi_info->io.slave_addr);
}
static DEVICE_ATTR(params, S_IRUGO, ipmi_params_show, NULL);

static struct attribute *ipmi_si_dev_attrs[] = {
	&dev_attr_type.attr,
	&dev_attr_interrupts_enabled.attr,
	&dev_attr_short_timeouts.attr,
	&dev_attr_long_timeouts.attr,
	&dev_attr_idles.attr,
	&dev_attr_interrupts.attr,
	&dev_attr_attentions.attr,
	&dev_attr_flag_fetches.attr,
	&dev_attr_hosed_count.attr,
	&dev_attr_complete_transactions.attr,
	&dev_attr_events.attr,
	&dev_attr_watchdog_pretimeouts.attr,
	&dev_attr_incoming_messages.attr,
	&dev_attr_params.attr,
	NULL
};

static const struct attribute_group ipmi_si_dev_attr_group = {
	.attrs		= ipmi_si_dev_attrs,
};

1779 1780 1781 1782 1783 1784 1785 1786 1787
/*
 * oem_data_avail_to_receive_msg_avail
 * @info - smi_info structure with msg_flags set
 *
 * Converts flags from OEM_DATA_AVAIL to RECEIVE_MSG_AVAIL
 * Returns 1 indicating need to re-run handle_flags().
 */
static int oem_data_avail_to_receive_msg_avail(struct smi_info *smi_info)
{
C
Corey Minyard 已提交
1788
	smi_info->msg_flags = ((smi_info->msg_flags & ~OEM_DATA_AVAIL) |
1789
			       RECEIVE_MSG_AVAIL);
1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813
	return 1;
}

/*
 * setup_dell_poweredge_oem_data_handler
 * @info - smi_info.device_id must be populated
 *
 * Systems that match, but have firmware version < 1.40 may assert
 * OEM0_DATA_AVAIL on their own, without being told via Set Flags that
 * it's safe to do so.  Such systems will de-assert OEM1_DATA_AVAIL
 * upon receipt of IPMI_GET_MSG_CMD, so we should treat these flags
 * as RECEIVE_MSG_AVAIL instead.
 *
 * As Dell has no plans to release IPMI 1.5 firmware that *ever*
 * assert the OEM[012] bits, and if it did, the driver would have to
 * change to handle that properly, we don't actually check for the
 * firmware version.
 * Device ID = 0x20                BMC on PowerEdge 8G servers
 * Device Revision = 0x80
 * Firmware Revision1 = 0x01       BMC version 1.40
 * Firmware Revision2 = 0x40       BCD encoded
 * IPMI Version = 0x51             IPMI 1.5
 * Manufacturer ID = A2 02 00      Dell IANA
 *
C
Corey Minyard 已提交
1814 1815 1816
 * Additionally, PowerEdge systems with IPMI < 1.5 may also assert
 * OEM0_DATA_AVAIL and needs to be treated as RECEIVE_MSG_AVAIL.
 *
1817 1818 1819 1820
 */
#define DELL_POWEREDGE_8G_BMC_DEVICE_ID  0x20
#define DELL_POWEREDGE_8G_BMC_DEVICE_REV 0x80
#define DELL_POWEREDGE_8G_BMC_IPMI_VERSION 0x51
1821
#define DELL_IANA_MFR_ID 0x0002a2
1822 1823 1824
static void setup_dell_poweredge_oem_data_handler(struct smi_info *smi_info)
{
	struct ipmi_device_id *id = &smi_info->device_id;
1825
	if (id->manufacturer_id == DELL_IANA_MFR_ID) {
C
Corey Minyard 已提交
1826 1827
		if (id->device_id       == DELL_POWEREDGE_8G_BMC_DEVICE_ID  &&
		    id->device_revision == DELL_POWEREDGE_8G_BMC_DEVICE_REV &&
1828
		    id->ipmi_version   == DELL_POWEREDGE_8G_BMC_IPMI_VERSION) {
C
Corey Minyard 已提交
1829 1830
			smi_info->oem_data_avail_handler =
				oem_data_avail_to_receive_msg_avail;
1831 1832 1833
		} else if (ipmi_version_major(id) < 1 ||
			   (ipmi_version_major(id) == 1 &&
			    ipmi_version_minor(id) < 5)) {
C
Corey Minyard 已提交
1834 1835 1836
			smi_info->oem_data_avail_handler =
				oem_data_avail_to_receive_msg_avail;
		}
1837 1838 1839
	}
}

1840 1841 1842 1843 1844
#define CANNOT_RETURN_REQUESTED_LENGTH 0xCA
static void return_hosed_msg_badsize(struct smi_info *smi_info)
{
	struct ipmi_smi_msg *msg = smi_info->curr_msg;

L
Lucas De Marchi 已提交
1845
	/* Make it a response */
1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898
	msg->rsp[0] = msg->data[0] | 4;
	msg->rsp[1] = msg->data[1];
	msg->rsp[2] = CANNOT_RETURN_REQUESTED_LENGTH;
	msg->rsp_size = 3;
	smi_info->curr_msg = NULL;
	deliver_recv_msg(smi_info, msg);
}

/*
 * dell_poweredge_bt_xaction_handler
 * @info - smi_info.device_id must be populated
 *
 * Dell PowerEdge servers with the BT interface (x6xx and 1750) will
 * not respond to a Get SDR command if the length of the data
 * requested is exactly 0x3A, which leads to command timeouts and no
 * data returned.  This intercepts such commands, and causes userspace
 * callers to try again with a different-sized buffer, which succeeds.
 */

#define STORAGE_NETFN 0x0A
#define STORAGE_CMD_GET_SDR 0x23
static int dell_poweredge_bt_xaction_handler(struct notifier_block *self,
					     unsigned long unused,
					     void *in)
{
	struct smi_info *smi_info = in;
	unsigned char *data = smi_info->curr_msg->data;
	unsigned int size   = smi_info->curr_msg->data_size;
	if (size >= 8 &&
	    (data[0]>>2) == STORAGE_NETFN &&
	    data[1] == STORAGE_CMD_GET_SDR &&
	    data[7] == 0x3A) {
		return_hosed_msg_badsize(smi_info);
		return NOTIFY_STOP;
	}
	return NOTIFY_DONE;
}

static struct notifier_block dell_poweredge_bt_xaction_notifier = {
	.notifier_call	= dell_poweredge_bt_xaction_handler,
};

/*
 * setup_dell_poweredge_bt_xaction_handler
 * @info - smi_info.device_id must be filled in already
 *
 * Fills in smi_info.device_id.start_transaction_pre_hook
 * when we know what function to use there.
 */
static void
setup_dell_poweredge_bt_xaction_handler(struct smi_info *smi_info)
{
	struct ipmi_device_id *id = &smi_info->device_id;
1899
	if (id->manufacturer_id == DELL_IANA_MFR_ID &&
1900
	    smi_info->io.si_type == SI_BT)
1901 1902 1903
		register_xaction_notifier(&dell_poweredge_bt_xaction_notifier);
}

1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916
/*
 * setup_oem_data_handler
 * @info - smi_info.device_id must be filled in already
 *
 * Fills in smi_info.device_id.oem_data_available_handler
 * when we know what function to use there.
 */

static void setup_oem_data_handler(struct smi_info *smi_info)
{
	setup_dell_poweredge_oem_data_handler(smi_info);
}

1917 1918 1919 1920 1921
static void setup_xaction_handlers(struct smi_info *smi_info)
{
	setup_dell_poweredge_bt_xaction_handler(smi_info);
}

1922 1923 1924 1925 1926 1927
static void check_for_broken_irqs(struct smi_info *smi_info)
{
	check_clr_rcv_irq(smi_info);
	check_set_rcv_irq(smi_info);
}

1928
static inline void stop_timer_and_thread(struct smi_info *smi_info)
C
Corey Minyard 已提交
1929
{
1930
	if (smi_info->thread != NULL) {
1931
		kthread_stop(smi_info->thread);
1932 1933
		smi_info->thread = NULL;
	}
1934 1935

	smi_info->timer_can_start = false;
1936
	if (smi_info->timer_running)
1937
		del_timer_sync(&smi_info->si_timer);
C
Corey Minyard 已提交
1938 1939
}

1940
static struct smi_info *find_dup_si(struct smi_info *info)
L
Linus Torvalds 已提交
1941
{
1942
	struct smi_info *e;
L
Linus Torvalds 已提交
1943

1944 1945 1946
	list_for_each_entry(e, &smi_infos, link) {
		if (e->io.addr_type != info->io.addr_type)
			continue;
1947 1948 1949 1950 1951 1952
		if (e->io.addr_data == info->io.addr_data) {
			/*
			 * This is a cheap hack, ACPI doesn't have a defined
			 * slave address but SMBIOS does.  Pick it up from
			 * any source that has it available.
			 */
1953 1954
			if (info->io.slave_addr && !e->io.slave_addr)
				e->io.slave_addr = info->io.slave_addr;
1955
			return e;
1956
		}
1957
	}
L
Linus Torvalds 已提交
1958

1959
	return NULL;
1960
}
L
Linus Torvalds 已提交
1961

1962
int ipmi_si_add_smi(struct si_sm_io *io)
1963
{
1964
	int rv = 0;
1965
	struct smi_info *new_smi, *dup;
1966

1967 1968
	if (!io->io_setup) {
		if (io->addr_type == IPMI_IO_ADDR_SPACE) {
1969
			io->io_setup = ipmi_si_port_setup;
1970
		} else if (io->addr_type == IPMI_MEM_ADDR_SPACE) {
1971
			io->io_setup = ipmi_si_mem_setup;
1972 1973 1974 1975 1976
		} else {
			return -EINVAL;
		}
	}

1977
	new_smi = kzalloc(sizeof(*new_smi), GFP_KERNEL);
1978 1979
	if (!new_smi)
		return -ENOMEM;
1980
	spin_lock_init(&new_smi->si_lock);
1981 1982 1983

	new_smi->io = *io;

1984
	mutex_lock(&smi_infos_lock);
1985 1986
	dup = find_dup_si(new_smi);
	if (dup) {
1987 1988
		if (new_smi->io.addr_source == SI_ACPI &&
		    dup->io.addr_source == SI_SMBIOS) {
1989
			/* We prefer ACPI over SMBIOS. */
1990
			dev_info(dup->io.dev,
1991
				 "Removing SMBIOS-specified %s state machine in favor of ACPI\n",
1992
				 si_to_str[new_smi->io.si_type]);
1993 1994
			cleanup_one_si(dup);
		} else {
1995
			dev_info(new_smi->io.dev,
1996
				 "%s-specified %s state machine: duplicate\n",
1997 1998
				 ipmi_addr_src_to_str(new_smi->io.addr_source),
				 si_to_str[new_smi->io.si_type]);
1999
			rv = -EBUSY;
2000
			kfree(new_smi);
2001 2002
			goto out_err;
		}
2003
	}
L
Linus Torvalds 已提交
2004

C
Corey Minyard 已提交
2005
	pr_info(PFX "Adding %s-specified %s state machine\n",
2006 2007
		ipmi_addr_src_to_str(new_smi->io.addr_source),
		si_to_str[new_smi->io.si_type]);
2008 2009 2010

	list_add_tail(&new_smi->link, &smi_infos);

2011 2012 2013 2014
	if (initialized) {
		rv = try_smi_init(new_smi);
		if (rv) {
			cleanup_one_si(new_smi);
2015
			mutex_unlock(&smi_infos_lock);
2016 2017 2018
			return rv;
		}
	}
2019 2020 2021 2022 2023
out_err:
	mutex_unlock(&smi_infos_lock);
	return rv;
}

T
Tony Camuso 已提交
2024 2025 2026 2027 2028
/*
 * Try to start up an interface.  Must be called with smi_infos_lock
 * held, primarily to keep smi_num consistent, we only one to do these
 * one at a time.
 */
2029 2030 2031 2032
static int try_smi_init(struct smi_info *new_smi)
{
	int rv = 0;
	int i;
2033
	char *init_name = NULL;
2034

C
Corey Minyard 已提交
2035
	pr_info(PFX "Trying %s-specified %s state machine at %s address 0x%lx, slave address 0x%x, irq %d\n",
2036 2037
		ipmi_addr_src_to_str(new_smi->io.addr_source),
		si_to_str[new_smi->io.si_type],
C
Corey Minyard 已提交
2038 2039
		addr_space_to_str[new_smi->io.addr_type],
		new_smi->io.addr_data,
2040
		new_smi->io.slave_addr, new_smi->io.irq);
2041

2042
	switch (new_smi->io.si_type) {
2043
	case SI_KCS:
L
Linus Torvalds 已提交
2044
		new_smi->handlers = &kcs_smi_handlers;
2045 2046 2047
		break;

	case SI_SMIC:
L
Linus Torvalds 已提交
2048
		new_smi->handlers = &smic_smi_handlers;
2049 2050 2051
		break;

	case SI_BT:
L
Linus Torvalds 已提交
2052
		new_smi->handlers = &bt_smi_handlers;
2053 2054 2055
		break;

	default:
L
Linus Torvalds 已提交
2056 2057 2058 2059 2060
		/* No support for anything else yet. */
		rv = -EIO;
		goto out_err;
	}

T
Tony Camuso 已提交
2061 2062
	new_smi->intf_num = smi_num;

2063
	/* Do this early so it's available for logs. */
2064
	if (!new_smi->io.dev) {
T
Tony Camuso 已提交
2065 2066
		init_name = kasprintf(GFP_KERNEL, "ipmi_si.%d",
				      new_smi->intf_num);
2067 2068 2069 2070 2071 2072 2073 2074 2075

		/*
		 * If we don't already have a device from something
		 * else (like PCI), then register a new one.
		 */
		new_smi->pdev = platform_device_alloc("ipmi_si",
						      new_smi->intf_num);
		if (!new_smi->pdev) {
			pr_err(PFX "Unable to allocate platform device\n");
2076
			rv = -ENOMEM;
2077 2078
			goto out_err;
		}
2079
		new_smi->io.dev = &new_smi->pdev->dev;
2080
		new_smi->io.dev->driver = &ipmi_platform_driver.driver;
2081
		/* Nulled by device_add() */
2082
		new_smi->io.dev->init_name = init_name;
2083 2084
	}

L
Linus Torvalds 已提交
2085 2086
	/* Allocate the state machine's data and initialize it. */
	new_smi->si_sm = kmalloc(new_smi->handlers->size(), GFP_KERNEL);
2087
	if (!new_smi->si_sm) {
L
Linus Torvalds 已提交
2088 2089 2090
		rv = -ENOMEM;
		goto out_err;
	}
2091 2092
	new_smi->io.io_size = new_smi->handlers->init_data(new_smi->si_sm,
							   &new_smi->io);
L
Linus Torvalds 已提交
2093 2094

	/* Now that we know the I/O size, we can set up the I/O. */
2095
	rv = new_smi->io.io_setup(&new_smi->io);
L
Linus Torvalds 已提交
2096
	if (rv) {
2097
		dev_err(new_smi->io.dev, "Could not set up I/O space\n");
L
Linus Torvalds 已提交
2098 2099 2100 2101 2102
		goto out_err;
	}

	/* Do low-level detection first. */
	if (new_smi->handlers->detect(new_smi->si_sm)) {
2103 2104 2105
		if (new_smi->io.addr_source)
			dev_err(new_smi->io.dev,
				"Interface detection failed\n");
L
Linus Torvalds 已提交
2106 2107 2108 2109
		rv = -ENODEV;
		goto out_err;
	}

2110 2111 2112 2113
	/*
	 * Attempt a get device id command.  If it fails, we probably
	 * don't have a BMC here.
	 */
L
Linus Torvalds 已提交
2114
	rv = try_get_dev_id(new_smi);
2115
	if (rv) {
2116 2117 2118
		if (new_smi->io.addr_source)
			dev_err(new_smi->io.dev,
			       "There appears to be no BMC at this location\n");
L
Linus Torvalds 已提交
2119
		goto out_err;
2120
	}
L
Linus Torvalds 已提交
2121

2122
	setup_oem_data_handler(new_smi);
2123
	setup_xaction_handlers(new_smi);
2124
	check_for_broken_irqs(new_smi);
2125

2126
	new_smi->waiting_msg = NULL;
L
Linus Torvalds 已提交
2127 2128
	new_smi->curr_msg = NULL;
	atomic_set(&new_smi->req_events, 0);
C
Corey Minyard 已提交
2129
	new_smi->run_to_completion = false;
2130 2131
	for (i = 0; i < SI_NUM_STATS; i++)
		atomic_set(&new_smi->stats[i], 0);
L
Linus Torvalds 已提交
2132

C
Corey Minyard 已提交
2133
	new_smi->interrupt_disabled = true;
2134
	atomic_set(&new_smi->need_watch, 0);
L
Linus Torvalds 已提交
2135

2136 2137
	rv = try_enable_event_buffer(new_smi);
	if (rv == 0)
C
Corey Minyard 已提交
2138
		new_smi->has_event_buffer = true;
2139

2140 2141 2142 2143
	/*
	 * Start clearing the flags before we enable interrupts or the
	 * timer to avoid racing with the timer.
	 */
2144
	start_clear_flags(new_smi);
2145 2146 2147 2148 2149

	/*
	 * IRQ is defined to be set when non-zero.  req_events will
	 * cause a global flags check that will enable interrupts.
	 */
2150
	if (new_smi->io.irq) {
2151 2152 2153
		new_smi->interrupt_disabled = false;
		atomic_set(&new_smi->req_events, 1);
	}
L
Linus Torvalds 已提交
2154

C
Corey Minyard 已提交
2155
	if (new_smi->pdev && !new_smi->pdev_registered) {
2156
		rv = platform_device_add(new_smi->pdev);
2157
		if (rv) {
2158
			dev_err(new_smi->io.dev,
C
Corey Minyard 已提交
2159 2160
				"Unable to register system interface device: %d\n",
				rv);
2161
			goto out_err;
2162
		}
C
Corey Minyard 已提交
2163
		new_smi->pdev_registered = true;
2164 2165
	}

2166 2167 2168 2169 2170 2171
	dev_set_drvdata(new_smi->io.dev, new_smi);
	rv = device_add_group(new_smi->io.dev, &ipmi_si_dev_attr_group);
	if (rv) {
		dev_err(new_smi->io.dev,
			"Unable to add device attributes: error %d\n",
			rv);
C
Corey Minyard 已提交
2172
		goto out_err;
2173
	}
2174
	new_smi->dev_group_added = true;
2175

L
Linus Torvalds 已提交
2176 2177
	rv = ipmi_register_smi(&handlers,
			       new_smi,
2178 2179
			       new_smi->io.dev,
			       new_smi->io.slave_addr);
L
Linus Torvalds 已提交
2180
	if (rv) {
2181 2182
		dev_err(new_smi->io.dev,
			"Unable to register device: error %d\n",
2183
			rv);
C
Corey Minyard 已提交
2184
		goto out_err;
L
Linus Torvalds 已提交
2185 2186
	}

2187
#ifdef CONFIG_IPMI_PROC_INTERFACE
L
Linus Torvalds 已提交
2188
	rv = ipmi_smi_add_proc_entry(new_smi->intf, "type",
2189
				     &smi_type_proc_ops,
2190
				     new_smi);
L
Linus Torvalds 已提交
2191
	if (rv) {
2192 2193
		dev_err(new_smi->io.dev,
			"Unable to create proc entry: %d\n", rv);
C
Corey Minyard 已提交
2194
		goto out_err;
L
Linus Torvalds 已提交
2195 2196 2197
	}

	rv = ipmi_smi_add_proc_entry(new_smi->intf, "si_stats",
2198
				     &smi_si_stats_proc_ops,
2199
				     new_smi);
L
Linus Torvalds 已提交
2200
	if (rv) {
2201 2202
		dev_err(new_smi->io.dev,
			"Unable to create proc entry: %d\n", rv);
C
Corey Minyard 已提交
2203
		goto out_err;
L
Linus Torvalds 已提交
2204 2205
	}

2206
	rv = ipmi_smi_add_proc_entry(new_smi->intf, "params",
2207
				     &smi_params_proc_ops,
2208
				     new_smi);
2209
	if (rv) {
2210 2211
		dev_err(new_smi->io.dev,
			"Unable to create proc entry: %d\n", rv);
C
Corey Minyard 已提交
2212
		goto out_err;
2213
	}
2214
#endif
2215

T
Tony Camuso 已提交
2216 2217 2218
	/* Don't increment till we know we have succeeded. */
	smi_num++;

2219 2220
	dev_info(new_smi->io.dev, "IPMI %s interface initialized\n",
		 si_to_str[new_smi->io.si_type]);
L
Linus Torvalds 已提交
2221

2222
	WARN_ON(new_smi->io.dev->init_name != NULL);
2223 2224
	kfree(init_name);

L
Linus Torvalds 已提交
2225 2226
	return 0;

2227
out_err:
C
Corey Minyard 已提交
2228
	shutdown_one_si(new_smi);
2229

2230 2231
	kfree(init_name);

L
Linus Torvalds 已提交
2232 2233 2234
	return rv;
}

B
Bill Pemberton 已提交
2235
static int init_ipmi_si(void)
L
Linus Torvalds 已提交
2236
{
2237
	struct smi_info *e;
2238
	enum ipmi_addr_src type = SI_INVALID;
L
Linus Torvalds 已提交
2239 2240 2241 2242

	if (initialized)
		return 0;

C
Corey Minyard 已提交
2243
	pr_info("IPMI System Interface driver.\n");
L
Linus Torvalds 已提交
2244

2245
	/* If the user gave us a device, they presumably want us to use it */
2246 2247
	if (!ipmi_si_hardcode_find_bmc())
		goto do_scan;
2248

2249 2250
	ipmi_si_platform_init();

2251
	ipmi_si_pci_init();
2252

2253
	ipmi_si_parisc_init();
2254

2255 2256 2257 2258
	/* We prefer devices with interrupts, but in the case of a machine
	   with multiple BMCs we assume that there will be several instances
	   of a given type so if we succeed in registering a type then also
	   try to register everything else of the same type */
2259
do_scan:
2260 2261
	mutex_lock(&smi_infos_lock);
	list_for_each_entry(e, &smi_infos, link) {
2262 2263 2264
		/* Try to register a device if it has an IRQ and we either
		   haven't successfully registered a device yet or this
		   device has the same type as one we successfully registered */
2265
		if (e->io.irq && (!type || e->io.addr_source == type)) {
2266
			if (!try_smi_init(e)) {
2267
				type = e->io.addr_source;
2268 2269 2270 2271
			}
		}
	}

2272
	/* type will only have been set if we successfully registered an si */
2273 2274
	if (type)
		goto skip_fallback_noirq;
2275

2276 2277 2278
	/* Fall back to the preferred device */

	list_for_each_entry(e, &smi_infos, link) {
2279
		if (!e->io.irq && (!type || e->io.addr_source == type)) {
2280
			if (!try_smi_init(e)) {
2281
				type = e->io.addr_source;
2282 2283
			}
		}
2284
	}
2285 2286 2287

skip_fallback_noirq:
	initialized = 1;
2288 2289
	mutex_unlock(&smi_infos_lock);

2290 2291 2292
	if (type)
		return 0;

2293
	mutex_lock(&smi_infos_lock);
2294
	if (unload_when_empty && list_empty(&smi_infos)) {
2295
		mutex_unlock(&smi_infos_lock);
2296
		cleanup_ipmi_si();
C
Corey Minyard 已提交
2297
		pr_warn(PFX "Unable to find any System Interface(s)\n");
L
Linus Torvalds 已提交
2298
		return -ENODEV;
2299
	} else {
2300
		mutex_unlock(&smi_infos_lock);
2301
		return 0;
L
Linus Torvalds 已提交
2302 2303 2304 2305
	}
}
module_init(init_ipmi_si);

2306
static void shutdown_smi(void *send_info)
L
Linus Torvalds 已提交
2307
{
2308
	struct smi_info *smi_info = send_info;
2309

C
Corey Minyard 已提交
2310 2311 2312 2313 2314 2315
	if (smi_info->dev_group_added) {
		device_remove_group(smi_info->io.dev, &ipmi_si_dev_attr_group);
		smi_info->dev_group_added = false;
	}
	if (smi_info->io.dev)
		dev_set_drvdata(smi_info->io.dev, NULL);
2316

2317
	/*
2318 2319
	 * Make sure that interrupts, the timer and the thread are
	 * stopped and will not run again.
2320
	 */
C
Corey Minyard 已提交
2321 2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333
	smi_info->interrupt_disabled = true;
	if (smi_info->io.irq_cleanup) {
		smi_info->io.irq_cleanup(&smi_info->io);
		smi_info->io.irq_cleanup = NULL;
	}
	stop_timer_and_thread(smi_info);

	/*
	 * Wait until we know that we are out of any interrupt
	 * handlers might have been running before we freed the
	 * interrupt.
	 */
	synchronize_sched();
L
Linus Torvalds 已提交
2334

2335 2336
	/*
	 * Timeouts are stopped, now make sure the interrupts are off
2337 2338
	 * in the BMC.  Note that timers and CPU interrupts are off,
	 * so no need for locks.
2339
	 */
C
Corey Minyard 已提交
2340 2341
	while (smi_info->curr_msg || (smi_info->si_state != SI_NORMAL)) {
		poll(smi_info);
C
Corey Minyard 已提交
2342 2343
		schedule_timeout_uninterruptible(1);
	}
C
Corey Minyard 已提交
2344 2345 2346 2347
	if (smi_info->handlers)
		disable_si_irq(smi_info);
	while (smi_info->curr_msg || (smi_info->si_state != SI_NORMAL)) {
		poll(smi_info);
2348
		schedule_timeout_uninterruptible(1);
L
Linus Torvalds 已提交
2349
	}
C
Corey Minyard 已提交
2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360
	if (smi_info->handlers)
		smi_info->handlers->cleanup(smi_info->si_sm);

	if (smi_info->io.addr_source_cleanup) {
		smi_info->io.addr_source_cleanup(&smi_info->io);
		smi_info->io.addr_source_cleanup = NULL;
	}
	if (smi_info->io.io_cleanup) {
		smi_info->io.io_cleanup(&smi_info->io);
		smi_info->io.io_cleanup = NULL;
	}
L
Linus Torvalds 已提交
2361

C
Corey Minyard 已提交
2362 2363 2364 2365
	kfree(smi_info->si_sm);
	smi_info->si_sm = NULL;
}

2366 2367 2368 2369 2370 2371 2372 2373
static void shutdown_one_si(struct smi_info *smi_info)
{
	ipmi_smi_t intf = smi_info->intf;

	if (!intf)
		return;

	smi_info->intf = NULL;
2374
	ipmi_unregister_smi(intf);
2375 2376
}

C
Corey Minyard 已提交
2377 2378 2379 2380
static void cleanup_one_si(struct smi_info *smi_info)
{
	if (!smi_info)
		return;
L
Linus Torvalds 已提交
2381

C
Corey Minyard 已提交
2382
	list_del(&smi_info->link);
L
Linus Torvalds 已提交
2383

C
Corey Minyard 已提交
2384
	shutdown_one_si(smi_info);
2385

C
Corey Minyard 已提交
2386 2387 2388 2389 2390 2391
	if (smi_info->pdev) {
		if (smi_info->pdev_registered)
			platform_device_unregister(smi_info->pdev);
		else
			platform_device_put(smi_info->pdev);
	}
2392

C
Corey Minyard 已提交
2393
	kfree(smi_info);
L
Linus Torvalds 已提交
2394 2395
}

2396 2397 2398 2399 2400 2401 2402 2403 2404 2405 2406 2407 2408 2409 2410 2411 2412 2413
int ipmi_si_remove_by_dev(struct device *dev)
{
	struct smi_info *e;
	int rv = -ENOENT;

	mutex_lock(&smi_infos_lock);
	list_for_each_entry(e, &smi_infos, link) {
		if (e->io.dev == dev) {
			cleanup_one_si(e);
			rv = 0;
			break;
		}
	}
	mutex_unlock(&smi_infos_lock);

	return rv;
}

2414 2415 2416 2417 2418 2419 2420 2421 2422 2423 2424 2425 2426 2427 2428 2429 2430 2431
void ipmi_si_remove_by_data(int addr_space, enum si_type si_type,
			    unsigned long addr)
{
	/* remove */
	struct smi_info *e, *tmp_e;

	mutex_lock(&smi_infos_lock);
	list_for_each_entry_safe(e, tmp_e, &smi_infos, link) {
		if (e->io.addr_type != addr_space)
			continue;
		if (e->io.si_type != si_type)
			continue;
		if (e->io.addr_data == addr)
			cleanup_one_si(e);
	}
	mutex_unlock(&smi_infos_lock);
}

2432
static void cleanup_ipmi_si(void)
L
Linus Torvalds 已提交
2433
{
2434
	struct smi_info *e, *tmp_e;
L
Linus Torvalds 已提交
2435

2436
	if (!initialized)
L
Linus Torvalds 已提交
2437 2438
		return;

2439
	ipmi_si_pci_shutdown();
2440 2441

	ipmi_si_parisc_shutdown();
2442

2443
	ipmi_si_platform_shutdown();
2444

2445
	mutex_lock(&smi_infos_lock);
2446 2447
	list_for_each_entry_safe(e, tmp_e, &smi_infos, link)
		cleanup_one_si(e);
2448
	mutex_unlock(&smi_infos_lock);
L
Linus Torvalds 已提交
2449 2450 2451
}
module_exit(cleanup_ipmi_si);

2452
MODULE_ALIAS("platform:dmi-ipmi-si");
L
Linus Torvalds 已提交
2453
MODULE_LICENSE("GPL");
2454
MODULE_AUTHOR("Corey Minyard <minyard@mvista.com>");
2455 2456
MODULE_DESCRIPTION("Interface to the IPMI driver for the KCS, SMIC, and BT"
		   " system interfaces.");