qla_isr.c 80.8 KB
Newer Older
L
Linus Torvalds 已提交
1
/*
A
Andrew Vasquez 已提交
2
 * QLogic Fibre Channel HBA Driver
3
 * Copyright (c)  2003-2013 QLogic Corporation
L
Linus Torvalds 已提交
4
 *
A
Andrew Vasquez 已提交
5
 * See LICENSE.qla2xxx for copyright and licensing details.
L
Linus Torvalds 已提交
6 7
 */
#include "qla_def.h"
8
#include "qla_target.h"
L
Linus Torvalds 已提交
9

10
#include <linux/delay.h>
11
#include <linux/slab.h>
12
#include <scsi/scsi_tcq.h>
13
#include <scsi/scsi_bsg_fc.h>
14
#include <scsi/scsi_eh.h>
15

L
Linus Torvalds 已提交
16
static void qla2x00_mbx_completion(scsi_qla_host_t *, uint16_t);
17
static void qla2x00_status_entry(scsi_qla_host_t *, struct rsp_que *, void *);
18
static void qla2x00_status_cont_entry(struct rsp_que *, sts_cont_entry_t *);
19 20
static void qla2x00_error_entry(scsi_qla_host_t *, struct rsp_que *,
	sts_entry_t *);
21

L
Linus Torvalds 已提交
22 23 24 25 26 27 28 29 30 31
/**
 * qla2100_intr_handler() - Process interrupts for the ISP2100 and ISP2200.
 * @irq:
 * @dev_id: SCSI driver HA context
 *
 * Called by system whenever the host adapter generates an interrupt.
 *
 * Returns handled flag.
 */
irqreturn_t
32
qla2100_intr_handler(int irq, void *dev_id)
L
Linus Torvalds 已提交
33
{
34 35
	scsi_qla_host_t	*vha;
	struct qla_hw_data *ha;
36
	struct device_reg_2xxx __iomem *reg;
L
Linus Torvalds 已提交
37 38
	int		status;
	unsigned long	iter;
39
	uint16_t	hccr;
40
	uint16_t	mb[4];
41
	struct rsp_que *rsp;
42
	unsigned long	flags;
L
Linus Torvalds 已提交
43

44 45
	rsp = (struct rsp_que *) dev_id;
	if (!rsp) {
46 47
		ql_log(ql_log_info, NULL, 0x505d,
		    "%s: NULL response queue pointer.\n", __func__);
L
Linus Torvalds 已提交
48 49 50
		return (IRQ_NONE);
	}

51
	ha = rsp->hw;
52
	reg = &ha->iobase->isp;
L
Linus Torvalds 已提交
53 54
	status = 0;

55
	spin_lock_irqsave(&ha->hardware_lock, flags);
56
	vha = pci_get_drvdata(ha->pdev);
L
Linus Torvalds 已提交
57
	for (iter = 50; iter--; ) {
58 59 60 61 62 63 64
		hccr = RD_REG_WORD(&reg->hccr);
		if (hccr & HCCR_RISC_PAUSE) {
			if (pci_channel_offline(ha->pdev))
				break;

			/*
			 * Issue a "HARD" reset in order for the RISC interrupt
65
			 * bit to be cleared.  Schedule a big hammer to get
66 67 68 69 70
			 * out of the RISC PAUSED state.
			 */
			WRT_REG_WORD(&reg->hccr, HCCR_RESET_RISC);
			RD_REG_WORD(&reg->hccr);

71 72
			ha->isp_ops->fw_dump(vha, 1);
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
73 74
			break;
		} else if ((RD_REG_WORD(&reg->istatus) & ISR_RISC_INT) == 0)
L
Linus Torvalds 已提交
75 76 77 78 79 80 81
			break;

		if (RD_REG_WORD(&reg->semaphore) & BIT_0) {
			WRT_REG_WORD(&reg->hccr, HCCR_CLR_RISC_INT);
			RD_REG_WORD(&reg->hccr);

			/* Get mailbox data. */
82 83
			mb[0] = RD_MAILBOX_REG(ha, reg, 0);
			if (mb[0] > 0x3fff && mb[0] < 0x8000) {
84
				qla2x00_mbx_completion(vha, mb[0]);
L
Linus Torvalds 已提交
85
				status |= MBX_INTERRUPT;
86 87 88 89
			} else if (mb[0] > 0x7fff && mb[0] < 0xc000) {
				mb[1] = RD_MAILBOX_REG(ha, reg, 1);
				mb[2] = RD_MAILBOX_REG(ha, reg, 2);
				mb[3] = RD_MAILBOX_REG(ha, reg, 3);
90
				qla2x00_async_event(vha, rsp, mb);
L
Linus Torvalds 已提交
91 92
			} else {
				/*EMPTY*/
93 94 95
				ql_dbg(ql_dbg_async, vha, 0x5025,
				    "Unrecognized interrupt type (%d).\n",
				    mb[0]);
L
Linus Torvalds 已提交
96 97 98 99 100
			}
			/* Release mailbox registers. */
			WRT_REG_WORD(&reg->semaphore, 0);
			RD_REG_WORD(&reg->semaphore);
		} else {
101
			qla2x00_process_response_queue(rsp);
L
Linus Torvalds 已提交
102 103 104 105 106

			WRT_REG_WORD(&reg->hccr, HCCR_CLR_RISC_INT);
			RD_REG_WORD(&reg->hccr);
		}
	}
107
	qla2x00_handle_mbx_completion(ha, status);
108
	spin_unlock_irqrestore(&ha->hardware_lock, flags);
L
Linus Torvalds 已提交
109 110 111 112 113 114 115 116 117 118 119 120 121 122

	return (IRQ_HANDLED);
}

/**
 * qla2300_intr_handler() - Process interrupts for the ISP23xx and ISP63xx.
 * @irq:
 * @dev_id: SCSI driver HA context
 *
 * Called by system whenever the host adapter generates an interrupt.
 *
 * Returns handled flag.
 */
irqreturn_t
123
qla2300_intr_handler(int irq, void *dev_id)
L
Linus Torvalds 已提交
124
{
125
	scsi_qla_host_t	*vha;
126
	struct device_reg_2xxx __iomem *reg;
L
Linus Torvalds 已提交
127 128 129 130
	int		status;
	unsigned long	iter;
	uint32_t	stat;
	uint16_t	hccr;
131
	uint16_t	mb[4];
132 133
	struct rsp_que *rsp;
	struct qla_hw_data *ha;
134
	unsigned long	flags;
L
Linus Torvalds 已提交
135

136 137
	rsp = (struct rsp_que *) dev_id;
	if (!rsp) {
138 139
		ql_log(ql_log_info, NULL, 0x5058,
		    "%s: NULL response queue pointer.\n", __func__);
L
Linus Torvalds 已提交
140 141 142
		return (IRQ_NONE);
	}

143
	ha = rsp->hw;
144
	reg = &ha->iobase->isp;
L
Linus Torvalds 已提交
145 146
	status = 0;

147
	spin_lock_irqsave(&ha->hardware_lock, flags);
148
	vha = pci_get_drvdata(ha->pdev);
L
Linus Torvalds 已提交
149 150 151
	for (iter = 50; iter--; ) {
		stat = RD_REG_DWORD(&reg->u.isp2300.host_status);
		if (stat & HSR_RISC_PAUSED) {
152
			if (unlikely(pci_channel_offline(ha->pdev)))
153 154
				break;

L
Linus Torvalds 已提交
155 156
			hccr = RD_REG_WORD(&reg->hccr);
			if (hccr & (BIT_15 | BIT_13 | BIT_11 | BIT_8))
157 158 159
				ql_log(ql_log_warn, vha, 0x5026,
				    "Parity error -- HCCR=%x, Dumping "
				    "firmware.\n", hccr);
L
Linus Torvalds 已提交
160
			else
161 162 163
				ql_log(ql_log_warn, vha, 0x5027,
				    "RISC paused -- HCCR=%x, Dumping "
				    "firmware.\n", hccr);
L
Linus Torvalds 已提交
164 165 166 167

			/*
			 * Issue a "HARD" reset in order for the RISC
			 * interrupt bit to be cleared.  Schedule a big
168
			 * hammer to get out of the RISC PAUSED state.
L
Linus Torvalds 已提交
169 170 171
			 */
			WRT_REG_WORD(&reg->hccr, HCCR_RESET_RISC);
			RD_REG_WORD(&reg->hccr);
172

173 174
			ha->isp_ops->fw_dump(vha, 1);
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
175 176 177 178 179 180 181 182 183
			break;
		} else if ((stat & HSR_RISC_INT) == 0)
			break;

		switch (stat & 0xff) {
		case 0x1:
		case 0x2:
		case 0x10:
		case 0x11:
184
			qla2x00_mbx_completion(vha, MSW(stat));
L
Linus Torvalds 已提交
185 186 187 188 189 190
			status |= MBX_INTERRUPT;

			/* Release mailbox registers. */
			WRT_REG_WORD(&reg->semaphore, 0);
			break;
		case 0x12:
191 192 193 194
			mb[0] = MSW(stat);
			mb[1] = RD_MAILBOX_REG(ha, reg, 1);
			mb[2] = RD_MAILBOX_REG(ha, reg, 2);
			mb[3] = RD_MAILBOX_REG(ha, reg, 3);
195
			qla2x00_async_event(vha, rsp, mb);
196 197
			break;
		case 0x13:
198
			qla2x00_process_response_queue(rsp);
L
Linus Torvalds 已提交
199 200
			break;
		case 0x15:
201 202
			mb[0] = MBA_CMPLT_1_16BIT;
			mb[1] = MSW(stat);
203
			qla2x00_async_event(vha, rsp, mb);
L
Linus Torvalds 已提交
204 205
			break;
		case 0x16:
206 207 208
			mb[0] = MBA_SCSI_COMPLETION;
			mb[1] = MSW(stat);
			mb[2] = RD_MAILBOX_REG(ha, reg, 2);
209
			qla2x00_async_event(vha, rsp, mb);
L
Linus Torvalds 已提交
210 211
			break;
		default:
212 213
			ql_dbg(ql_dbg_async, vha, 0x5028,
			    "Unrecognized interrupt type (%d).\n", stat & 0xff);
L
Linus Torvalds 已提交
214 215 216 217 218
			break;
		}
		WRT_REG_WORD(&reg->hccr, HCCR_CLR_RISC_INT);
		RD_REG_WORD_RELAXED(&reg->hccr);
	}
219
	qla2x00_handle_mbx_completion(ha, status);
220
	spin_unlock_irqrestore(&ha->hardware_lock, flags);
L
Linus Torvalds 已提交
221 222 223 224 225 226 227 228 229 230

	return (IRQ_HANDLED);
}

/**
 * qla2x00_mbx_completion() - Process mailbox command completions.
 * @ha: SCSI driver HA context
 * @mb0: Mailbox0 register
 */
static void
231
qla2x00_mbx_completion(scsi_qla_host_t *vha, uint16_t mb0)
L
Linus Torvalds 已提交
232 233
{
	uint16_t	cnt;
234
	uint32_t	mboxes;
L
Linus Torvalds 已提交
235
	uint16_t __iomem *wptr;
236
	struct qla_hw_data *ha = vha->hw;
237
	struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
L
Linus Torvalds 已提交
238

239 240 241
	/* Read all mbox registers? */
	mboxes = (1 << ha->mbx_count) - 1;
	if (!ha->mcp)
242
		ql_dbg(ql_dbg_async, vha, 0x5001, "MBX pointer ERROR.\n");
243 244 245
	else
		mboxes = ha->mcp->in_mb;

L
Linus Torvalds 已提交
246 247 248
	/* Load return mailbox registers. */
	ha->flags.mbox_int = 1;
	ha->mailbox_out[0] = mb0;
249
	mboxes >>= 1;
L
Linus Torvalds 已提交
250 251 252
	wptr = (uint16_t __iomem *)MAILBOX_REG(ha, reg, 1);

	for (cnt = 1; cnt < ha->mbx_count; cnt++) {
A
Andrew Vasquez 已提交
253
		if (IS_QLA2200(ha) && cnt == 8)
L
Linus Torvalds 已提交
254
			wptr = (uint16_t __iomem *)MAILBOX_REG(ha, reg, 8);
255
		if ((cnt == 4 || cnt == 5) && (mboxes & BIT_0))
L
Linus Torvalds 已提交
256
			ha->mailbox_out[cnt] = qla2x00_debounce_register(wptr);
257
		else if (mboxes & BIT_0)
L
Linus Torvalds 已提交
258
			ha->mailbox_out[cnt] = RD_REG_WORD(wptr);
A
Andrew Vasquez 已提交
259

L
Linus Torvalds 已提交
260
		wptr++;
261
		mboxes >>= 1;
L
Linus Torvalds 已提交
262 263 264
	}
}

265 266 267 268 269 270 271 272 273 274 275 276 277 278 279
static void
qla81xx_idc_event(scsi_qla_host_t *vha, uint16_t aen, uint16_t descr)
{
	static char *event[] =
		{ "Complete", "Request Notification", "Time Extension" };
	int rval;
	struct device_reg_24xx __iomem *reg24 = &vha->hw->iobase->isp24;
	uint16_t __iomem *wptr;
	uint16_t cnt, timeout, mb[QLA_IDC_ACK_REGS];

	/* Seed data -- mailbox1 -> mailbox7. */
	wptr = (uint16_t __iomem *)&reg24->mailbox1;
	for (cnt = 0; cnt < QLA_IDC_ACK_REGS; cnt++, wptr++)
		mb[cnt] = RD_REG_WORD(wptr);

280
	ql_dbg(ql_dbg_async, vha, 0x5021,
281
	    "Inter-Driver Communication %s -- "
282 283 284
	    "%04x %04x %04x %04x %04x %04x %04x.\n",
	    event[aen & 0xff], mb[0], mb[1], mb[2], mb[3],
	    mb[4], mb[5], mb[6]);
285 286 287 288 289
	if ((aen == MBA_IDC_COMPLETE && mb[1] >> 15)) {
		vha->hw->flags.idc_compl_status = 1;
		if (vha->hw->notify_dcbx_comp)
			complete(&vha->hw->dcbx_comp);
	}
290

291 292 293 294
	/* Acknowledgement needed? [Notify && non-zero timeout]. */
	timeout = (descr >> 8) & 0xf;
	if (aen != MBA_IDC_NOTIFY || !timeout)
		return;
295

296 297 298
	ql_dbg(ql_dbg_async, vha, 0x5022,
	    "%lu Inter-Driver Communication %s -- ACK timeout=%d.\n",
	    vha->host_no, event[aen & 0xff], timeout);
299 300 301

	rval = qla2x00_post_idc_ack_work(vha, mb);
	if (rval != QLA_SUCCESS)
302
		ql_log(ql_log_warn, vha, 0x5023,
303 304 305
		    "IDC failed to post ACK.\n");
}

306
#define LS_UNKNOWN	2
307 308
const char *
qla2x00_get_link_speed_str(struct qla_hw_data *ha, uint16_t speed)
309
{
310 311 312
	static const char * const link_speeds[] = {
		"1", "2", "?", "4", "8", "16", "10"
	};
313 314

	if (IS_QLA2100(ha) || IS_QLA2200(ha))
315 316 317 318 319 320 321
		return link_speeds[0];
	else if (speed == 0x13)
		return link_speeds[6];
	else if (speed < 6)
		return link_speeds[speed];
	else
		return link_speeds[LS_UNKNOWN];
322 323
}

324
static void
325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479
qla83xx_handle_8200_aen(scsi_qla_host_t *vha, uint16_t *mb)
{
	struct qla_hw_data *ha = vha->hw;

	/*
	 * 8200 AEN Interpretation:
	 * mb[0] = AEN code
	 * mb[1] = AEN Reason code
	 * mb[2] = LSW of Peg-Halt Status-1 Register
	 * mb[6] = MSW of Peg-Halt Status-1 Register
	 * mb[3] = LSW of Peg-Halt Status-2 register
	 * mb[7] = MSW of Peg-Halt Status-2 register
	 * mb[4] = IDC Device-State Register value
	 * mb[5] = IDC Driver-Presence Register value
	 */
	ql_dbg(ql_dbg_async, vha, 0x506b, "AEN Code: mb[0] = 0x%x AEN reason: "
	    "mb[1] = 0x%x PH-status1: mb[2] = 0x%x PH-status1: mb[6] = 0x%x.\n",
	    mb[0], mb[1], mb[2], mb[6]);
	ql_dbg(ql_dbg_async, vha, 0x506c, "PH-status2: mb[3] = 0x%x "
	    "PH-status2: mb[7] = 0x%x Device-State: mb[4] = 0x%x "
	    "Drv-Presence: mb[5] = 0x%x.\n", mb[3], mb[7], mb[4], mb[5]);

	if (mb[1] & (IDC_PEG_HALT_STATUS_CHANGE | IDC_NIC_FW_REPORTED_FAILURE |
				IDC_HEARTBEAT_FAILURE)) {
		ha->flags.nic_core_hung = 1;
		ql_log(ql_log_warn, vha, 0x5060,
		    "83XX: F/W Error Reported: Check if reset required.\n");

		if (mb[1] & IDC_PEG_HALT_STATUS_CHANGE) {
			uint32_t protocol_engine_id, fw_err_code, err_level;

			/*
			 * IDC_PEG_HALT_STATUS_CHANGE interpretation:
			 *  - PEG-Halt Status-1 Register:
			 *	(LSW = mb[2], MSW = mb[6])
			 *	Bits 0-7   = protocol-engine ID
			 *	Bits 8-28  = f/w error code
			 *	Bits 29-31 = Error-level
			 *	    Error-level 0x1 = Non-Fatal error
			 *	    Error-level 0x2 = Recoverable Fatal error
			 *	    Error-level 0x4 = UnRecoverable Fatal error
			 *  - PEG-Halt Status-2 Register:
			 *	(LSW = mb[3], MSW = mb[7])
			 */
			protocol_engine_id = (mb[2] & 0xff);
			fw_err_code = (((mb[2] & 0xff00) >> 8) |
			    ((mb[6] & 0x1fff) << 8));
			err_level = ((mb[6] & 0xe000) >> 13);
			ql_log(ql_log_warn, vha, 0x5061, "PegHalt Status-1 "
			    "Register: protocol_engine_id=0x%x "
			    "fw_err_code=0x%x err_level=0x%x.\n",
			    protocol_engine_id, fw_err_code, err_level);
			ql_log(ql_log_warn, vha, 0x5062, "PegHalt Status-2 "
			    "Register: 0x%x%x.\n", mb[7], mb[3]);
			if (err_level == ERR_LEVEL_NON_FATAL) {
				ql_log(ql_log_warn, vha, 0x5063,
				    "Not a fatal error, f/w has recovered "
				    "iteself.\n");
			} else if (err_level == ERR_LEVEL_RECOVERABLE_FATAL) {
				ql_log(ql_log_fatal, vha, 0x5064,
				    "Recoverable Fatal error: Chip reset "
				    "required.\n");
				qla83xx_schedule_work(vha,
				    QLA83XX_NIC_CORE_RESET);
			} else if (err_level == ERR_LEVEL_UNRECOVERABLE_FATAL) {
				ql_log(ql_log_fatal, vha, 0x5065,
				    "Unrecoverable Fatal error: Set FAILED "
				    "state, reboot required.\n");
				qla83xx_schedule_work(vha,
				    QLA83XX_NIC_CORE_UNRECOVERABLE);
			}
		}

		if (mb[1] & IDC_NIC_FW_REPORTED_FAILURE) {
			uint16_t peg_fw_state, nw_interface_link_up;
			uint16_t nw_interface_signal_detect, sfp_status;
			uint16_t htbt_counter, htbt_monitor_enable;
			uint16_t sfp_additonal_info, sfp_multirate;
			uint16_t sfp_tx_fault, link_speed, dcbx_status;

			/*
			 * IDC_NIC_FW_REPORTED_FAILURE interpretation:
			 *  - PEG-to-FC Status Register:
			 *	(LSW = mb[2], MSW = mb[6])
			 *	Bits 0-7   = Peg-Firmware state
			 *	Bit 8      = N/W Interface Link-up
			 *	Bit 9      = N/W Interface signal detected
			 *	Bits 10-11 = SFP Status
			 *	  SFP Status 0x0 = SFP+ transceiver not expected
			 *	  SFP Status 0x1 = SFP+ transceiver not present
			 *	  SFP Status 0x2 = SFP+ transceiver invalid
			 *	  SFP Status 0x3 = SFP+ transceiver present and
			 *	  valid
			 *	Bits 12-14 = Heartbeat Counter
			 *	Bit 15     = Heartbeat Monitor Enable
			 *	Bits 16-17 = SFP Additional Info
			 *	  SFP info 0x0 = Unregocnized transceiver for
			 *	  Ethernet
			 *	  SFP info 0x1 = SFP+ brand validation failed
			 *	  SFP info 0x2 = SFP+ speed validation failed
			 *	  SFP info 0x3 = SFP+ access error
			 *	Bit 18     = SFP Multirate
			 *	Bit 19     = SFP Tx Fault
			 *	Bits 20-22 = Link Speed
			 *	Bits 23-27 = Reserved
			 *	Bits 28-30 = DCBX Status
			 *	  DCBX Status 0x0 = DCBX Disabled
			 *	  DCBX Status 0x1 = DCBX Enabled
			 *	  DCBX Status 0x2 = DCBX Exchange error
			 *	Bit 31     = Reserved
			 */
			peg_fw_state = (mb[2] & 0x00ff);
			nw_interface_link_up = ((mb[2] & 0x0100) >> 8);
			nw_interface_signal_detect = ((mb[2] & 0x0200) >> 9);
			sfp_status = ((mb[2] & 0x0c00) >> 10);
			htbt_counter = ((mb[2] & 0x7000) >> 12);
			htbt_monitor_enable = ((mb[2] & 0x8000) >> 15);
			sfp_additonal_info = (mb[6] & 0x0003);
			sfp_multirate = ((mb[6] & 0x0004) >> 2);
			sfp_tx_fault = ((mb[6] & 0x0008) >> 3);
			link_speed = ((mb[6] & 0x0070) >> 4);
			dcbx_status = ((mb[6] & 0x7000) >> 12);

			ql_log(ql_log_warn, vha, 0x5066,
			    "Peg-to-Fc Status Register:\n"
			    "peg_fw_state=0x%x, nw_interface_link_up=0x%x, "
			    "nw_interface_signal_detect=0x%x"
			    "\nsfp_statis=0x%x.\n ", peg_fw_state,
			    nw_interface_link_up, nw_interface_signal_detect,
			    sfp_status);
			ql_log(ql_log_warn, vha, 0x5067,
			    "htbt_counter=0x%x, htbt_monitor_enable=0x%x, "
			    "sfp_additonal_info=0x%x, sfp_multirate=0x%x.\n ",
			    htbt_counter, htbt_monitor_enable,
			    sfp_additonal_info, sfp_multirate);
			ql_log(ql_log_warn, vha, 0x5068,
			    "sfp_tx_fault=0x%x, link_state=0x%x, "
			    "dcbx_status=0x%x.\n", sfp_tx_fault, link_speed,
			    dcbx_status);

			qla83xx_schedule_work(vha, QLA83XX_NIC_CORE_RESET);
		}

		if (mb[1] & IDC_HEARTBEAT_FAILURE) {
			ql_log(ql_log_warn, vha, 0x5069,
			    "Heartbeat Failure encountered, chip reset "
			    "required.\n");

			qla83xx_schedule_work(vha, QLA83XX_NIC_CORE_RESET);
		}
	}

	if (mb[1] & IDC_DEVICE_STATE_CHANGE) {
		ql_log(ql_log_info, vha, 0x506a,
		    "IDC Device-State changed = 0x%x.\n", mb[4]);
480 481
		if (ha->flags.nic_core_reset_owner)
			return;
482 483 484 485
		qla83xx_schedule_work(vha, MBA_IDC_AEN);
	}
}

486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510
int
qla2x00_is_a_vp_did(scsi_qla_host_t *vha, uint32_t rscn_entry)
{
	struct qla_hw_data *ha = vha->hw;
	scsi_qla_host_t *vp;
	uint32_t vp_did;
	unsigned long flags;
	int ret = 0;

	if (!ha->num_vhosts)
		return ret;

	spin_lock_irqsave(&ha->vport_slock, flags);
	list_for_each_entry(vp, &ha->vp_list, list) {
		vp_did = vp->d_id.b24;
		if (vp_did == rscn_entry) {
			ret = 1;
			break;
		}
	}
	spin_unlock_irqrestore(&ha->vport_slock, flags);

	return ret;
}

L
Linus Torvalds 已提交
511 512 513
/**
 * qla2x00_async_event() - Process aynchronous events.
 * @ha: SCSI driver HA context
514
 * @mb: Mailbox registers (0 - 3)
L
Linus Torvalds 已提交
515
 */
516
void
517
qla2x00_async_event(scsi_qla_host_t *vha, struct rsp_que *rsp, uint16_t *mb)
L
Linus Torvalds 已提交
518 519
{
	uint16_t	handle_cnt;
520
	uint16_t	cnt, mbx;
L
Linus Torvalds 已提交
521
	uint32_t	handles[5];
522
	struct qla_hw_data *ha = vha->hw;
523
	struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
524
	struct device_reg_24xx __iomem *reg24 = &ha->iobase->isp24;
525
	struct device_reg_82xx __iomem *reg82 = &ha->iobase->isp82;
L
Linus Torvalds 已提交
526
	uint32_t	rscn_entry, host_pid;
527
	unsigned long	flags;
L
Linus Torvalds 已提交
528 529 530

	/* Setup to process RIO completion. */
	handle_cnt = 0;
531
	if (IS_CNA_CAPABLE(ha))
532
		goto skip_rio;
L
Linus Torvalds 已提交
533 534
	switch (mb[0]) {
	case MBA_SCSI_COMPLETION:
535
		handles[0] = le32_to_cpu((uint32_t)((mb[2] << 16) | mb[1]));
L
Linus Torvalds 已提交
536 537 538
		handle_cnt = 1;
		break;
	case MBA_CMPLT_1_16BIT:
539
		handles[0] = mb[1];
L
Linus Torvalds 已提交
540 541 542 543
		handle_cnt = 1;
		mb[0] = MBA_SCSI_COMPLETION;
		break;
	case MBA_CMPLT_2_16BIT:
544 545
		handles[0] = mb[1];
		handles[1] = mb[2];
L
Linus Torvalds 已提交
546 547 548 549
		handle_cnt = 2;
		mb[0] = MBA_SCSI_COMPLETION;
		break;
	case MBA_CMPLT_3_16BIT:
550 551 552
		handles[0] = mb[1];
		handles[1] = mb[2];
		handles[2] = mb[3];
L
Linus Torvalds 已提交
553 554 555 556
		handle_cnt = 3;
		mb[0] = MBA_SCSI_COMPLETION;
		break;
	case MBA_CMPLT_4_16BIT:
557 558 559
		handles[0] = mb[1];
		handles[1] = mb[2];
		handles[2] = mb[3];
L
Linus Torvalds 已提交
560 561 562 563 564
		handles[3] = (uint32_t)RD_MAILBOX_REG(ha, reg, 6);
		handle_cnt = 4;
		mb[0] = MBA_SCSI_COMPLETION;
		break;
	case MBA_CMPLT_5_16BIT:
565 566 567
		handles[0] = mb[1];
		handles[1] = mb[2];
		handles[2] = mb[3];
L
Linus Torvalds 已提交
568 569 570 571 572 573
		handles[3] = (uint32_t)RD_MAILBOX_REG(ha, reg, 6);
		handles[4] = (uint32_t)RD_MAILBOX_REG(ha, reg, 7);
		handle_cnt = 5;
		mb[0] = MBA_SCSI_COMPLETION;
		break;
	case MBA_CMPLT_2_32BIT:
574
		handles[0] = le32_to_cpu((uint32_t)((mb[2] << 16) | mb[1]));
L
Linus Torvalds 已提交
575 576 577 578 579 580 581 582 583
		handles[1] = le32_to_cpu(
		    ((uint32_t)(RD_MAILBOX_REG(ha, reg, 7) << 16)) |
		    RD_MAILBOX_REG(ha, reg, 6));
		handle_cnt = 2;
		mb[0] = MBA_SCSI_COMPLETION;
		break;
	default:
		break;
	}
584
skip_rio:
L
Linus Torvalds 已提交
585 586
	switch (mb[0]) {
	case MBA_SCSI_COMPLETION:	/* Fast Post */
587
		if (!vha->flags.online)
L
Linus Torvalds 已提交
588 589 590
			break;

		for (cnt = 0; cnt < handle_cnt; cnt++)
591 592
			qla2x00_process_completed_request(vha, rsp->req,
				handles[cnt]);
L
Linus Torvalds 已提交
593 594 595
		break;

	case MBA_RESET:			/* Reset */
596 597
		ql_dbg(ql_dbg_async, vha, 0x5002,
		    "Asynchronous RESET.\n");
L
Linus Torvalds 已提交
598

599
		set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
600 601 602
		break;

	case MBA_SYSTEM_ERR:		/* System Error */
603 604
		mbx = (IS_QLA81XX(ha) || IS_QLA83XX(ha)) ?
			RD_REG_WORD(&reg24->mailbox7) : 0;
605
		ql_log(ql_log_warn, vha, 0x5003,
606 607
		    "ISP System Error - mbx1=%xh mbx2=%xh mbx3=%xh "
		    "mbx7=%xh.\n", mb[1], mb[2], mb[3], mbx);
L
Linus Torvalds 已提交
608

609
		ha->isp_ops->fw_dump(vha, 1);
L
Linus Torvalds 已提交
610

611
		if (IS_FWI2_CAPABLE(ha)) {
612
			if (mb[1] == 0 && mb[2] == 0) {
613
				ql_log(ql_log_fatal, vha, 0x5004,
614 615
				    "Unrecoverable Hardware Error: adapter "
				    "marked OFFLINE!\n");
616
				vha->flags.online = 0;
617
				vha->device_flags |= DFLG_DEV_FAILED;
618
			} else {
L
Lucas De Marchi 已提交
619
				/* Check to see if MPI timeout occurred */
620 621 622 623
				if ((mbx & MBX_3) && (ha->flags.port0))
					set_bit(MPI_RESET_NEEDED,
					    &vha->dpc_flags);

624
				set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
625
			}
626
		} else if (mb[1] == 0) {
627
			ql_log(ql_log_fatal, vha, 0x5005,
L
Linus Torvalds 已提交
628 629
			    "Unrecoverable Hardware Error: adapter marked "
			    "OFFLINE!\n");
630
			vha->flags.online = 0;
631
			vha->device_flags |= DFLG_DEV_FAILED;
L
Linus Torvalds 已提交
632
		} else
633
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
634 635 636
		break;

	case MBA_REQ_TRANSFER_ERR:	/* Request Transfer Error */
637 638
		ql_log(ql_log_warn, vha, 0x5006,
		    "ISP Request Transfer Error (%x).\n",  mb[1]);
L
Linus Torvalds 已提交
639

640
		set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
641 642 643
		break;

	case MBA_RSP_TRANSFER_ERR:	/* Response Transfer Error */
644 645
		ql_log(ql_log_warn, vha, 0x5007,
		    "ISP Response Transfer Error.\n");
L
Linus Torvalds 已提交
646

647
		set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
648 649 650
		break;

	case MBA_WAKEUP_THRES:		/* Request Queue Wake-up */
651 652
		ql_dbg(ql_dbg_async, vha, 0x5008,
		    "Asynchronous WAKEUP_THRES.\n");
L
Linus Torvalds 已提交
653

654
		break;
L
Linus Torvalds 已提交
655
	case MBA_LIP_OCCURRED:		/* Loop Initialization Procedure */
656
		ql_dbg(ql_dbg_async, vha, 0x5009,
657
		    "LIP occurred (%x).\n", mb[1]);
L
Linus Torvalds 已提交
658

659 660 661 662
		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
			atomic_set(&vha->loop_state, LOOP_DOWN);
			atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
			qla2x00_mark_all_devices_lost(vha, 1);
L
Linus Torvalds 已提交
663 664
		}

665 666 667
		if (vha->vp_idx) {
			atomic_set(&vha->vp_state, VP_FAILED);
			fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
668 669
		}

670 671
		set_bit(REGISTER_FC4_NEEDED, &vha->dpc_flags);
		set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
672

673 674
		vha->flags.management_server_logged_in = 0;
		qla2x00_post_aen_work(vha, FCH_EVT_LIP, mb[1]);
L
Linus Torvalds 已提交
675 676 677
		break;

	case MBA_LOOP_UP:		/* Loop Up Event */
678
		if (IS_QLA2100(ha) || IS_QLA2200(ha))
679
			ha->link_data_rate = PORT_SPEED_1GB;
680
		else
L
Linus Torvalds 已提交
681 682
			ha->link_data_rate = mb[1];

683
		ql_dbg(ql_dbg_async, vha, 0x500a,
684
		    "LOOP UP detected (%s Gbps).\n",
685
		    qla2x00_get_link_speed_str(ha, ha->link_data_rate));
L
Linus Torvalds 已提交
686

687 688
		vha->flags.management_server_logged_in = 0;
		qla2x00_post_aen_work(vha, FCH_EVT_LINKUP, ha->link_data_rate);
L
Linus Torvalds 已提交
689 690 691
		break;

	case MBA_LOOP_DOWN:		/* Loop Down Event */
692 693
		mbx = (IS_QLA81XX(ha) || IS_QLA8031(ha))
			? RD_REG_WORD(&reg24->mailbox4) : 0;
694
		mbx = IS_QLA82XX(ha) ? RD_REG_WORD(&reg82->mailbox_out[4]) : mbx;
695
		ql_dbg(ql_dbg_async, vha, 0x500b,
696 697
		    "LOOP DOWN detected (%x %x %x %x).\n",
		    mb[1], mb[2], mb[3], mbx);
L
Linus Torvalds 已提交
698

699 700 701 702 703
		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
			atomic_set(&vha->loop_state, LOOP_DOWN);
			atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
			vha->device_flags |= DFLG_NO_CABLE;
			qla2x00_mark_all_devices_lost(vha, 1);
L
Linus Torvalds 已提交
704 705
		}

706 707 708
		if (vha->vp_idx) {
			atomic_set(&vha->vp_state, VP_FAILED);
			fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
709 710
		}

711
		vha->flags.management_server_logged_in = 0;
712
		ha->link_data_rate = PORT_SPEED_UNKNOWN;
713
		qla2x00_post_aen_work(vha, FCH_EVT_LINKDOWN, 0);
L
Linus Torvalds 已提交
714 715 716
		break;

	case MBA_LIP_RESET:		/* LIP reset occurred */
717
		ql_dbg(ql_dbg_async, vha, 0x500c,
718
		    "LIP reset occurred (%x).\n", mb[1]);
L
Linus Torvalds 已提交
719

720 721 722 723
		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
			atomic_set(&vha->loop_state, LOOP_DOWN);
			atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME);
			qla2x00_mark_all_devices_lost(vha, 1);
L
Linus Torvalds 已提交
724 725
		}

726 727 728
		if (vha->vp_idx) {
			atomic_set(&vha->vp_state, VP_FAILED);
			fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
729 730
		}

731
		set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
732 733

		ha->operating_mode = LOOP;
734 735
		vha->flags.management_server_logged_in = 0;
		qla2x00_post_aen_work(vha, FCH_EVT_LIPRESET, mb[1]);
L
Linus Torvalds 已提交
736 737
		break;

738
	/* case MBA_DCBX_COMPLETE: */
L
Linus Torvalds 已提交
739 740 741 742
	case MBA_POINT_TO_POINT:	/* Point-to-Point */
		if (IS_QLA2100(ha))
			break;

743
		if (IS_QLA81XX(ha) || IS_QLA82XX(ha) || IS_QLA8031(ha)) {
744 745 746
			ql_dbg(ql_dbg_async, vha, 0x500d,
			    "DCBX Completed -- %04x %04x %04x.\n",
			    mb[1], mb[2], mb[3]);
747 748 749 750
			if (ha->notify_dcbx_comp)
				complete(&ha->dcbx_comp);

		} else
751 752
			ql_dbg(ql_dbg_async, vha, 0x500e,
			    "Asynchronous P2P MODE received.\n");
L
Linus Torvalds 已提交
753 754 755 756 757

		/*
		 * Until there's a transition from loop down to loop up, treat
		 * this as loop down only.
		 */
758 759 760 761
		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
			atomic_set(&vha->loop_state, LOOP_DOWN);
			if (!atomic_read(&vha->loop_down_timer))
				atomic_set(&vha->loop_down_timer,
L
Linus Torvalds 已提交
762
				    LOOP_DOWN_TIME);
763
			qla2x00_mark_all_devices_lost(vha, 1);
L
Linus Torvalds 已提交
764 765
		}

766 767 768
		if (vha->vp_idx) {
			atomic_set(&vha->vp_state, VP_FAILED);
			fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
769 770
		}

771 772 773 774 775
		if (!(test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags)))
			set_bit(RESET_MARKER_NEEDED, &vha->dpc_flags);

		set_bit(REGISTER_FC4_NEEDED, &vha->dpc_flags);
		set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
776 777

		ha->flags.gpsc_supported = 1;
778
		vha->flags.management_server_logged_in = 0;
L
Linus Torvalds 已提交
779 780 781 782 783 784
		break;

	case MBA_CHG_IN_CONNECTION:	/* Change in connection mode */
		if (IS_QLA2100(ha))
			break;

785
		ql_dbg(ql_dbg_async, vha, 0x500f,
L
Linus Torvalds 已提交
786 787
		    "Configuration change detected: value=%x.\n", mb[1]);

788 789 790 791
		if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
			atomic_set(&vha->loop_state, LOOP_DOWN);
			if (!atomic_read(&vha->loop_down_timer))
				atomic_set(&vha->loop_down_timer,
L
Linus Torvalds 已提交
792
				    LOOP_DOWN_TIME);
793
			qla2x00_mark_all_devices_lost(vha, 1);
L
Linus Torvalds 已提交
794 795
		}

796 797 798
		if (vha->vp_idx) {
			atomic_set(&vha->vp_state, VP_FAILED);
			fc_vport_set_state(vha->fc_vport, FC_VPORT_FAILED);
799 800
		}

801 802
		set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
		set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags);
L
Linus Torvalds 已提交
803 804 805
		break;

	case MBA_PORT_UPDATE:		/* Port database update */
806 807 808 809 810 811 812 813 814 815 816 817 818 819 820
		/*
		 * Handle only global and vn-port update events
		 *
		 * Relevant inputs:
		 * mb[1] = N_Port handle of changed port
		 * OR 0xffff for global event
		 * mb[2] = New login state
		 * 7 = Port logged out
		 * mb[3] = LSB is vp_idx, 0xff = all vps
		 *
		 * Skip processing if:
		 *       Event is global, vp_idx is NOT all vps,
		 *           vp_idx does not match
		 *       Event is not global, vp_idx does not match
		 */
821 822 823 824
		if (IS_QLA2XXX_MIDTYPE(ha) &&
		    ((mb[1] == 0xffff && (mb[3] & 0xff) != 0xff) ||
			(mb[1] != 0xffff)) && vha->vp_idx != (mb[3] & 0xff))
			break;
825

826 827
		/* Global event -- port logout or port unavailable. */
		if (mb[1] == 0xffff && mb[2] == 0x7) {
828 829 830
			ql_dbg(ql_dbg_async, vha, 0x5010,
			    "Port unavailable %04x %04x %04x.\n",
			    mb[1], mb[2], mb[3]);
831 832
			ql_log(ql_log_warn, vha, 0x505e,
			    "Link is offline.\n");
833 834 835 836 837 838 839 840 841 842 843 844 845

			if (atomic_read(&vha->loop_state) != LOOP_DOWN) {
				atomic_set(&vha->loop_state, LOOP_DOWN);
				atomic_set(&vha->loop_down_timer,
				    LOOP_DOWN_TIME);
				vha->device_flags |= DFLG_NO_CABLE;
				qla2x00_mark_all_devices_lost(vha, 1);
			}

			if (vha->vp_idx) {
				atomic_set(&vha->vp_state, VP_FAILED);
				fc_vport_set_state(vha->fc_vport,
				    FC_VPORT_FAILED);
846
				qla2x00_mark_all_devices_lost(vha, 1);
847 848 849 850 851 852 853
			}

			vha->flags.management_server_logged_in = 0;
			ha->link_data_rate = PORT_SPEED_UNKNOWN;
			break;
		}

L
Linus Torvalds 已提交
854
		/*
855
		 * If PORT UPDATE is global (received LIP_OCCURRED/LIP_RESET
L
Linus Torvalds 已提交
856 857 858
		 * event etc. earlier indicating loop is down) then process
		 * it.  Otherwise ignore it and Wait for RSCN to come in.
		 */
859
		atomic_set(&vha->loop_down_timer, 0);
860
		if (mb[1] != 0xffff || (mb[2] != 0x6 && mb[2] != 0x4)) {
861 862 863
			ql_dbg(ql_dbg_async, vha, 0x5011,
			    "Asynchronous PORT UPDATE ignored %04x/%04x/%04x.\n",
			    mb[1], mb[2], mb[3]);
864 865

			qlt_async_event(mb[0], vha, mb);
L
Linus Torvalds 已提交
866 867 868
			break;
		}

869 870 871
		ql_dbg(ql_dbg_async, vha, 0x5012,
		    "Port database changed %04x %04x %04x.\n",
		    mb[1], mb[2], mb[3]);
872 873
		ql_log(ql_log_warn, vha, 0x505f,
		    "Link is operational (%s Gbps).\n",
874
		    qla2x00_get_link_speed_str(ha, ha->link_data_rate));
L
Linus Torvalds 已提交
875 876 877 878

		/*
		 * Mark all devices as missing so we will login again.
		 */
879
		atomic_set(&vha->loop_state, LOOP_UP);
L
Linus Torvalds 已提交
880

881
		qla2x00_mark_all_devices_lost(vha, 1);
L
Linus Torvalds 已提交
882

883 884 885
		if (vha->vp_idx == 0 && !qla_ini_mode_enabled(vha))
			set_bit(SCR_PENDING, &vha->dpc_flags);

886 887
		set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
		set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags);
888 889

		qlt_async_event(mb[0], vha, mb);
L
Linus Torvalds 已提交
890 891 892
		break;

	case MBA_RSCN_UPDATE:		/* State Change Registration */
893
		/* Check if the Vport has issued a SCR */
894
		if (vha->vp_idx && test_bit(VP_SCR_NEEDED, &vha->vp_flags))
895 896
			break;
		/* Only handle SCNs for our Vport index. */
897
		if (ha->flags.npiv_supported && vha->vp_idx != (mb[3] & 0xff))
898
			break;
899

900 901 902
		ql_dbg(ql_dbg_async, vha, 0x5013,
		    "RSCN database changed -- %04x %04x %04x.\n",
		    mb[1], mb[2], mb[3]);
L
Linus Torvalds 已提交
903

904
		rscn_entry = ((mb[1] & 0xff) << 16) | mb[2];
905 906
		host_pid = (vha->d_id.b.domain << 16) | (vha->d_id.b.area << 8)
				| vha->d_id.b.al_pa;
L
Linus Torvalds 已提交
907
		if (rscn_entry == host_pid) {
908 909 910
			ql_dbg(ql_dbg_async, vha, 0x5014,
			    "Ignoring RSCN update to local host "
			    "port ID (%06x).\n", host_pid);
L
Linus Torvalds 已提交
911 912 913
			break;
		}

914 915
		/* Ignore reserved bits from RSCN-payload. */
		rscn_entry = ((mb[1] & 0x3ff) << 16) | mb[2];
L
Linus Torvalds 已提交
916

917 918 919 920
		/* Skip RSCNs for virtual ports on the same physical port */
		if (qla2x00_is_a_vp_did(vha, rscn_entry))
			break;

921 922
		atomic_set(&vha->loop_down_timer, 0);
		vha->flags.management_server_logged_in = 0;
L
Linus Torvalds 已提交
923

924 925 926
		set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags);
		set_bit(RSCN_UPDATE, &vha->dpc_flags);
		qla2x00_post_aen_work(vha, FCH_EVT_RSCN, rscn_entry);
L
Linus Torvalds 已提交
927 928 929 930
		break;

	/* case MBA_RIO_RESPONSE: */
	case MBA_ZIO_RESPONSE:
931 932
		ql_dbg(ql_dbg_async, vha, 0x5015,
		    "[R|Z]IO update completion.\n");
L
Linus Torvalds 已提交
933

934
		if (IS_FWI2_CAPABLE(ha))
935
			qla24xx_process_response_queue(vha, rsp);
936
		else
937
			qla2x00_process_response_queue(rsp);
L
Linus Torvalds 已提交
938
		break;
939 940

	case MBA_DISCARD_RND_FRAME:
941 942 943
		ql_dbg(ql_dbg_async, vha, 0x5016,
		    "Discard RND Frame -- %04x %04x %04x.\n",
		    mb[1], mb[2], mb[3]);
944
		break;
945 946

	case MBA_TRACE_NOTIFICATION:
947 948
		ql_dbg(ql_dbg_async, vha, 0x5017,
		    "Trace Notification -- %04x %04x.\n", mb[1], mb[2]);
949
		break;
950 951

	case MBA_ISP84XX_ALERT:
952 953 954
		ql_dbg(ql_dbg_async, vha, 0x5018,
		    "ISP84XX Alert Notification -- %04x %04x %04x.\n",
		    mb[1], mb[2], mb[3]);
955 956 957 958

		spin_lock_irqsave(&ha->cs84xx->access_lock, flags);
		switch (mb[1]) {
		case A84_PANIC_RECOVERY:
959 960 961
			ql_log(ql_log_info, vha, 0x5019,
			    "Alert 84XX: panic recovery %04x %04x.\n",
			    mb[2], mb[3]);
962 963 964
			break;
		case A84_OP_LOGIN_COMPLETE:
			ha->cs84xx->op_fw_version = mb[3] << 16 | mb[2];
965 966 967
			ql_log(ql_log_info, vha, 0x501a,
			    "Alert 84XX: firmware version %x.\n",
			    ha->cs84xx->op_fw_version);
968 969 970
			break;
		case A84_DIAG_LOGIN_COMPLETE:
			ha->cs84xx->diag_fw_version = mb[3] << 16 | mb[2];
971 972 973
			ql_log(ql_log_info, vha, 0x501b,
			    "Alert 84XX: diagnostic firmware version %x.\n",
			    ha->cs84xx->diag_fw_version);
974 975 976 977
			break;
		case A84_GOLD_LOGIN_COMPLETE:
			ha->cs84xx->diag_fw_version = mb[3] << 16 | mb[2];
			ha->cs84xx->fw_update = 1;
978 979 980
			ql_log(ql_log_info, vha, 0x501c,
			    "Alert 84XX: gold firmware version %x.\n",
			    ha->cs84xx->gold_fw_version);
981 982
			break;
		default:
983 984
			ql_log(ql_log_warn, vha, 0x501d,
			    "Alert 84xx: Invalid Alert %04x %04x %04x.\n",
985 986 987 988
			    mb[1], mb[2], mb[3]);
		}
		spin_unlock_irqrestore(&ha->cs84xx->access_lock, flags);
		break;
989
	case MBA_DCBX_START:
990 991 992
		ql_dbg(ql_dbg_async, vha, 0x501e,
		    "DCBX Started -- %04x %04x %04x.\n",
		    mb[1], mb[2], mb[3]);
993 994
		break;
	case MBA_DCBX_PARAM_UPDATE:
995 996 997
		ql_dbg(ql_dbg_async, vha, 0x501f,
		    "DCBX Parameters Updated -- %04x %04x %04x.\n",
		    mb[1], mb[2], mb[3]);
998 999
		break;
	case MBA_FCF_CONF_ERR:
1000 1001 1002
		ql_dbg(ql_dbg_async, vha, 0x5020,
		    "FCF Configuration Error -- %04x %04x %04x.\n",
		    mb[1], mb[2], mb[3]);
1003 1004
		break;
	case MBA_IDC_NOTIFY:
1005 1006 1007 1008 1009
		if (IS_QLA8031(vha->hw)) {
			mb[4] = RD_REG_WORD(&reg24->mailbox4);
			if (((mb[2] & 0x7fff) == MBC_PORT_RESET ||
			    (mb[2] & 0x7fff) == MBC_SET_PORT_CONFIG) &&
			    (mb[4] & INTERNAL_LOOPBACK_MASK) != 0) {
1010
				set_bit(ISP_QUIESCE_NEEDED, &vha->dpc_flags);
1011 1012 1013 1014 1015 1016
				/*
				 * Extend loop down timer since port is active.
				 */
				if (atomic_read(&vha->loop_state) == LOOP_DOWN)
					atomic_set(&vha->loop_down_timer,
					    LOOP_DOWN_TIME);
1017 1018
				qla2xxx_wake_dpc(vha);
			}
1019
		}
1020
	case MBA_IDC_COMPLETE:
1021 1022 1023
		if (ha->notify_lb_portup_comp)
			complete(&ha->lb_portup_comp);
		/* Fallthru */
1024
	case MBA_IDC_TIME_EXT:
1025
		if (IS_QLA81XX(vha->hw) || IS_QLA8031(vha->hw))
1026 1027 1028 1029 1030 1031 1032 1033 1034
			qla81xx_idc_event(vha, mb[0], mb[1]);
		break;

	case MBA_IDC_AEN:
		mb[4] = RD_REG_WORD(&reg24->mailbox4);
		mb[5] = RD_REG_WORD(&reg24->mailbox5);
		mb[6] = RD_REG_WORD(&reg24->mailbox6);
		mb[7] = RD_REG_WORD(&reg24->mailbox7);
		qla83xx_handle_8200_aen(vha, mb);
1035
		break;
1036

1037 1038 1039 1040
	default:
		ql_dbg(ql_dbg_async, vha, 0x5057,
		    "Unknown AEN:%04x %04x %04x %04x\n",
		    mb[0], mb[1], mb[2], mb[3]);
L
Linus Torvalds 已提交
1041
	}
1042

1043 1044
	qlt_async_event(mb[0], vha, mb);

1045
	if (!vha->vp_idx && ha->num_vhosts)
1046
		qla2x00_alert_all_vps(rsp, mb);
L
Linus Torvalds 已提交
1047 1048 1049 1050 1051 1052 1053
}

/**
 * qla2x00_process_completed_request() - Process a Fast Post response.
 * @ha: SCSI driver HA context
 * @index: SRB index
 */
1054
void
1055
qla2x00_process_completed_request(struct scsi_qla_host *vha,
1056
				  struct req_que *req, uint32_t index)
L
Linus Torvalds 已提交
1057 1058
{
	srb_t *sp;
1059
	struct qla_hw_data *ha = vha->hw;
L
Linus Torvalds 已提交
1060 1061

	/* Validate handle. */
1062
	if (index >= req->num_outstanding_cmds) {
1063 1064
		ql_log(ql_log_warn, vha, 0x3014,
		    "Invalid SCSI command index (%x).\n", index);
L
Linus Torvalds 已提交
1065

1066 1067 1068 1069
		if (IS_QLA82XX(ha))
			set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
		else
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
1070 1071 1072
		return;
	}

1073
	sp = req->outstanding_cmds[index];
L
Linus Torvalds 已提交
1074 1075
	if (sp) {
		/* Free outstanding command slot. */
1076
		req->outstanding_cmds[index] = NULL;
L
Linus Torvalds 已提交
1077 1078

		/* Save ISP completion status */
1079
		sp->done(ha, sp, DID_OK << 16);
L
Linus Torvalds 已提交
1080
	} else {
1081
		ql_log(ql_log_warn, vha, 0x3016, "Invalid SCSI SRB.\n");
L
Linus Torvalds 已提交
1082

1083 1084 1085 1086
		if (IS_QLA82XX(ha))
			set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
		else
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
L
Linus Torvalds 已提交
1087 1088 1089
	}
}

1090
srb_t *
1091 1092 1093 1094 1095 1096 1097 1098 1099
qla2x00_get_sp_from_handle(scsi_qla_host_t *vha, const char *func,
    struct req_que *req, void *iocb)
{
	struct qla_hw_data *ha = vha->hw;
	sts_entry_t *pkt = iocb;
	srb_t *sp = NULL;
	uint16_t index;

	index = LSW(pkt->handle);
1100
	if (index >= req->num_outstanding_cmds) {
1101 1102
		ql_log(ql_log_warn, vha, 0x5031,
		    "Invalid command index (%x).\n", index);
1103 1104 1105 1106
		if (IS_QLA82XX(ha))
			set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
		else
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1107 1108 1109 1110
		goto done;
	}
	sp = req->outstanding_cmds[index];
	if (!sp) {
1111 1112
		ql_log(ql_log_warn, vha, 0x5032,
		    "Invalid completion handle (%x) -- timed-out.\n", index);
1113 1114 1115
		return sp;
	}
	if (sp->handle != index) {
1116 1117
		ql_log(ql_log_warn, vha, 0x5033,
		    "SRB handle (%x) mismatch %x.\n", sp->handle, index);
1118 1119
		return NULL;
	}
1120

1121
	req->outstanding_cmds[index] = NULL;
1122

1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134
done:
	return sp;
}

static void
qla2x00_mbx_iocb_entry(scsi_qla_host_t *vha, struct req_que *req,
    struct mbx_entry *mbx)
{
	const char func[] = "MBX-IOCB";
	const char *type;
	fc_port_t *fcport;
	srb_t *sp;
1135
	struct srb_iocb *lio;
1136
	uint16_t *data;
1137
	uint16_t status;
1138 1139 1140 1141 1142

	sp = qla2x00_get_sp_from_handle(vha, func, req, mbx);
	if (!sp)
		return;

1143 1144
	lio = &sp->u.iocb_cmd;
	type = sp->name;
1145
	fcport = sp->fcport;
1146
	data = lio->u.logio.data;
1147

1148
	data[0] = MBS_COMMAND_ERROR;
1149
	data[1] = lio->u.logio.flags & SRB_LOGIN_RETRIED ?
1150
	    QLA_LOGIO_LOGIN_RETRIED : 0;
1151
	if (mbx->entry_status) {
1152
		ql_dbg(ql_dbg_async, vha, 0x5043,
1153
		    "Async-%s error entry - hdl=%x portid=%02x%02x%02x "
1154
		    "entry-status=%x status=%x state-flag=%x "
1155 1156
		    "status-flags=%x.\n", type, sp->handle,
		    fcport->d_id.b.domain, fcport->d_id.b.area,
1157 1158
		    fcport->d_id.b.al_pa, mbx->entry_status,
		    le16_to_cpu(mbx->status), le16_to_cpu(mbx->state_flags),
1159
		    le16_to_cpu(mbx->status_flags));
1160

1161
		ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x5029,
1162
		    (uint8_t *)mbx, sizeof(*mbx));
1163

1164
		goto logio_done;
1165 1166
	}

1167
	status = le16_to_cpu(mbx->status);
1168
	if (status == 0x30 && sp->type == SRB_LOGIN_CMD &&
1169 1170 1171
	    le16_to_cpu(mbx->mb0) == MBS_COMMAND_COMPLETE)
		status = 0;
	if (!status && le16_to_cpu(mbx->mb0) == MBS_COMMAND_COMPLETE) {
1172
		ql_dbg(ql_dbg_async, vha, 0x5045,
1173 1174 1175 1176
		    "Async-%s complete - hdl=%x portid=%02x%02x%02x mbx1=%x.\n",
		    type, sp->handle, fcport->d_id.b.domain,
		    fcport->d_id.b.area, fcport->d_id.b.al_pa,
		    le16_to_cpu(mbx->mb1));
1177 1178

		data[0] = MBS_COMMAND_COMPLETE;
1179
		if (sp->type == SRB_LOGIN_CMD) {
1180 1181 1182
			fcport->port_type = FCT_TARGET;
			if (le16_to_cpu(mbx->mb1) & BIT_0)
				fcport->port_type = FCT_INITIATOR;
1183
			else if (le16_to_cpu(mbx->mb1) & BIT_1)
1184
				fcport->flags |= FCF_FCP2_DEVICE;
1185
		}
1186
		goto logio_done;
1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200
	}

	data[0] = le16_to_cpu(mbx->mb0);
	switch (data[0]) {
	case MBS_PORT_ID_USED:
		data[1] = le16_to_cpu(mbx->mb1);
		break;
	case MBS_LOOP_ID_USED:
		break;
	default:
		data[0] = MBS_COMMAND_ERROR;
		break;
	}

1201
	ql_log(ql_log_warn, vha, 0x5046,
1202 1203 1204 1205
	    "Async-%s failed - hdl=%x portid=%02x%02x%02x status=%x "
	    "mb0=%x mb1=%x mb2=%x mb6=%x mb7=%x.\n", type, sp->handle,
	    fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa,
	    status, le16_to_cpu(mbx->mb0), le16_to_cpu(mbx->mb1),
1206
	    le16_to_cpu(mbx->mb2), le16_to_cpu(mbx->mb6),
1207
	    le16_to_cpu(mbx->mb7));
1208

1209
logio_done:
1210
	sp->done(vha, sp, 0);
1211 1212
}

1213 1214 1215 1216 1217 1218 1219 1220 1221
static void
qla2x00_ct_entry(scsi_qla_host_t *vha, struct req_que *req,
    sts_entry_t *pkt, int iocb_type)
{
	const char func[] = "CT_IOCB";
	const char *type;
	srb_t *sp;
	struct fc_bsg_job *bsg_job;
	uint16_t comp_status;
1222
	int res;
1223 1224 1225 1226 1227

	sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
	if (!sp)
		return;

1228
	bsg_job = sp->u.bsg_job;
1229

1230
	type = "ct pass-through";
1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241

	comp_status = le16_to_cpu(pkt->comp_status);

	/* return FC_CTELS_STATUS_OK and leave the decoding of the ELS/CT
	 * fc payload  to the caller
	 */
	bsg_job->reply->reply_data.ctels_reply.status = FC_CTELS_STATUS_OK;
	bsg_job->reply_len = sizeof(struct fc_bsg_reply);

	if (comp_status != CS_COMPLETE) {
		if (comp_status == CS_DATA_UNDERRUN) {
1242
			res = DID_OK << 16;
1243 1244 1245
			bsg_job->reply->reply_payload_rcv_len =
			    le16_to_cpu(((sts_entry_t *)pkt)->rsp_info_len);

1246 1247
			ql_log(ql_log_warn, vha, 0x5048,
			    "CT pass-through-%s error "
1248
			    "comp_status-status=0x%x total_byte = 0x%x.\n",
1249 1250
			    type, comp_status,
			    bsg_job->reply->reply_payload_rcv_len);
1251
		} else {
1252 1253 1254
			ql_log(ql_log_warn, vha, 0x5049,
			    "CT pass-through-%s error "
			    "comp_status-status=0x%x.\n", type, comp_status);
1255
			res = DID_ERROR << 16;
1256 1257
			bsg_job->reply->reply_payload_rcv_len = 0;
		}
1258
		ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x5035,
1259
		    (uint8_t *)pkt, sizeof(*pkt));
1260
	} else {
1261
		res = DID_OK << 16;
1262 1263 1264 1265 1266
		bsg_job->reply->reply_payload_rcv_len =
		    bsg_job->reply_payload.payload_len;
		bsg_job->reply_len = 0;
	}

1267
	sp->done(vha, sp, res);
1268 1269
}

1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280
static void
qla24xx_els_ct_entry(scsi_qla_host_t *vha, struct req_que *req,
    struct sts_entry_24xx *pkt, int iocb_type)
{
	const char func[] = "ELS_CT_IOCB";
	const char *type;
	srb_t *sp;
	struct fc_bsg_job *bsg_job;
	uint16_t comp_status;
	uint32_t fw_status[3];
	uint8_t* fw_sts_ptr;
1281
	int res;
1282 1283 1284 1285

	sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
	if (!sp)
		return;
1286
	bsg_job = sp->u.bsg_job;
1287 1288

	type = NULL;
1289
	switch (sp->type) {
1290 1291 1292 1293 1294 1295 1296 1297
	case SRB_ELS_CMD_RPT:
	case SRB_ELS_CMD_HST:
		type = "els";
		break;
	case SRB_CT_CMD:
		type = "ct pass-through";
		break;
	default:
1298
		ql_dbg(ql_dbg_user, vha, 0x503e,
1299
		    "Unrecognized SRB: (%p) type=%d.\n", sp, sp->type);
1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314
		return;
	}

	comp_status = fw_status[0] = le16_to_cpu(pkt->comp_status);
	fw_status[1] = le16_to_cpu(((struct els_sts_entry_24xx*)pkt)->error_subcode_1);
	fw_status[2] = le16_to_cpu(((struct els_sts_entry_24xx*)pkt)->error_subcode_2);

	/* return FC_CTELS_STATUS_OK and leave the decoding of the ELS/CT
	 * fc payload  to the caller
	 */
	bsg_job->reply->reply_data.ctels_reply.status = FC_CTELS_STATUS_OK;
	bsg_job->reply_len = sizeof(struct fc_bsg_reply) + sizeof(fw_status);

	if (comp_status != CS_COMPLETE) {
		if (comp_status == CS_DATA_UNDERRUN) {
1315
			res = DID_OK << 16;
1316
			bsg_job->reply->reply_payload_rcv_len =
1317
			    le16_to_cpu(((struct els_sts_entry_24xx *)pkt)->total_byte_count);
1318

1319
			ql_dbg(ql_dbg_user, vha, 0x503f,
1320
			    "ELS-CT pass-through-%s error hdl=%x comp_status-status=0x%x "
1321
			    "error subcode 1=0x%x error subcode 2=0x%x total_byte = 0x%x.\n",
1322
			    type, sp->handle, comp_status, fw_status[1], fw_status[2],
1323 1324
			    le16_to_cpu(((struct els_sts_entry_24xx *)
				pkt)->total_byte_count));
1325 1326 1327 1328
			fw_sts_ptr = ((uint8_t*)bsg_job->req->sense) + sizeof(struct fc_bsg_reply);
			memcpy( fw_sts_ptr, fw_status, sizeof(fw_status));
		}
		else {
1329
			ql_dbg(ql_dbg_user, vha, 0x5040,
1330
			    "ELS-CT pass-through-%s error hdl=%x comp_status-status=0x%x "
1331
			    "error subcode 1=0x%x error subcode 2=0x%x.\n",
1332
			    type, sp->handle, comp_status,
1333 1334 1335 1336
			    le16_to_cpu(((struct els_sts_entry_24xx *)
				pkt)->error_subcode_1),
			    le16_to_cpu(((struct els_sts_entry_24xx *)
				    pkt)->error_subcode_2));
1337
			res = DID_ERROR << 16;
1338 1339 1340 1341
			bsg_job->reply->reply_payload_rcv_len = 0;
			fw_sts_ptr = ((uint8_t*)bsg_job->req->sense) + sizeof(struct fc_bsg_reply);
			memcpy( fw_sts_ptr, fw_status, sizeof(fw_status));
		}
1342
		ql_dump_buffer(ql_dbg_user + ql_dbg_buffer, vha, 0x5056,
1343
				(uint8_t *)pkt, sizeof(*pkt));
1344 1345
	}
	else {
1346
		res =  DID_OK << 16;
1347 1348 1349 1350
		bsg_job->reply->reply_payload_rcv_len = bsg_job->reply_payload.payload_len;
		bsg_job->reply_len = 0;
	}

1351
	sp->done(vha, sp, res);
1352 1353
}

1354 1355 1356 1357 1358 1359 1360 1361
static void
qla24xx_logio_entry(scsi_qla_host_t *vha, struct req_que *req,
    struct logio_entry_24xx *logio)
{
	const char func[] = "LOGIO-IOCB";
	const char *type;
	fc_port_t *fcport;
	srb_t *sp;
1362
	struct srb_iocb *lio;
1363
	uint16_t *data;
1364 1365 1366 1367 1368 1369
	uint32_t iop[2];

	sp = qla2x00_get_sp_from_handle(vha, func, req, logio);
	if (!sp)
		return;

1370 1371
	lio = &sp->u.iocb_cmd;
	type = sp->name;
1372
	fcport = sp->fcport;
1373
	data = lio->u.logio.data;
1374

1375
	data[0] = MBS_COMMAND_ERROR;
1376
	data[1] = lio->u.logio.flags & SRB_LOGIN_RETRIED ?
1377
		QLA_LOGIO_LOGIN_RETRIED : 0;
1378
	if (logio->entry_status) {
1379
		ql_log(ql_log_warn, fcport->vha, 0x5034,
1380
		    "Async-%s error entry - hdl=%x"
1381
		    "portid=%02x%02x%02x entry-status=%x.\n",
1382 1383 1384 1385
		    type, sp->handle, fcport->d_id.b.domain,
		    fcport->d_id.b.area, fcport->d_id.b.al_pa,
		    logio->entry_status);
		ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x504d,
1386
		    (uint8_t *)logio, sizeof(*logio));
1387

1388
		goto logio_done;
1389 1390 1391
	}

	if (le16_to_cpu(logio->comp_status) == CS_COMPLETE) {
1392
		ql_dbg(ql_dbg_async, fcport->vha, 0x5036,
1393 1394 1395
		    "Async-%s complete - hdl=%x portid=%02x%02x%02x "
		    "iop0=%x.\n", type, sp->handle, fcport->d_id.b.domain,
		    fcport->d_id.b.area, fcport->d_id.b.al_pa,
1396
		    le32_to_cpu(logio->io_parameter[0]));
1397 1398

		data[0] = MBS_COMMAND_COMPLETE;
1399
		if (sp->type != SRB_LOGIN_CMD)
1400
			goto logio_done;
1401 1402 1403 1404 1405

		iop[0] = le32_to_cpu(logio->io_parameter[0]);
		if (iop[0] & BIT_4) {
			fcport->port_type = FCT_TARGET;
			if (iop[0] & BIT_8)
1406
				fcport->flags |= FCF_FCP2_DEVICE;
1407
		} else if (iop[0] & BIT_5)
1408
			fcport->port_type = FCT_INITIATOR;
1409

1410 1411 1412
		if (iop[0] & BIT_7)
			fcport->flags |= FCF_CONF_COMP_SUPPORTED;

1413 1414 1415 1416 1417
		if (logio->io_parameter[7] || logio->io_parameter[8])
			fcport->supported_classes |= FC_COS_CLASS2;
		if (logio->io_parameter[9] || logio->io_parameter[10])
			fcport->supported_classes |= FC_COS_CLASS3;

1418
		goto logio_done;
1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435
	}

	iop[0] = le32_to_cpu(logio->io_parameter[0]);
	iop[1] = le32_to_cpu(logio->io_parameter[1]);
	switch (iop[0]) {
	case LSC_SCODE_PORTID_USED:
		data[0] = MBS_PORT_ID_USED;
		data[1] = LSW(iop[1]);
		break;
	case LSC_SCODE_NPORT_USED:
		data[0] = MBS_LOOP_ID_USED;
		break;
	default:
		data[0] = MBS_COMMAND_ERROR;
		break;
	}

1436
	ql_dbg(ql_dbg_async, fcport->vha, 0x5037,
1437 1438
	    "Async-%s failed - hdl=%x portid=%02x%02x%02x comp=%x "
	    "iop0=%x iop1=%x.\n", type, sp->handle, fcport->d_id.b.domain,
1439
	    fcport->d_id.b.area, fcport->d_id.b.al_pa,
1440 1441
	    le16_to_cpu(logio->comp_status),
	    le32_to_cpu(logio->io_parameter[0]),
1442
	    le32_to_cpu(logio->io_parameter[1]));
1443

1444
logio_done:
1445
	sp->done(vha, sp, 0);
1446 1447
}

1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463
static void
qla24xx_tm_iocb_entry(scsi_qla_host_t *vha, struct req_que *req,
    struct tsk_mgmt_entry *tsk)
{
	const char func[] = "TMF-IOCB";
	const char *type;
	fc_port_t *fcport;
	srb_t *sp;
	struct srb_iocb *iocb;
	struct sts_entry_24xx *sts = (struct sts_entry_24xx *)tsk;
	int error = 1;

	sp = qla2x00_get_sp_from_handle(vha, func, req, tsk);
	if (!sp)
		return;

1464 1465
	iocb = &sp->u.iocb_cmd;
	type = sp->name;
1466 1467 1468
	fcport = sp->fcport;

	if (sts->entry_status) {
1469
		ql_log(ql_log_warn, fcport->vha, 0x5038,
1470 1471
		    "Async-%s error - hdl=%x entry-status(%x).\n",
		    type, sp->handle, sts->entry_status);
1472
	} else if (sts->comp_status != __constant_cpu_to_le16(CS_COMPLETE)) {
1473
		ql_log(ql_log_warn, fcport->vha, 0x5039,
1474 1475
		    "Async-%s error - hdl=%x completion status(%x).\n",
		    type, sp->handle, sts->comp_status);
1476 1477
	} else if (!(le16_to_cpu(sts->scsi_status) &
	    SS_RESPONSE_INFO_LEN_VALID)) {
1478
		ql_log(ql_log_warn, fcport->vha, 0x503a,
1479 1480
		    "Async-%s error - hdl=%x no response info(%x).\n",
		    type, sp->handle, sts->scsi_status);
1481
	} else if (le32_to_cpu(sts->rsp_data_len) < 4) {
1482
		ql_log(ql_log_warn, fcport->vha, 0x503b,
1483 1484
		    "Async-%s error - hdl=%x not enough response(%d).\n",
		    type, sp->handle, sts->rsp_data_len);
1485
	} else if (sts->data[3]) {
1486
		ql_log(ql_log_warn, fcport->vha, 0x503c,
1487 1488
		    "Async-%s error - hdl=%x response(%x).\n",
		    type, sp->handle, sts->data[3]);
1489 1490 1491 1492 1493 1494
	} else {
		error = 0;
	}

	if (error) {
		iocb->u.tmf.data = error;
1495 1496
		ql_dump_buffer(ql_dbg_async + ql_dbg_buffer, vha, 0x5055,
		    (uint8_t *)sts, sizeof(*sts));
1497 1498
	}

1499
	sp->done(vha, sp, 0);
1500 1501
}

L
Linus Torvalds 已提交
1502 1503 1504 1505 1506
/**
 * qla2x00_process_response_queue() - Process response queue entries.
 * @ha: SCSI driver HA context
 */
void
1507
qla2x00_process_response_queue(struct rsp_que *rsp)
L
Linus Torvalds 已提交
1508
{
1509 1510
	struct scsi_qla_host *vha;
	struct qla_hw_data *ha = rsp->hw;
1511
	struct device_reg_2xxx __iomem *reg = &ha->iobase->isp;
L
Linus Torvalds 已提交
1512 1513 1514
	sts_entry_t	*pkt;
	uint16_t        handle_cnt;
	uint16_t        cnt;
1515

1516
	vha = pci_get_drvdata(ha->pdev);
L
Linus Torvalds 已提交
1517

1518
	if (!vha->flags.online)
L
Linus Torvalds 已提交
1519 1520
		return;

1521 1522
	while (rsp->ring_ptr->signature != RESPONSE_PROCESSED) {
		pkt = (sts_entry_t *)rsp->ring_ptr;
L
Linus Torvalds 已提交
1523

1524 1525 1526 1527
		rsp->ring_index++;
		if (rsp->ring_index == rsp->length) {
			rsp->ring_index = 0;
			rsp->ring_ptr = rsp->ring;
L
Linus Torvalds 已提交
1528
		} else {
1529
			rsp->ring_ptr++;
L
Linus Torvalds 已提交
1530 1531 1532
		}

		if (pkt->entry_status != 0) {
1533
			qla2x00_error_entry(vha, rsp, pkt);
L
Linus Torvalds 已提交
1534 1535 1536 1537 1538 1539 1540
			((response_t *)pkt)->signature = RESPONSE_PROCESSED;
			wmb();
			continue;
		}

		switch (pkt->entry_type) {
		case STATUS_TYPE:
1541
			qla2x00_status_entry(vha, rsp, pkt);
L
Linus Torvalds 已提交
1542 1543 1544 1545
			break;
		case STATUS_TYPE_21:
			handle_cnt = ((sts21_entry_t *)pkt)->handle_count;
			for (cnt = 0; cnt < handle_cnt; cnt++) {
1546
				qla2x00_process_completed_request(vha, rsp->req,
L
Linus Torvalds 已提交
1547 1548 1549 1550 1551 1552
				    ((sts21_entry_t *)pkt)->handle[cnt]);
			}
			break;
		case STATUS_TYPE_22:
			handle_cnt = ((sts22_entry_t *)pkt)->handle_count;
			for (cnt = 0; cnt < handle_cnt; cnt++) {
1553
				qla2x00_process_completed_request(vha, rsp->req,
L
Linus Torvalds 已提交
1554 1555 1556 1557
				    ((sts22_entry_t *)pkt)->handle[cnt]);
			}
			break;
		case STATUS_CONT_TYPE:
1558
			qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt);
L
Linus Torvalds 已提交
1559
			break;
1560 1561 1562
		case MBX_IOCB_TYPE:
			qla2x00_mbx_iocb_entry(vha, rsp->req,
			    (struct mbx_entry *)pkt);
1563
			break;
1564 1565 1566
		case CT_IOCB_TYPE:
			qla2x00_ct_entry(vha, rsp->req, pkt, CT_IOCB_TYPE);
			break;
L
Linus Torvalds 已提交
1567 1568
		default:
			/* Type Not Supported. */
1569 1570
			ql_log(ql_log_warn, vha, 0x504a,
			    "Received unknown response pkt type %x "
L
Linus Torvalds 已提交
1571
			    "entry status=%x.\n",
1572
			    pkt->entry_type, pkt->entry_status);
L
Linus Torvalds 已提交
1573 1574 1575 1576 1577 1578 1579
			break;
		}
		((response_t *)pkt)->signature = RESPONSE_PROCESSED;
		wmb();
	}

	/* Adjust ring index */
1580
	WRT_REG_WORD(ISP_RSP_Q_OUT(ha, reg), rsp->ring_index);
L
Linus Torvalds 已提交
1581 1582
}

1583
static inline void
1584
qla2x00_handle_sense(srb_t *sp, uint8_t *sense_data, uint32_t par_sense_len,
1585
		     uint32_t sense_len, struct rsp_que *rsp, int res)
1586
{
1587
	struct scsi_qla_host *vha = sp->fcport->vha;
1588 1589
	struct scsi_cmnd *cp = GET_CMD_SP(sp);
	uint32_t track_sense_len;
1590 1591 1592 1593

	if (sense_len >= SCSI_SENSE_BUFFERSIZE)
		sense_len = SCSI_SENSE_BUFFERSIZE;

1594 1595 1596 1597 1598
	SET_CMD_SENSE_LEN(sp, sense_len);
	SET_CMD_SENSE_PTR(sp, cp->sense_buffer);
	track_sense_len = sense_len;

	if (sense_len > par_sense_len)
1599
		sense_len = par_sense_len;
1600 1601 1602

	memcpy(cp->sense_buffer, sense_data, sense_len);

1603 1604 1605 1606 1607
	SET_CMD_SENSE_PTR(sp, cp->sense_buffer + sense_len);
	track_sense_len -= sense_len;
	SET_CMD_SENSE_LEN(sp, track_sense_len);

	if (track_sense_len != 0) {
1608
		rsp->status_srb = sp;
1609 1610
		cp->result = res;
	}
1611

1612 1613 1614 1615 1616
	if (sense_len) {
		ql_dbg(ql_dbg_io + ql_dbg_buffer, vha, 0x301c,
		    "Check condition Sense data, nexus%ld:%d:%d cmd=%p.\n",
		    sp->fcport->vha->host_no, cp->device->id, cp->device->lun,
		    cp);
1617 1618
		ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x302b,
		    cp->sense_buffer, sense_len);
1619
	}
1620 1621
}

1622 1623
struct scsi_dif_tuple {
	__be16 guard;       /* Checksum */
1624
	__be16 app_tag;         /* APPL identifier */
1625 1626 1627 1628 1629 1630 1631 1632 1633
	__be32 ref_tag;         /* Target LBA or indirect LBA */
};

/*
 * Checks the guard or meta-data for the type of error
 * detected by the HBA. In case of errors, we set the
 * ASC/ASCQ fields in the sense buffer with ILLEGAL_REQUEST
 * to indicate to the kernel that the HBA detected error.
 */
1634
static inline int
1635 1636
qla2x00_handle_dif_error(srb_t *sp, struct sts_entry_24xx *sts24)
{
1637
	struct scsi_qla_host *vha = sp->fcport->vha;
1638
	struct scsi_cmnd *cmd = GET_CMD_SP(sp);
1639 1640
	uint8_t		*ap = &sts24->data[12];
	uint8_t		*ep = &sts24->data[20];
1641 1642 1643 1644
	uint32_t	e_ref_tag, a_ref_tag;
	uint16_t	e_app_tag, a_app_tag;
	uint16_t	e_guard, a_guard;

1645 1646 1647 1648 1649 1650 1651 1652 1653 1654
	/*
	 * swab32 of the "data" field in the beginning of qla2x00_status_entry()
	 * would make guard field appear at offset 2
	 */
	a_guard   = le16_to_cpu(*(uint16_t *)(ap + 2));
	a_app_tag = le16_to_cpu(*(uint16_t *)(ap + 0));
	a_ref_tag = le32_to_cpu(*(uint32_t *)(ap + 4));
	e_guard   = le16_to_cpu(*(uint16_t *)(ep + 2));
	e_app_tag = le16_to_cpu(*(uint16_t *)(ep + 0));
	e_ref_tag = le32_to_cpu(*(uint32_t *)(ep + 4));
1655

1656 1657
	ql_dbg(ql_dbg_io, vha, 0x3023,
	    "iocb(s) %p Returned STATUS.\n", sts24);
1658

1659 1660
	ql_dbg(ql_dbg_io, vha, 0x3024,
	    "DIF ERROR in cmd 0x%x lba 0x%llx act ref"
1661
	    " tag=0x%x, exp ref_tag=0x%x, act app tag=0x%x, exp app"
1662
	    " tag=0x%x, act guard=0x%x, exp guard=0x%x.\n",
1663
	    cmd->cmnd[0], (u64)scsi_get_lba(cmd), a_ref_tag, e_ref_tag,
1664
	    a_app_tag, e_app_tag, a_guard, e_guard);
1665

1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705
	/*
	 * Ignore sector if:
	 * For type     3: ref & app tag is all 'f's
	 * For type 0,1,2: app tag is all 'f's
	 */
	if ((a_app_tag == 0xffff) &&
	    ((scsi_get_prot_type(cmd) != SCSI_PROT_DIF_TYPE3) ||
	     (a_ref_tag == 0xffffffff))) {
		uint32_t blocks_done, resid;
		sector_t lba_s = scsi_get_lba(cmd);

		/* 2TB boundary case covered automatically with this */
		blocks_done = e_ref_tag - (uint32_t)lba_s + 1;

		resid = scsi_bufflen(cmd) - (blocks_done *
		    cmd->device->sector_size);

		scsi_set_resid(cmd, resid);
		cmd->result = DID_OK << 16;

		/* Update protection tag */
		if (scsi_prot_sg_count(cmd)) {
			uint32_t i, j = 0, k = 0, num_ent;
			struct scatterlist *sg;
			struct sd_dif_tuple *spt;

			/* Patch the corresponding protection tags */
			scsi_for_each_prot_sg(cmd, sg,
			    scsi_prot_sg_count(cmd), i) {
				num_ent = sg_dma_len(sg) / 8;
				if (k + num_ent < blocks_done) {
					k += num_ent;
					continue;
				}
				j = blocks_done - k - 1;
				k = blocks_done;
				break;
			}

			if (k != blocks_done) {
1706
				ql_log(ql_log_warn, vha, 0x302f,
1707 1708
				    "unexpected tag values tag:lba=%x:%llx)\n",
				    e_ref_tag, (unsigned long long)lba_s);
1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722
				return 1;
			}

			spt = page_address(sg_page(sg)) + sg->offset;
			spt += j;

			spt->app_tag = 0xffff;
			if (scsi_get_prot_type(cmd) == SCSI_PROT_DIF_TYPE3)
				spt->ref_tag = 0xffffffff;
		}

		return 0;
	}

1723 1724 1725 1726 1727 1728 1729
	/* check guard */
	if (e_guard != a_guard) {
		scsi_build_sense_buffer(1, cmd->sense_buffer, ILLEGAL_REQUEST,
		    0x10, 0x1);
		set_driver_byte(cmd, DRIVER_SENSE);
		set_host_byte(cmd, DID_ABORT);
		cmd->result |= SAM_STAT_CHECK_CONDITION << 1;
1730
		return 1;
1731 1732
	}

1733 1734
	/* check ref tag */
	if (e_ref_tag != a_ref_tag) {
1735
		scsi_build_sense_buffer(1, cmd->sense_buffer, ILLEGAL_REQUEST,
1736
		    0x10, 0x3);
1737 1738 1739
		set_driver_byte(cmd, DRIVER_SENSE);
		set_host_byte(cmd, DID_ABORT);
		cmd->result |= SAM_STAT_CHECK_CONDITION << 1;
1740
		return 1;
1741 1742
	}

1743 1744
	/* check appl tag */
	if (e_app_tag != a_app_tag) {
1745
		scsi_build_sense_buffer(1, cmd->sense_buffer, ILLEGAL_REQUEST,
1746
		    0x10, 0x2);
1747 1748 1749
		set_driver_byte(cmd, DRIVER_SENSE);
		set_host_byte(cmd, DID_ABORT);
		cmd->result |= SAM_STAT_CHECK_CONDITION << 1;
1750
		return 1;
1751
	}
1752

1753
	return 1;
1754 1755
}

1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772
static void
qla25xx_process_bidir_status_iocb(scsi_qla_host_t *vha, void *pkt,
				  struct req_que *req, uint32_t index)
{
	struct qla_hw_data *ha = vha->hw;
	srb_t *sp;
	uint16_t	comp_status;
	uint16_t	scsi_status;
	uint16_t thread_id;
	uint32_t rval = EXT_STATUS_OK;
	struct fc_bsg_job *bsg_job = NULL;
	sts_entry_t *sts;
	struct sts_entry_24xx *sts24;
	sts = (sts_entry_t *) pkt;
	sts24 = (struct sts_entry_24xx *) pkt;

	/* Validate handle. */
1773
	if (index >= req->num_outstanding_cmds) {
1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898
		ql_log(ql_log_warn, vha, 0x70af,
		    "Invalid SCSI completion handle 0x%x.\n", index);
		set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
		return;
	}

	sp = req->outstanding_cmds[index];
	if (sp) {
		/* Free outstanding command slot. */
		req->outstanding_cmds[index] = NULL;
		bsg_job = sp->u.bsg_job;
	} else {
		ql_log(ql_log_warn, vha, 0x70b0,
		    "Req:%d: Invalid ISP SCSI completion handle(0x%x)\n",
		    req->id, index);

		set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
		return;
	}

	if (IS_FWI2_CAPABLE(ha)) {
		comp_status = le16_to_cpu(sts24->comp_status);
		scsi_status = le16_to_cpu(sts24->scsi_status) & SS_MASK;
	} else {
		comp_status = le16_to_cpu(sts->comp_status);
		scsi_status = le16_to_cpu(sts->scsi_status) & SS_MASK;
	}

	thread_id = bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
	switch (comp_status) {
	case CS_COMPLETE:
		if (scsi_status == 0) {
			bsg_job->reply->reply_payload_rcv_len =
					bsg_job->reply_payload.payload_len;
			rval = EXT_STATUS_OK;
		}
		goto done;

	case CS_DATA_OVERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b1,
		    "Command completed with date overrun thread_id=%d\n",
		    thread_id);
		rval = EXT_STATUS_DATA_OVERRUN;
		break;

	case CS_DATA_UNDERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b2,
		    "Command completed with date underrun thread_id=%d\n",
		    thread_id);
		rval = EXT_STATUS_DATA_UNDERRUN;
		break;
	case CS_BIDIR_RD_OVERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b3,
		    "Command completed with read data overrun thread_id=%d\n",
		    thread_id);
		rval = EXT_STATUS_DATA_OVERRUN;
		break;

	case CS_BIDIR_RD_WR_OVERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b4,
		    "Command completed with read and write data overrun "
		    "thread_id=%d\n", thread_id);
		rval = EXT_STATUS_DATA_OVERRUN;
		break;

	case CS_BIDIR_RD_OVERRUN_WR_UNDERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b5,
		    "Command completed with read data over and write data "
		    "underrun thread_id=%d\n", thread_id);
		rval = EXT_STATUS_DATA_OVERRUN;
		break;

	case CS_BIDIR_RD_UNDERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b6,
		    "Command completed with read data data underrun "
		    "thread_id=%d\n", thread_id);
		rval = EXT_STATUS_DATA_UNDERRUN;
		break;

	case CS_BIDIR_RD_UNDERRUN_WR_OVERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b7,
		    "Command completed with read data under and write data "
		    "overrun thread_id=%d\n", thread_id);
		rval = EXT_STATUS_DATA_UNDERRUN;
		break;

	case CS_BIDIR_RD_WR_UNDERRUN:
		ql_dbg(ql_dbg_user, vha, 0x70b8,
		    "Command completed with read and write data underrun "
		    "thread_id=%d\n", thread_id);
		rval = EXT_STATUS_DATA_UNDERRUN;
		break;

	case CS_BIDIR_DMA:
		ql_dbg(ql_dbg_user, vha, 0x70b9,
		    "Command completed with data DMA error thread_id=%d\n",
		    thread_id);
		rval = EXT_STATUS_DMA_ERR;
		break;

	case CS_TIMEOUT:
		ql_dbg(ql_dbg_user, vha, 0x70ba,
		    "Command completed with timeout thread_id=%d\n",
		    thread_id);
		rval = EXT_STATUS_TIMEOUT;
		break;
	default:
		ql_dbg(ql_dbg_user, vha, 0x70bb,
		    "Command completed with completion status=0x%x "
		    "thread_id=%d\n", comp_status, thread_id);
		rval = EXT_STATUS_ERR;
		break;
	}
		bsg_job->reply->reply_payload_rcv_len = 0;

done:
	/* Return the vendor specific reply to API */
	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] = rval;
	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
	/* Always return DID_OK, bsg will send the vendor specific response
	 * in this case only */
	sp->done(vha, sp, (DID_OK << 6));

}

L
Linus Torvalds 已提交
1899 1900 1901 1902 1903 1904
/**
 * qla2x00_status_entry() - Process a Status IOCB entry.
 * @ha: SCSI driver HA context
 * @pkt: Entry pointer
 */
static void
1905
qla2x00_status_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, void *pkt)
L
Linus Torvalds 已提交
1906 1907 1908 1909
{
	srb_t		*sp;
	fc_port_t	*fcport;
	struct scsi_cmnd *cp;
1910 1911
	sts_entry_t *sts;
	struct sts_entry_24xx *sts24;
L
Linus Torvalds 已提交
1912 1913
	uint16_t	comp_status;
	uint16_t	scsi_status;
1914
	uint16_t	ox_id;
L
Linus Torvalds 已提交
1915 1916
	uint8_t		lscsi_status;
	int32_t		resid;
1917 1918
	uint32_t sense_len, par_sense_len, rsp_info_len, resid_len,
	    fw_resid_len;
1919
	uint8_t		*rsp_info, *sense_data;
1920
	struct qla_hw_data *ha = vha->hw;
1921 1922 1923
	uint32_t handle;
	uint16_t que;
	struct req_que *req;
1924
	int logit = 1;
1925
	int res = 0;
1926
	uint16_t state_flags = 0;
1927 1928 1929

	sts = (sts_entry_t *) pkt;
	sts24 = (struct sts_entry_24xx *) pkt;
1930
	if (IS_FWI2_CAPABLE(ha)) {
1931 1932
		comp_status = le16_to_cpu(sts24->comp_status);
		scsi_status = le16_to_cpu(sts24->scsi_status) & SS_MASK;
1933
		state_flags = le16_to_cpu(sts24->state_flags);
1934 1935 1936 1937
	} else {
		comp_status = le16_to_cpu(sts->comp_status);
		scsi_status = le16_to_cpu(sts->scsi_status) & SS_MASK;
	}
1938 1939 1940
	handle = (uint32_t) LSW(sts->handle);
	que = MSW(sts->handle);
	req = ha->req_q_map[que];
1941

L
Linus Torvalds 已提交
1942
	/* Validate handle. */
1943
	if (handle < req->num_outstanding_cmds)
1944
		sp = req->outstanding_cmds[handle];
1945
	else
L
Linus Torvalds 已提交
1946 1947 1948
		sp = NULL;

	if (sp == NULL) {
1949
		ql_dbg(ql_dbg_io, vha, 0x3017,
1950
		    "Invalid status handle (0x%x).\n", sts->handle);
L
Linus Torvalds 已提交
1951

1952 1953 1954 1955
		if (IS_QLA82XX(ha))
			set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
		else
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1956
		qla2xxx_wake_dpc(vha);
L
Linus Torvalds 已提交
1957 1958
		return;
	}
1959 1960 1961 1962 1963 1964 1965 1966

	if (unlikely((state_flags & BIT_1) && (sp->type == SRB_BIDI_CMD))) {
		qla25xx_process_bidir_status_iocb(vha, pkt, req, handle);
		return;
	}

	/* Fast path completion. */
	if (comp_status == CS_COMPLETE && scsi_status == 0) {
1967
		qla2x00_do_host_ramp_up(vha);
1968 1969 1970 1971 1972 1973
		qla2x00_process_completed_request(vha, req, handle);

		return;
	}

	req->outstanding_cmds[handle] = NULL;
1974
	cp = GET_CMD_SP(sp);
L
Linus Torvalds 已提交
1975
	if (cp == NULL) {
1976
		ql_dbg(ql_dbg_io, vha, 0x3018,
1977 1978
		    "Command already returned (0x%x/%p).\n",
		    sts->handle, sp);
L
Linus Torvalds 已提交
1979 1980 1981 1982

		return;
	}

1983
	lscsi_status = scsi_status & STATUS_MASK;
L
Linus Torvalds 已提交
1984

1985
	fcport = sp->fcport;
L
Linus Torvalds 已提交
1986

1987
	ox_id = 0;
1988 1989
	sense_len = par_sense_len = rsp_info_len = resid_len =
	    fw_resid_len = 0;
1990
	if (IS_FWI2_CAPABLE(ha)) {
1991 1992 1993 1994 1995 1996 1997 1998
		if (scsi_status & SS_SENSE_LEN_VALID)
			sense_len = le32_to_cpu(sts24->sense_len);
		if (scsi_status & SS_RESPONSE_INFO_LEN_VALID)
			rsp_info_len = le32_to_cpu(sts24->rsp_data_len);
		if (scsi_status & (SS_RESIDUAL_UNDER | SS_RESIDUAL_OVER))
			resid_len = le32_to_cpu(sts24->rsp_residual_count);
		if (comp_status == CS_DATA_UNDERRUN)
			fw_resid_len = le32_to_cpu(sts24->residual_len);
1999 2000 2001
		rsp_info = sts24->data;
		sense_data = sts24->data;
		host_to_fcp_swap(sts24->data, sizeof(sts24->data));
2002
		ox_id = le16_to_cpu(sts24->ox_id);
2003
		par_sense_len = sizeof(sts24->data);
2004
	} else {
2005 2006 2007 2008
		if (scsi_status & SS_SENSE_LEN_VALID)
			sense_len = le16_to_cpu(sts->req_sense_length);
		if (scsi_status & SS_RESPONSE_INFO_LEN_VALID)
			rsp_info_len = le16_to_cpu(sts->rsp_info_len);
2009 2010 2011
		resid_len = le32_to_cpu(sts->residual_length);
		rsp_info = sts->rsp_info;
		sense_data = sts->req_sense_data;
2012
		par_sense_len = sizeof(sts->req_sense_data);
2013 2014
	}

L
Linus Torvalds 已提交
2015 2016
	/* Check for any FCP transport errors. */
	if (scsi_status & SS_RESPONSE_INFO_LEN_VALID) {
2017
		/* Sense data lies beyond any FCP RESPONSE data. */
2018
		if (IS_FWI2_CAPABLE(ha)) {
2019
			sense_data += rsp_info_len;
2020 2021
			par_sense_len -= rsp_info_len;
		}
2022
		if (rsp_info_len > 3 && rsp_info[3]) {
2023
			ql_dbg(ql_dbg_io, fcport->vha, 0x3019,
2024 2025
			    "FCP I/O protocol failure (0x%x/0x%x).\n",
			    rsp_info_len, rsp_info[3]);
L
Linus Torvalds 已提交
2026

2027
			res = DID_BUS_BUSY << 16;
2028
			goto out;
L
Linus Torvalds 已提交
2029 2030 2031
		}
	}

2032 2033 2034 2035 2036
	/* Check for overrun. */
	if (IS_FWI2_CAPABLE(ha) && comp_status == CS_COMPLETE &&
	    scsi_status & SS_RESIDUAL_OVER)
		comp_status = CS_DATA_OVERRUN;

L
Linus Torvalds 已提交
2037 2038 2039 2040 2041
	/*
	 * Based on Host and scsi status generate status code for Linux
	 */
	switch (comp_status) {
	case CS_COMPLETE:
2042
	case CS_QUEUE_FULL:
L
Linus Torvalds 已提交
2043
		if (scsi_status == 0) {
2044
			res = DID_OK << 16;
L
Linus Torvalds 已提交
2045 2046 2047
			break;
		}
		if (scsi_status & (SS_RESIDUAL_UNDER | SS_RESIDUAL_OVER)) {
2048
			resid = resid_len;
2049
			scsi_set_resid(cp, resid);
2050 2051

			if (!lscsi_status &&
2052
			    ((unsigned)(scsi_bufflen(cp) - resid) <
2053
			     cp->underflow)) {
2054
				ql_dbg(ql_dbg_io, fcport->vha, 0x301a,
2055
				    "Mid-layer underflow "
2056
				    "detected (0x%x of 0x%x bytes).\n",
2057
				    resid, scsi_bufflen(cp));
2058

2059
				res = DID_ERROR << 16;
2060 2061
				break;
			}
L
Linus Torvalds 已提交
2062
		}
2063
		res = DID_OK << 16 | lscsi_status;
L
Linus Torvalds 已提交
2064

2065
		if (lscsi_status == SAM_STAT_TASK_SET_FULL) {
2066
			ql_dbg(ql_dbg_io, fcport->vha, 0x301b,
2067
			    "QUEUE FULL detected.\n");
2068 2069
			break;
		}
2070
		logit = 0;
L
Linus Torvalds 已提交
2071 2072 2073
		if (lscsi_status != SS_CHECK_CONDITION)
			break;

2074
		memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
L
Linus Torvalds 已提交
2075 2076 2077
		if (!(scsi_status & SS_SENSE_LEN_VALID))
			break;

2078
		qla2x00_handle_sense(sp, sense_data, par_sense_len, sense_len,
2079
		    rsp, res);
L
Linus Torvalds 已提交
2080 2081 2082
		break;

	case CS_DATA_UNDERRUN:
2083
		/* Use F/W calculated residual length. */
2084 2085 2086 2087
		resid = IS_FWI2_CAPABLE(ha) ? fw_resid_len : resid_len;
		scsi_set_resid(cp, resid);
		if (scsi_status & SS_RESIDUAL_UNDER) {
			if (IS_FWI2_CAPABLE(ha) && fw_resid_len != resid_len) {
2088
				ql_dbg(ql_dbg_io, fcport->vha, 0x301d,
2089 2090 2091
				    "Dropped frame(s) detected "
				    "(0x%x of 0x%x bytes).\n",
				    resid, scsi_bufflen(cp));
2092

2093
				res = DID_ERROR << 16 | lscsi_status;
2094
				goto check_scsi_status;
2095
			}
2096

2097 2098 2099
			if (!lscsi_status &&
			    ((unsigned)(scsi_bufflen(cp) - resid) <
			    cp->underflow)) {
2100
				ql_dbg(ql_dbg_io, fcport->vha, 0x301e,
2101
				    "Mid-layer underflow "
2102
				    "detected (0x%x of 0x%x bytes).\n",
2103
				    resid, scsi_bufflen(cp));
2104

2105
				res = DID_ERROR << 16;
2106 2107
				break;
			}
2108 2109 2110 2111 2112 2113 2114
		} else if (lscsi_status != SAM_STAT_TASK_SET_FULL &&
			    lscsi_status != SAM_STAT_BUSY) {
			/*
			 * scsi status of task set and busy are considered to be
			 * task not completed.
			 */

2115
			ql_dbg(ql_dbg_io, fcport->vha, 0x301f,
2116
			    "Dropped frame(s) detected (0x%x "
2117 2118
			    "of 0x%x bytes).\n", resid,
			    scsi_bufflen(cp));
2119

2120
			res = DID_ERROR << 16 | lscsi_status;
2121
			goto check_scsi_status;
2122 2123 2124 2125
		} else {
			ql_dbg(ql_dbg_io, fcport->vha, 0x3030,
			    "scsi_status: 0x%x, lscsi_status: 0x%x\n",
			    scsi_status, lscsi_status);
L
Linus Torvalds 已提交
2126 2127
		}

2128
		res = DID_OK << 16 | lscsi_status;
2129
		logit = 0;
2130

2131
check_scsi_status:
L
Linus Torvalds 已提交
2132
		/*
A
Andrew Vasquez 已提交
2133
		 * Check to see if SCSI Status is non zero. If so report SCSI
L
Linus Torvalds 已提交
2134 2135 2136
		 * Status.
		 */
		if (lscsi_status != 0) {
2137
			if (lscsi_status == SAM_STAT_TASK_SET_FULL) {
2138
				ql_dbg(ql_dbg_io, fcport->vha, 0x3020,
2139
				    "QUEUE FULL detected.\n");
2140
				logit = 1;
2141 2142
				break;
			}
L
Linus Torvalds 已提交
2143 2144 2145
			if (lscsi_status != SS_CHECK_CONDITION)
				break;

2146
			memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
L
Linus Torvalds 已提交
2147 2148 2149
			if (!(scsi_status & SS_SENSE_LEN_VALID))
				break;

2150
			qla2x00_handle_sense(sp, sense_data, par_sense_len,
2151
			    sense_len, rsp, res);
L
Linus Torvalds 已提交
2152 2153 2154 2155 2156 2157 2158 2159
		}
		break;

	case CS_PORT_LOGGED_OUT:
	case CS_PORT_CONFIG_CHG:
	case CS_PORT_BUSY:
	case CS_INCOMPLETE:
	case CS_PORT_UNAVAILABLE:
2160
	case CS_TIMEOUT:
2161 2162
	case CS_RESET:

2163 2164 2165 2166 2167
		/*
		 * We are going to have the fc class block the rport
		 * while we try to recover so instruct the mid layer
		 * to requeue until the class decides how to handle this.
		 */
2168
		res = DID_TRANSPORT_DISRUPTED << 16;
2169 2170 2171 2172 2173 2174 2175 2176 2177

		if (comp_status == CS_TIMEOUT) {
			if (IS_FWI2_CAPABLE(ha))
				break;
			else if ((le16_to_cpu(sts->status_flags) &
			    SF_LOGOUT_SENT) == 0)
				break;
		}

2178
		ql_dbg(ql_dbg_io, fcport->vha, 0x3021,
2179 2180
		    "Port down status: port-state=0x%x.\n",
		    atomic_read(&fcport->state));
2181

2182
		if (atomic_read(&fcport->state) == FCS_ONLINE)
2183
			qla2x00_mark_device_lost(fcport->vha, fcport, 1, 1);
L
Linus Torvalds 已提交
2184 2185 2186
		break;

	case CS_ABORTED:
2187
		res = DID_RESET << 16;
L
Linus Torvalds 已提交
2188
		break;
2189 2190

	case CS_DIF_ERROR:
2191
		logit = qla2x00_handle_dif_error(sp, sts24);
2192
		res = cp->result;
2193
		break;
2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206

	case CS_TRANSPORT:
		res = DID_ERROR << 16;

		if (!IS_PI_SPLIT_DET_CAPABLE(ha))
			break;

		if (state_flags & BIT_4)
			scmd_printk(KERN_WARNING, cp,
			    "Unsupported device '%s' found.\n",
			    cp->device->vendor);
		break;

L
Linus Torvalds 已提交
2207
	default:
2208
		res = DID_ERROR << 16;
L
Linus Torvalds 已提交
2209 2210 2211
		break;
	}

2212 2213
out:
	if (logit)
2214
		ql_dbg(ql_dbg_io, fcport->vha, 0x3022,
2215 2216
		    "FCP command status: 0x%x-0x%x (0x%x) nexus=%ld:%d:%d "
		    "portid=%02x%02x%02x oxid=0x%x cdb=%10phN len=0x%x "
2217
		    "rsp_info=0x%x resid=0x%x fw_resid=0x%x.\n",
2218
		    comp_status, scsi_status, res, vha->host_no,
2219 2220
		    cp->device->id, cp->device->lun, fcport->d_id.b.domain,
		    fcport->d_id.b.area, fcport->d_id.b.al_pa, ox_id,
2221
		    cp->cmnd, scsi_bufflen(cp), rsp_info_len,
2222
		    resid_len, fw_resid_len);
2223

2224 2225 2226
	if (!res)
		qla2x00_do_host_ramp_up(vha);

2227
	if (rsp->status_srb == NULL)
2228
		sp->done(ha, sp, res);
L
Linus Torvalds 已提交
2229 2230 2231 2232 2233 2234 2235 2236 2237 2238
}

/**
 * qla2x00_status_cont_entry() - Process a Status Continuations entry.
 * @ha: SCSI driver HA context
 * @pkt: Entry pointer
 *
 * Extended sense data.
 */
static void
2239
qla2x00_status_cont_entry(struct rsp_que *rsp, sts_cont_entry_t *pkt)
L
Linus Torvalds 已提交
2240
{
2241
	uint8_t	sense_sz = 0;
2242
	struct qla_hw_data *ha = rsp->hw;
2243
	struct scsi_qla_host *vha = pci_get_drvdata(ha->pdev);
2244
	srb_t *sp = rsp->status_srb;
L
Linus Torvalds 已提交
2245
	struct scsi_cmnd *cp;
2246 2247
	uint32_t sense_len;
	uint8_t *sense_ptr;
L
Linus Torvalds 已提交
2248

2249 2250
	if (!sp || !GET_CMD_SENSE_LEN(sp))
		return;
L
Linus Torvalds 已提交
2251

2252 2253
	sense_len = GET_CMD_SENSE_LEN(sp);
	sense_ptr = GET_CMD_SENSE_PTR(sp);
L
Linus Torvalds 已提交
2254

2255 2256 2257 2258
	cp = GET_CMD_SP(sp);
	if (cp == NULL) {
		ql_log(ql_log_warn, vha, 0x3025,
		    "cmd is NULL: already returned to OS (sp=%p).\n", sp);
L
Linus Torvalds 已提交
2259

2260 2261
		rsp->status_srb = NULL;
		return;
L
Linus Torvalds 已提交
2262 2263
	}

2264 2265 2266 2267
	if (sense_len > sizeof(pkt->data))
		sense_sz = sizeof(pkt->data);
	else
		sense_sz = sense_len;
2268

2269 2270 2271 2272 2273 2274
	/* Move sense data. */
	if (IS_FWI2_CAPABLE(ha))
		host_to_fcp_swap(pkt->data, sizeof(pkt->data));
	memcpy(sense_ptr, pkt->data, sense_sz);
	ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x302c,
		sense_ptr, sense_sz);
2275

2276 2277
	sense_len -= sense_sz;
	sense_ptr += sense_sz;
2278

2279 2280 2281 2282 2283 2284 2285
	SET_CMD_SENSE_PTR(sp, sense_ptr);
	SET_CMD_SENSE_LEN(sp, sense_len);

	/* Place command on done queue. */
	if (sense_len == 0) {
		rsp->status_srb = NULL;
		sp->done(ha, sp, cp->result);
2286 2287 2288
	}
}

L
Linus Torvalds 已提交
2289 2290 2291 2292 2293 2294
/**
 * qla2x00_error_entry() - Process an error entry.
 * @ha: SCSI driver HA context
 * @pkt: Entry pointer
 */
static void
2295
qla2x00_error_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, sts_entry_t *pkt)
L
Linus Torvalds 已提交
2296 2297
{
	srb_t *sp;
2298
	struct qla_hw_data *ha = vha->hw;
2299
	const char func[] = "ERROR-IOCB";
2300
	uint16_t que = MSW(pkt->handle);
2301
	struct req_que *req = NULL;
2302
	int res = DID_ERROR << 16;
2303

2304 2305 2306
	ql_dbg(ql_dbg_async, vha, 0x502a,
	    "type of error status in response: 0x%x\n", pkt->entry_status);

2307 2308 2309 2310 2311
	if (que >= ha->max_req_queues || !ha->req_q_map[que])
		goto fatal;

	req = ha->req_q_map[que];

2312 2313
	if (pkt->entry_status & RF_BUSY)
		res = DID_BUS_BUSY << 16;
L
Linus Torvalds 已提交
2314

2315
	sp = qla2x00_get_sp_from_handle(vha, func, req, pkt);
2316
	if (sp) {
2317
		sp->done(ha, sp, res);
2318
		return;
L
Linus Torvalds 已提交
2319
	}
2320 2321 2322 2323 2324 2325 2326 2327 2328
fatal:
	ql_log(ql_log_warn, vha, 0x5030,
	    "Error entry - invalid handle/queue.\n");

	if (IS_QLA82XX(ha))
		set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
	else
		set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
	qla2xxx_wake_dpc(vha);
L
Linus Torvalds 已提交
2329 2330
}

2331 2332 2333 2334 2335 2336
/**
 * qla24xx_mbx_completion() - Process mailbox command completions.
 * @ha: SCSI driver HA context
 * @mb0: Mailbox0 register
 */
static void
2337
qla24xx_mbx_completion(scsi_qla_host_t *vha, uint16_t mb0)
2338 2339
{
	uint16_t	cnt;
2340
	uint32_t	mboxes;
2341
	uint16_t __iomem *wptr;
2342
	struct qla_hw_data *ha = vha->hw;
2343 2344
	struct device_reg_24xx __iomem *reg = &ha->iobase->isp24;

2345 2346 2347
	/* Read all mbox registers? */
	mboxes = (1 << ha->mbx_count) - 1;
	if (!ha->mcp)
2348
		ql_dbg(ql_dbg_async, vha, 0x504e, "MBX pointer ERROR.\n");
2349 2350 2351
	else
		mboxes = ha->mcp->in_mb;

2352 2353 2354
	/* Load return mailbox registers. */
	ha->flags.mbox_int = 1;
	ha->mailbox_out[0] = mb0;
2355
	mboxes >>= 1;
2356 2357 2358
	wptr = (uint16_t __iomem *)&reg->mailbox1;

	for (cnt = 1; cnt < ha->mbx_count; cnt++) {
2359 2360 2361 2362
		if (mboxes & BIT_0)
			ha->mailbox_out[cnt] = RD_REG_WORD(wptr);

		mboxes >>= 1;
2363 2364 2365 2366 2367 2368 2369 2370
		wptr++;
	}
}

/**
 * qla24xx_process_response_queue() - Process response queue entries.
 * @ha: SCSI driver HA context
 */
2371 2372
void qla24xx_process_response_queue(struct scsi_qla_host *vha,
	struct rsp_que *rsp)
2373 2374
{
	struct sts_entry_24xx *pkt;
2375
	struct qla_hw_data *ha = vha->hw;
2376

2377
	if (!vha->flags.online)
2378 2379
		return;

2380 2381
	while (rsp->ring_ptr->signature != RESPONSE_PROCESSED) {
		pkt = (struct sts_entry_24xx *)rsp->ring_ptr;
2382

2383 2384 2385 2386
		rsp->ring_index++;
		if (rsp->ring_index == rsp->length) {
			rsp->ring_index = 0;
			rsp->ring_ptr = rsp->ring;
2387
		} else {
2388
			rsp->ring_ptr++;
2389 2390 2391
		}

		if (pkt->entry_status != 0) {
2392
			qla2x00_error_entry(vha, rsp, (sts_entry_t *) pkt);
2393 2394 2395

			(void)qlt_24xx_process_response_error(vha, pkt);

2396 2397 2398 2399 2400 2401 2402
			((response_t *)pkt)->signature = RESPONSE_PROCESSED;
			wmb();
			continue;
		}

		switch (pkt->entry_type) {
		case STATUS_TYPE:
2403
			qla2x00_status_entry(vha, rsp, pkt);
2404 2405
			break;
		case STATUS_CONT_TYPE:
2406
			qla2x00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt);
2407
			break;
2408
		case VP_RPT_ID_IOCB_TYPE:
2409
			qla24xx_report_id_acquisition(vha,
2410 2411
			    (struct vp_rpt_id_entry_24xx *)pkt);
			break;
2412 2413 2414 2415
		case LOGINOUT_PORT_IOCB_TYPE:
			qla24xx_logio_entry(vha, rsp->req,
			    (struct logio_entry_24xx *)pkt);
			break;
2416 2417 2418 2419
		case TSK_MGMT_IOCB_TYPE:
			qla24xx_tm_iocb_entry(vha, rsp->req,
			    (struct tsk_mgmt_entry *)pkt);
			break;
2420 2421 2422 2423 2424 2425
                case CT_IOCB_TYPE:
			qla24xx_els_ct_entry(vha, rsp->req, pkt, CT_IOCB_TYPE);
			break;
                case ELS_IOCB_TYPE:
			qla24xx_els_ct_entry(vha, rsp->req, pkt, ELS_IOCB_TYPE);
			break;
2426 2427 2428 2429 2430 2431 2432 2433
		case ABTS_RECV_24XX:
			/* ensure that the ATIO queue is empty */
			qlt_24xx_process_atio_queue(vha);
		case ABTS_RESP_24XX:
		case CTIO_TYPE7:
		case NOTIFY_ACK_TYPE:
			qlt_response_pkt_all_vps(vha, (response_t *)pkt);
			break;
2434 2435 2436 2437 2438
		case MARKER_TYPE:
			/* Do nothing in this case, this check is to prevent it
			 * from falling into default case
			 */
			break;
2439 2440
		default:
			/* Type Not Supported. */
2441 2442
			ql_dbg(ql_dbg_async, vha, 0x5042,
			    "Received unknown response pkt type %x "
2443
			    "entry status=%x.\n",
2444
			    pkt->entry_type, pkt->entry_status);
2445 2446 2447 2448 2449 2450 2451
			break;
		}
		((response_t *)pkt)->signature = RESPONSE_PROCESSED;
		wmb();
	}

	/* Adjust ring index */
2452 2453 2454 2455 2456
	if (IS_QLA82XX(ha)) {
		struct device_reg_82xx __iomem *reg = &ha->iobase->isp82;
		WRT_REG_DWORD(&reg->rsp_q_out[0], rsp->ring_index);
	} else
		WRT_REG_DWORD(rsp->rsp_q_out, rsp->ring_index);
2457 2458
}

2459
static void
2460
qla2xxx_check_risc_status(scsi_qla_host_t *vha)
2461 2462 2463
{
	int rval;
	uint32_t cnt;
2464
	struct qla_hw_data *ha = vha->hw;
2465 2466
	struct device_reg_24xx __iomem *reg = &ha->iobase->isp24;

2467
	if (!IS_QLA25XX(ha) && !IS_QLA81XX(ha) && !IS_QLA83XX(ha))
2468 2469 2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481 2482 2483 2484
		return;

	rval = QLA_SUCCESS;
	WRT_REG_DWORD(&reg->iobase_addr, 0x7C00);
	RD_REG_DWORD(&reg->iobase_addr);
	WRT_REG_DWORD(&reg->iobase_window, 0x0001);
	for (cnt = 10000; (RD_REG_DWORD(&reg->iobase_window) & BIT_0) == 0 &&
	    rval == QLA_SUCCESS; cnt--) {
		if (cnt) {
			WRT_REG_DWORD(&reg->iobase_window, 0x0001);
			udelay(10);
		} else
			rval = QLA_FUNCTION_TIMEOUT;
	}
	if (rval == QLA_SUCCESS)
		goto next_test;

2485
	rval = QLA_SUCCESS;
2486 2487 2488 2489 2490 2491 2492 2493 2494 2495 2496 2497 2498 2499
	WRT_REG_DWORD(&reg->iobase_window, 0x0003);
	for (cnt = 100; (RD_REG_DWORD(&reg->iobase_window) & BIT_0) == 0 &&
	    rval == QLA_SUCCESS; cnt--) {
		if (cnt) {
			WRT_REG_DWORD(&reg->iobase_window, 0x0003);
			udelay(10);
		} else
			rval = QLA_FUNCTION_TIMEOUT;
	}
	if (rval != QLA_SUCCESS)
		goto done;

next_test:
	if (RD_REG_DWORD(&reg->iobase_c8) & BIT_3)
2500 2501
		ql_log(ql_log_info, vha, 0x504c,
		    "Additional code -- 0x55AA.\n");
2502 2503 2504 2505 2506 2507

done:
	WRT_REG_DWORD(&reg->iobase_window, 0x0000);
	RD_REG_DWORD(&reg->iobase_window);
}

2508
/**
2509
 * qla24xx_intr_handler() - Process interrupts for the ISP23xx and ISP24xx.
2510 2511 2512 2513 2514 2515 2516 2517
 * @irq:
 * @dev_id: SCSI driver HA context
 *
 * Called by system whenever the host adapter generates an interrupt.
 *
 * Returns handled flag.
 */
irqreturn_t
2518
qla24xx_intr_handler(int irq, void *dev_id)
2519
{
2520 2521
	scsi_qla_host_t	*vha;
	struct qla_hw_data *ha;
2522 2523 2524 2525 2526
	struct device_reg_24xx __iomem *reg;
	int		status;
	unsigned long	iter;
	uint32_t	stat;
	uint32_t	hccr;
2527
	uint16_t	mb[8];
2528
	struct rsp_que *rsp;
2529
	unsigned long	flags;
2530

2531 2532
	rsp = (struct rsp_que *) dev_id;
	if (!rsp) {
2533 2534
		ql_log(ql_log_info, NULL, 0x5059,
		    "%s: NULL response queue pointer.\n", __func__);
2535 2536 2537
		return IRQ_NONE;
	}

2538
	ha = rsp->hw;
2539 2540 2541
	reg = &ha->iobase->isp24;
	status = 0;

2542 2543 2544
	if (unlikely(pci_channel_offline(ha->pdev)))
		return IRQ_HANDLED;

2545
	spin_lock_irqsave(&ha->hardware_lock, flags);
2546
	vha = pci_get_drvdata(ha->pdev);
2547 2548 2549
	for (iter = 50; iter--; ) {
		stat = RD_REG_DWORD(&reg->host_status);
		if (stat & HSRX_RISC_PAUSED) {
2550
			if (unlikely(pci_channel_offline(ha->pdev)))
2551 2552
				break;

2553 2554
			hccr = RD_REG_DWORD(&reg->hccr);

2555 2556 2557
			ql_log(ql_log_warn, vha, 0x504b,
			    "RISC paused -- HCCR=%x, Dumping firmware.\n",
			    hccr);
2558

2559
			qla2xxx_check_risc_status(vha);
2560

2561 2562
			ha->isp_ops->fw_dump(vha, 1);
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2563 2564 2565 2566 2567
			break;
		} else if ((stat & HSRX_RISC_INT) == 0)
			break;

		switch (stat & 0xff) {
2568 2569 2570 2571
		case INTR_ROM_MB_SUCCESS:
		case INTR_ROM_MB_FAILED:
		case INTR_MB_SUCCESS:
		case INTR_MB_FAILED:
2572
			qla24xx_mbx_completion(vha, MSW(stat));
2573 2574 2575
			status |= MBX_INTERRUPT;

			break;
2576
		case INTR_ASYNC_EVENT:
2577 2578 2579 2580
			mb[0] = MSW(stat);
			mb[1] = RD_REG_WORD(&reg->mailbox1);
			mb[2] = RD_REG_WORD(&reg->mailbox2);
			mb[3] = RD_REG_WORD(&reg->mailbox3);
2581
			qla2x00_async_event(vha, rsp, mb);
2582
			break;
2583 2584
		case INTR_RSP_QUE_UPDATE:
		case INTR_RSP_QUE_UPDATE_83XX:
2585
			qla24xx_process_response_queue(vha, rsp);
2586
			break;
2587
		case INTR_ATIO_QUE_UPDATE:
2588 2589
			qlt_24xx_process_atio_queue(vha);
			break;
2590
		case INTR_ATIO_RSP_QUE_UPDATE:
2591 2592 2593
			qlt_24xx_process_atio_queue(vha);
			qla24xx_process_response_queue(vha, rsp);
			break;
2594
		default:
2595 2596
			ql_dbg(ql_dbg_async, vha, 0x504f,
			    "Unrecognized interrupt type (%d).\n", stat * 0xff);
2597 2598 2599 2600
			break;
		}
		WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
		RD_REG_DWORD_RELAXED(&reg->hccr);
2601 2602
		if (unlikely(IS_QLA83XX(ha) && (ha->pdev->revision == 1)))
			ndelay(3500);
2603
	}
2604
	qla2x00_handle_mbx_completion(ha, status);
2605
	spin_unlock_irqrestore(&ha->hardware_lock, flags);
2606 2607 2608 2609

	return IRQ_HANDLED;
}

2610 2611 2612
static irqreturn_t
qla24xx_msix_rsp_q(int irq, void *dev_id)
{
2613 2614
	struct qla_hw_data *ha;
	struct rsp_que *rsp;
2615
	struct device_reg_24xx __iomem *reg;
2616
	struct scsi_qla_host *vha;
2617
	unsigned long flags;
2618

2619 2620
	rsp = (struct rsp_que *) dev_id;
	if (!rsp) {
2621 2622
		ql_log(ql_log_info, NULL, 0x505a,
		    "%s: NULL response queue pointer.\n", __func__);
2623 2624 2625
		return IRQ_NONE;
	}
	ha = rsp->hw;
2626 2627
	reg = &ha->iobase->isp24;

2628
	spin_lock_irqsave(&ha->hardware_lock, flags);
2629

2630
	vha = pci_get_drvdata(ha->pdev);
2631
	qla24xx_process_response_queue(vha, rsp);
2632
	if (!ha->flags.disable_msix_handshake) {
2633 2634 2635
		WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
		RD_REG_DWORD_RELAXED(&reg->hccr);
	}
2636
	spin_unlock_irqrestore(&ha->hardware_lock, flags);
2637 2638 2639 2640

	return IRQ_HANDLED;
}

2641 2642 2643 2644 2645
static irqreturn_t
qla25xx_msix_rsp_q(int irq, void *dev_id)
{
	struct qla_hw_data *ha;
	struct rsp_que *rsp;
2646
	struct device_reg_24xx __iomem *reg;
2647
	unsigned long flags;
2648 2649 2650

	rsp = (struct rsp_que *) dev_id;
	if (!rsp) {
2651 2652
		ql_log(ql_log_info, NULL, 0x505b,
		    "%s: NULL response queue pointer.\n", __func__);
2653 2654 2655 2656
		return IRQ_NONE;
	}
	ha = rsp->hw;

2657
	/* Clear the interrupt, if enabled, for this response queue */
2658
	if (!ha->flags.disable_msix_handshake) {
2659
		reg = &ha->iobase->isp24;
2660
		spin_lock_irqsave(&ha->hardware_lock, flags);
2661 2662
		WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
		RD_REG_DWORD_RELAXED(&reg->hccr);
2663
		spin_unlock_irqrestore(&ha->hardware_lock, flags);
2664
	}
2665 2666 2667 2668 2669
	queue_work_on((int) (rsp->id - 1), ha->wq, &rsp->q_work);

	return IRQ_HANDLED;
}

2670 2671 2672
static irqreturn_t
qla24xx_msix_default(int irq, void *dev_id)
{
2673 2674 2675
	scsi_qla_host_t	*vha;
	struct qla_hw_data *ha;
	struct rsp_que *rsp;
2676 2677 2678 2679
	struct device_reg_24xx __iomem *reg;
	int		status;
	uint32_t	stat;
	uint32_t	hccr;
2680
	uint16_t	mb[8];
2681
	unsigned long flags;
2682

2683 2684
	rsp = (struct rsp_que *) dev_id;
	if (!rsp) {
2685 2686
		ql_log(ql_log_info, NULL, 0x505c,
		    "%s: NULL response queue pointer.\n", __func__);
2687 2688 2689
		return IRQ_NONE;
	}
	ha = rsp->hw;
2690 2691 2692
	reg = &ha->iobase->isp24;
	status = 0;

2693
	spin_lock_irqsave(&ha->hardware_lock, flags);
2694
	vha = pci_get_drvdata(ha->pdev);
2695
	do {
2696 2697
		stat = RD_REG_DWORD(&reg->host_status);
		if (stat & HSRX_RISC_PAUSED) {
2698
			if (unlikely(pci_channel_offline(ha->pdev)))
2699 2700
				break;

2701 2702
			hccr = RD_REG_DWORD(&reg->hccr);

2703 2704 2705
			ql_log(ql_log_info, vha, 0x5050,
			    "RISC paused -- HCCR=%x, Dumping firmware.\n",
			    hccr);
2706

2707
			qla2xxx_check_risc_status(vha);
2708

2709 2710
			ha->isp_ops->fw_dump(vha, 1);
			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2711 2712 2713 2714 2715
			break;
		} else if ((stat & HSRX_RISC_INT) == 0)
			break;

		switch (stat & 0xff) {
2716 2717 2718 2719
		case INTR_ROM_MB_SUCCESS:
		case INTR_ROM_MB_FAILED:
		case INTR_MB_SUCCESS:
		case INTR_MB_FAILED:
2720
			qla24xx_mbx_completion(vha, MSW(stat));
2721 2722 2723
			status |= MBX_INTERRUPT;

			break;
2724
		case INTR_ASYNC_EVENT:
2725 2726 2727 2728
			mb[0] = MSW(stat);
			mb[1] = RD_REG_WORD(&reg->mailbox1);
			mb[2] = RD_REG_WORD(&reg->mailbox2);
			mb[3] = RD_REG_WORD(&reg->mailbox3);
2729
			qla2x00_async_event(vha, rsp, mb);
2730
			break;
2731 2732
		case INTR_RSP_QUE_UPDATE:
		case INTR_RSP_QUE_UPDATE_83XX:
2733
			qla24xx_process_response_queue(vha, rsp);
2734
			break;
2735
		case INTR_ATIO_QUE_UPDATE:
2736 2737
			qlt_24xx_process_atio_queue(vha);
			break;
2738
		case INTR_ATIO_RSP_QUE_UPDATE:
2739 2740 2741
			qlt_24xx_process_atio_queue(vha);
			qla24xx_process_response_queue(vha, rsp);
			break;
2742
		default:
2743 2744
			ql_dbg(ql_dbg_async, vha, 0x5051,
			    "Unrecognized interrupt type (%d).\n", stat & 0xff);
2745 2746 2747
			break;
		}
		WRT_REG_DWORD(&reg->hccr, HCCRX_CLR_RISC_INT);
2748
	} while (0);
2749
	qla2x00_handle_mbx_completion(ha, status);
2750
	spin_unlock_irqrestore(&ha->hardware_lock, flags);
2751 2752 2753 2754 2755 2756 2757 2758

	return IRQ_HANDLED;
}

/* Interrupt handling helpers. */

struct qla_init_msix_entry {
	const char *name;
2759
	irq_handler_t handler;
2760 2761
};

2762
static struct qla_init_msix_entry msix_entries[3] = {
2763 2764
	{ "qla2xxx (default)", qla24xx_msix_default },
	{ "qla2xxx (rsp_q)", qla24xx_msix_rsp_q },
2765
	{ "qla2xxx (multiq)", qla25xx_msix_rsp_q },
2766 2767
};

2768 2769 2770 2771 2772
static struct qla_init_msix_entry qla82xx_msix_entries[2] = {
	{ "qla2xxx (default)", qla82xx_msix_default },
	{ "qla2xxx (rsp_q)", qla82xx_msix_rsp_q },
};

2773 2774 2775 2776 2777 2778
static struct qla_init_msix_entry qla83xx_msix_entries[3] = {
	{ "qla2xxx (default)", qla24xx_msix_default },
	{ "qla2xxx (rsp_q)", qla24xx_msix_rsp_q },
	{ "qla2xxx (atio_q)", qla83xx_msix_atio_q },
};

2779
static void
2780
qla24xx_disable_msix(struct qla_hw_data *ha)
2781 2782 2783
{
	int i;
	struct qla_msix_entry *qentry;
2784
	scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
2785

2786 2787
	for (i = 0; i < ha->msix_count; i++) {
		qentry = &ha->msix_entries[i];
2788
		if (qentry->have_irq)
2789
			free_irq(qentry->vector, qentry->rsp);
2790 2791
	}
	pci_disable_msix(ha->pdev);
2792 2793 2794
	kfree(ha->msix_entries);
	ha->msix_entries = NULL;
	ha->flags.msix_enabled = 0;
2795 2796
	ql_dbg(ql_dbg_init, vha, 0x0042,
	    "Disabled the MSI.\n");
2797 2798 2799
}

static int
2800
qla24xx_enable_msix(struct qla_hw_data *ha, struct rsp_que *rsp)
2801
{
2802
#define MIN_MSIX_COUNT	2
2803
	int i, ret;
2804
	struct msix_entry *entries;
2805
	struct qla_msix_entry *qentry;
2806
	scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
2807 2808

	entries = kzalloc(sizeof(struct msix_entry) * ha->msix_count,
2809
			GFP_KERNEL);
2810 2811 2812
	if (!entries) {
		ql_log(ql_log_warn, vha, 0x00bc,
		    "Failed to allocate memory for msix_entry.\n");
2813
		return -ENOMEM;
2814
	}
2815

2816 2817
	for (i = 0; i < ha->msix_count; i++)
		entries[i].entry = i;
2818

2819
	ret = pci_enable_msix(ha->pdev, entries, ha->msix_count);
2820
	if (ret) {
2821 2822 2823
		if (ret < MIN_MSIX_COUNT)
			goto msix_failed;

2824 2825 2826 2827
		ql_log(ql_log_warn, vha, 0x00c6,
		    "MSI-X: Failed to enable support "
		    "-- %d/%d\n Retry with %d vectors.\n",
		    ha->msix_count, ret, ret);
2828 2829 2830
		ha->msix_count = ret;
		ret = pci_enable_msix(ha->pdev, entries, ha->msix_count);
		if (ret) {
2831
msix_failed:
2832 2833 2834 2835
			ql_log(ql_log_fatal, vha, 0x00c7,
			    "MSI-X: Failed to enable support, "
			    "giving   up -- %d/%d.\n",
			    ha->msix_count, ret);
2836 2837
			goto msix_out;
		}
2838
		ha->max_rsp_queues = ha->msix_count - 1;
2839 2840 2841 2842
	}
	ha->msix_entries = kzalloc(sizeof(struct qla_msix_entry) *
				ha->msix_count, GFP_KERNEL);
	if (!ha->msix_entries) {
2843 2844
		ql_log(ql_log_fatal, vha, 0x00c8,
		    "Failed to allocate memory for ha->msix_entries.\n");
2845
		ret = -ENOMEM;
2846 2847 2848 2849
		goto msix_out;
	}
	ha->flags.msix_enabled = 1;

2850 2851 2852 2853
	for (i = 0; i < ha->msix_count; i++) {
		qentry = &ha->msix_entries[i];
		qentry->vector = entries[i].vector;
		qentry->entry = entries[i].entry;
2854
		qentry->have_irq = 0;
2855
		qentry->rsp = NULL;
2856 2857
	}

2858
	/* Enable MSI-X vectors for the base queue */
2859
	for (i = 0; i < ha->msix_count; i++) {
2860
		qentry = &ha->msix_entries[i];
2861 2862 2863 2864 2865
		if (QLA_TGT_MODE_ENABLED() && IS_ATIO_MSIX_CAPABLE(ha)) {
			ret = request_irq(qentry->vector,
				qla83xx_msix_entries[i].handler,
				0, qla83xx_msix_entries[i].name, rsp);
		} else if (IS_QLA82XX(ha)) {
2866 2867 2868 2869 2870 2871 2872 2873
			ret = request_irq(qentry->vector,
				qla82xx_msix_entries[i].handler,
				0, qla82xx_msix_entries[i].name, rsp);
		} else {
			ret = request_irq(qentry->vector,
				msix_entries[i].handler,
				0, msix_entries[i].name, rsp);
		}
2874
		if (ret) {
2875 2876 2877
			ql_log(ql_log_fatal, vha, 0x00cb,
			    "MSI-X: unable to register handler -- %x/%d.\n",
			    qentry->vector, ret);
2878 2879 2880 2881 2882 2883 2884
			qla24xx_disable_msix(ha);
			ha->mqenable = 0;
			goto msix_out;
		}
		qentry->have_irq = 1;
		qentry->rsp = rsp;
		rsp->msix = qentry;
2885 2886 2887
	}

	/* Enable MSI-X vector for response queue update for queue 0 */
2888 2889 2890 2891 2892 2893 2894 2895
	if (IS_QLA83XX(ha)) {
		if (ha->msixbase && ha->mqiobase &&
		    (ha->max_rsp_queues > 1 || ha->max_req_queues > 1))
			ha->mqenable = 1;
	} else
		if (ha->mqiobase
		    && (ha->max_rsp_queues > 1 || ha->max_req_queues > 1))
			ha->mqenable = 1;
2896 2897 2898 2899 2900 2901
	ql_dbg(ql_dbg_multiq, vha, 0xc005,
	    "mqiobase=%p, max_rsp_queues=%d, max_req_queues=%d.\n",
	    ha->mqiobase, ha->max_rsp_queues, ha->max_req_queues);
	ql_dbg(ql_dbg_init, vha, 0x0055,
	    "mqiobase=%p, max_rsp_queues=%d, max_req_queues=%d.\n",
	    ha->mqiobase, ha->max_rsp_queues, ha->max_req_queues);
2902

2903
msix_out:
2904
	kfree(entries);
2905 2906 2907 2908
	return ret;
}

int
2909
qla2x00_request_irqs(struct qla_hw_data *ha, struct rsp_que *rsp)
2910 2911
{
	int ret;
2912
	device_reg_t __iomem *reg = ha->iobase;
2913
	scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
2914 2915

	/* If possible, enable MSI-X. */
2916
	if (!IS_QLA2432(ha) && !IS_QLA2532(ha) && !IS_QLA8432(ha) &&
2917
		!IS_CNA_CAPABLE(ha) && !IS_QLA2031(ha) && !IS_QLAFX00(ha))
2918 2919 2920 2921 2922 2923
		goto skip_msi;

	if (ha->pdev->subsystem_vendor == PCI_VENDOR_ID_HP &&
		(ha->pdev->subsystem_device == 0x7040 ||
		ha->pdev->subsystem_device == 0x7041 ||
		ha->pdev->subsystem_device == 0x1705)) {
2924 2925
		ql_log(ql_log_warn, vha, 0x0034,
		    "MSI-X: Unsupported ISP 2432 SSVID/SSDID (0x%X,0x%X).\n",
2926
			ha->pdev->subsystem_vendor,
2927
			ha->pdev->subsystem_device);
2928 2929
		goto skip_msi;
	}
2930

2931
	if (IS_QLA2432(ha) && (ha->pdev->revision < QLA_MSIX_CHIP_REV_24XX)) {
2932 2933
		ql_log(ql_log_warn, vha, 0x0035,
		    "MSI-X; Unsupported ISP2432 (0x%X, 0x%X).\n",
2934
		    ha->pdev->revision, QLA_MSIX_CHIP_REV_24XX);
2935 2936 2937
		goto skip_msix;
	}

2938
	ret = qla24xx_enable_msix(ha, rsp);
2939
	if (!ret) {
2940 2941 2942
		ql_dbg(ql_dbg_init, vha, 0x0036,
		    "MSI-X: Enabled (0x%X, 0x%X).\n",
		    ha->chip_revision, ha->fw_attributes);
2943
		goto clear_risc_ints;
2944
	}
2945 2946
	ql_log(ql_log_info, vha, 0x0037,
	    "MSI-X Falling back-to MSI mode -%d.\n", ret);
2947
skip_msix:
2948

2949
	if (!IS_QLA24XX(ha) && !IS_QLA2532(ha) && !IS_QLA8432(ha) &&
2950
	    !IS_QLA8001(ha) && !IS_QLA82XX(ha) && !IS_QLAFX00(ha))
2951 2952 2953 2954
		goto skip_msi;

	ret = pci_enable_msi(ha->pdev);
	if (!ret) {
2955 2956
		ql_dbg(ql_dbg_init, vha, 0x0038,
		    "MSI: Enabled.\n");
2957
		ha->flags.msi_enabled = 1;
2958
	} else
2959 2960
		ql_log(ql_log_warn, vha, 0x0039,
		    "MSI-X; Falling back-to INTa mode -- %d.\n", ret);
2961 2962 2963 2964 2965

	/* Skip INTx on ISP82xx. */
	if (!ha->flags.msi_enabled && IS_QLA82XX(ha))
		return QLA_FUNCTION_FAILED;

2966 2967
skip_msi:

2968
	ret = request_irq(ha->pdev->irq, ha->isp_ops->intr_handler,
2969 2970
	    ha->flags.msi_enabled ? 0 : IRQF_SHARED,
	    QLA2XXX_DRIVER_NAME, rsp);
2971
	if (ret) {
2972
		ql_log(ql_log_warn, vha, 0x003a,
2973 2974
		    "Failed to reserve interrupt %d already in use.\n",
		    ha->pdev->irq);
2975
		goto fail;
2976
	} else if (!ha->flags.msi_enabled) {
2977 2978
		ql_dbg(ql_dbg_init, vha, 0x0125,
		    "INTa mode: Enabled.\n");
2979 2980
		ha->flags.mr_intr_valid = 1;
	}
2981

2982 2983
clear_risc_ints:

2984
	spin_lock_irq(&ha->hardware_lock);
2985
	if (!IS_FWI2_CAPABLE(ha))
2986
		WRT_REG_WORD(&reg->isp.semaphore, 0);
2987
	spin_unlock_irq(&ha->hardware_lock);
2988

2989
fail:
2990 2991 2992 2993
	return ret;
}

void
2994
qla2x00_free_irqs(scsi_qla_host_t *vha)
2995
{
2996
	struct qla_hw_data *ha = vha->hw;
2997 2998 2999 3000 3001 3002 3003 3004 3005
	struct rsp_que *rsp;

	/*
	 * We need to check that ha->rsp_q_map is valid in case we are called
	 * from a probe failure context.
	 */
	if (!ha->rsp_q_map || !ha->rsp_q_map[0])
		return;
	rsp = ha->rsp_q_map[0];
3006 3007 3008

	if (ha->flags.msix_enabled)
		qla24xx_disable_msix(ha);
3009
	else if (ha->flags.msi_enabled) {
3010
		free_irq(ha->pdev->irq, rsp);
3011
		pci_disable_msi(ha->pdev);
3012 3013
	} else
		free_irq(ha->pdev->irq, rsp);
3014
}
3015

3016 3017 3018 3019

int qla25xx_request_irq(struct rsp_que *rsp)
{
	struct qla_hw_data *ha = rsp->hw;
3020
	struct qla_init_msix_entry *intr = &msix_entries[2];
3021
	struct qla_msix_entry *msix = rsp->msix;
3022
	scsi_qla_host_t *vha = pci_get_drvdata(ha->pdev);
3023 3024 3025 3026
	int ret;

	ret = request_irq(msix->vector, intr->handler, 0, intr->name, rsp);
	if (ret) {
3027 3028 3029
		ql_log(ql_log_fatal, vha, 0x00e6,
		    "MSI-X: Unable to register handler -- %x/%d.\n",
		    msix->vector, ret);
3030 3031 3032 3033 3034 3035
		return ret;
	}
	msix->have_irq = 1;
	msix->rsp = rsp;
	return ret;
}