msi.c 33.8 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8
/*
 * File:	msi.c
 * Purpose:	PCI Message Signaled Interrupt (MSI)
 *
 * Copyright (C) 2003-2004 Intel
 * Copyright (C) Tom Long Nguyen (tom.l.nguyen@intel.com)
 */

9
#include <linux/err.h>
L
Linus Torvalds 已提交
10 11 12
#include <linux/mm.h>
#include <linux/irq.h>
#include <linux/interrupt.h>
13
#include <linux/export.h>
L
Linus Torvalds 已提交
14 15 16
#include <linux/ioport.h>
#include <linux/pci.h>
#include <linux/proc_fs.h>
17
#include <linux/msi.h>
D
Dan Williams 已提交
18
#include <linux/smp.h>
H
Hidetoshi Seto 已提交
19 20
#include <linux/errno.h>
#include <linux/io.h>
21
#include <linux/slab.h>
22
#include <linux/irqdomain.h>
L
Linus Torvalds 已提交
23 24 25 26

#include "pci.h"

static int pci_msi_enable = 1;
27
int pci_msi_ignore_mask;
L
Linus Torvalds 已提交
28

29 30
#define msix_table_size(flags)	((flags & PCI_MSIX_FLAGS_QSIZE) + 1)

31 32 33 34 35 36 37 38 39
#ifdef CONFIG_PCI_MSI_IRQ_DOMAIN
static struct irq_domain *pci_msi_default_domain;
static DEFINE_MUTEX(pci_msi_domain_lock);

struct irq_domain * __weak arch_get_pci_msi_domain(struct pci_dev *dev)
{
	return pci_msi_default_domain;
}

40 41
static struct irq_domain *pci_msi_get_domain(struct pci_dev *dev)
{
42
	struct irq_domain *domain;
43

44 45 46
	domain = dev_get_msi_domain(&dev->dev);
	if (domain)
		return domain;
47

48
	return arch_get_pci_msi_domain(dev);
49 50
}

51 52 53 54
static int pci_msi_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
{
	struct irq_domain *domain;

55
	domain = pci_msi_get_domain(dev);
56 57 58 59 60 61 62 63 64 65
	if (domain)
		return pci_msi_domain_alloc_irqs(domain, dev, nvec, type);

	return arch_setup_msi_irqs(dev, nvec, type);
}

static void pci_msi_teardown_msi_irqs(struct pci_dev *dev)
{
	struct irq_domain *domain;

66
	domain = pci_msi_get_domain(dev);
67 68 69 70 71 72 73 74 75
	if (domain)
		pci_msi_domain_free_irqs(domain, dev);
	else
		arch_teardown_msi_irqs(dev);
}
#else
#define pci_msi_setup_msi_irqs		arch_setup_msi_irqs
#define pci_msi_teardown_msi_irqs	arch_teardown_msi_irqs
#endif
76

77 78
/* Arch hooks */

79 80
int __weak arch_setup_msi_irq(struct pci_dev *dev, struct msi_desc *desc)
{
81
	struct msi_controller *chip = dev->bus->msi;
82 83 84 85 86 87 88 89 90 91 92 93
	int err;

	if (!chip || !chip->setup_irq)
		return -EINVAL;

	err = chip->setup_irq(chip, dev, desc);
	if (err < 0)
		return err;

	irq_set_chip_data(desc->irq, chip);

	return 0;
94 95 96
}

void __weak arch_teardown_msi_irq(unsigned int irq)
97
{
98
	struct msi_controller *chip = irq_get_chip_data(irq);
99 100 101 102 103

	if (!chip || !chip->teardown_irq)
		return;

	chip->teardown_irq(chip, irq);
104 105
}

106
int __weak arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
107 108 109 110
{
	struct msi_desc *entry;
	int ret;

111 112 113 114 115 116 117
	/*
	 * If an architecture wants to support multiple MSI, it needs to
	 * override arch_setup_msi_irqs()
	 */
	if (type == PCI_CAP_ID_MSI && nvec > 1)
		return 1;

118
	for_each_pci_msi_entry(entry, dev) {
119
		ret = arch_setup_msi_irq(dev, entry);
120
		if (ret < 0)
121
			return ret;
122 123
		if (ret > 0)
			return -ENOSPC;
124 125 126 127
	}

	return 0;
}
128

129 130 131 132
/*
 * We have a default implementation available as a separate non-weak
 * function, as it is used by the Xen x86 PCI code
 */
133
void default_teardown_msi_irqs(struct pci_dev *dev)
134
{
135
	int i;
136 137
	struct msi_desc *entry;

138
	for_each_pci_msi_entry(entry, dev)
139 140 141
		if (entry->irq)
			for (i = 0; i < entry->nvec_used; i++)
				arch_teardown_msi_irq(entry->irq + i);
142 143
}

144 145 146 147
void __weak arch_teardown_msi_irqs(struct pci_dev *dev)
{
	return default_teardown_msi_irqs(dev);
}
148

149
static void default_restore_msi_irq(struct pci_dev *dev, int irq)
150 151 152 153 154
{
	struct msi_desc *entry;

	entry = NULL;
	if (dev->msix_enabled) {
155
		for_each_pci_msi_entry(entry, dev) {
156 157 158 159 160 161 162 163
			if (irq == entry->irq)
				break;
		}
	} else if (dev->msi_enabled)  {
		entry = irq_get_msi_desc(irq);
	}

	if (entry)
164
		__pci_write_msi_msg(entry, &entry->msg);
165
}
166

167
void __weak arch_restore_msi_irqs(struct pci_dev *dev)
168
{
169
	return default_restore_msi_irqs(dev);
170
}
171

172 173
static inline __attribute_const__ u32 msi_mask(unsigned x)
{
174 175 176 177
	/* Don't shift by >= width of type */
	if (x >= 5)
		return 0xffffffff;
	return (1 << (1 << x)) - 1;
178 179
}

180 181 182 183 184 185
/*
 * PCI 2.3 does not specify mask bits for each MSI interrupt.  Attempting to
 * mask all MSI interrupts by clearing the MSI enable bit does not work
 * reliably as devices without an INTx disable bit will then generate a
 * level IRQ which will never be cleared.
 */
186
u32 __pci_msi_desc_mask_irq(struct msi_desc *desc, u32 mask, u32 flag)
L
Linus Torvalds 已提交
187
{
188
	u32 mask_bits = desc->masked;
L
Linus Torvalds 已提交
189

190
	if (pci_msi_ignore_mask || !desc->msi_attrib.maskbit)
191
		return 0;
192 193 194

	mask_bits &= ~mask;
	mask_bits |= flag;
195 196
	pci_write_config_dword(msi_desc_to_pci_dev(desc), desc->mask_pos,
			       mask_bits);
197 198 199 200 201 202

	return mask_bits;
}

static void msi_mask_irq(struct msi_desc *desc, u32 mask, u32 flag)
{
203
	desc->masked = __pci_msi_desc_mask_irq(desc, mask, flag);
204 205 206 207 208 209 210 211 212
}

/*
 * This internal function does not flush PCI writes to the device.
 * All users must ensure that they read from the device before either
 * assuming that the device state is up to date, or returning out of this
 * file.  This saves a few milliseconds when initialising devices with lots
 * of MSI-X interrupts.
 */
213
u32 __pci_msix_desc_mask_irq(struct msi_desc *desc, u32 flag)
214 215 216
{
	u32 mask_bits = desc->masked;
	unsigned offset = desc->msi_attrib.entry_nr * PCI_MSIX_ENTRY_SIZE +
217
						PCI_MSIX_ENTRY_VECTOR_CTRL;
218 219 220 221

	if (pci_msi_ignore_mask)
		return 0;

222 223 224
	mask_bits &= ~PCI_MSIX_ENTRY_CTRL_MASKBIT;
	if (flag)
		mask_bits |= PCI_MSIX_ENTRY_CTRL_MASKBIT;
225
	writel(mask_bits, desc->mask_base + offset);
226 227 228 229 230 231

	return mask_bits;
}

static void msix_mask_irq(struct msi_desc *desc, u32 flag)
{
232
	desc->masked = __pci_msix_desc_mask_irq(desc, flag);
233
}
234

235
static void msi_set_mask_bit(struct irq_data *data, u32 flag)
236
{
237
	struct msi_desc *desc = irq_data_get_msi_desc(data);
238

239 240 241 242
	if (desc->msi_attrib.is_msix) {
		msix_mask_irq(desc, flag);
		readl(desc->mask_base);		/* Flush write to device */
	} else {
243
		unsigned offset = data->irq - desc->irq;
244
		msi_mask_irq(desc, 1 << offset, flag << offset);
L
Linus Torvalds 已提交
245
	}
246 247
}

248 249 250 251 252
/**
 * pci_msi_mask_irq - Generic irq chip callback to mask PCI/MSI interrupts
 * @data:	pointer to irqdata associated to that interrupt
 */
void pci_msi_mask_irq(struct irq_data *data)
253
{
254
	msi_set_mask_bit(data, 1);
255 256
}

257 258 259 260 261
/**
 * pci_msi_unmask_irq - Generic irq chip callback to unmask PCI/MSI interrupts
 * @data:	pointer to irqdata associated to that interrupt
 */
void pci_msi_unmask_irq(struct irq_data *data)
262
{
263
	msi_set_mask_bit(data, 0);
L
Linus Torvalds 已提交
264 265
}

266 267 268 269
void default_restore_msi_irqs(struct pci_dev *dev)
{
	struct msi_desc *entry;

270
	for_each_pci_msi_entry(entry, dev)
271 272 273
		default_restore_msi_irq(dev, entry->irq);
}

274
void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg)
L
Linus Torvalds 已提交
275
{
276 277 278
	struct pci_dev *dev = msi_desc_to_pci_dev(entry);

	BUG_ON(dev->current_state != PCI_D0);
279 280 281 282 283 284 285 286 287

	if (entry->msi_attrib.is_msix) {
		void __iomem *base = entry->mask_base +
			entry->msi_attrib.entry_nr * PCI_MSIX_ENTRY_SIZE;

		msg->address_lo = readl(base + PCI_MSIX_ENTRY_LOWER_ADDR);
		msg->address_hi = readl(base + PCI_MSIX_ENTRY_UPPER_ADDR);
		msg->data = readl(base + PCI_MSIX_ENTRY_DATA);
	} else {
288
		int pos = dev->msi_cap;
289 290
		u16 data;

291 292
		pci_read_config_dword(dev, pos + PCI_MSI_ADDRESS_LO,
				      &msg->address_lo);
293
		if (entry->msi_attrib.is_64) {
294 295
			pci_read_config_dword(dev, pos + PCI_MSI_ADDRESS_HI,
					      &msg->address_hi);
B
Bjorn Helgaas 已提交
296
			pci_read_config_word(dev, pos + PCI_MSI_DATA_64, &data);
297 298
		} else {
			msg->address_hi = 0;
B
Bjorn Helgaas 已提交
299
			pci_read_config_word(dev, pos + PCI_MSI_DATA_32, &data);
300 301 302 303 304
		}
		msg->data = data;
	}
}

305
void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg)
Y
Yinghai Lu 已提交
306
{
307 308 309
	struct pci_dev *dev = msi_desc_to_pci_dev(entry);

	if (dev->current_state != PCI_D0) {
310 311
		/* Don't touch the hardware now */
	} else if (entry->msi_attrib.is_msix) {
312 313 314 315
		void __iomem *base;
		base = entry->mask_base +
			entry->msi_attrib.entry_nr * PCI_MSIX_ENTRY_SIZE;

316 317 318
		writel(msg->address_lo, base + PCI_MSIX_ENTRY_LOWER_ADDR);
		writel(msg->address_hi, base + PCI_MSIX_ENTRY_UPPER_ADDR);
		writel(msg->data, base + PCI_MSIX_ENTRY_DATA);
319
	} else {
320
		int pos = dev->msi_cap;
321 322
		u16 msgctl;

323
		pci_read_config_word(dev, pos + PCI_MSI_FLAGS, &msgctl);
324 325
		msgctl &= ~PCI_MSI_FLAGS_QSIZE;
		msgctl |= entry->msi_attrib.multiple << 4;
326
		pci_write_config_word(dev, pos + PCI_MSI_FLAGS, msgctl);
327

328 329
		pci_write_config_dword(dev, pos + PCI_MSI_ADDRESS_LO,
				       msg->address_lo);
330
		if (entry->msi_attrib.is_64) {
331 332
			pci_write_config_dword(dev, pos + PCI_MSI_ADDRESS_HI,
					       msg->address_hi);
B
Bjorn Helgaas 已提交
333 334
			pci_write_config_word(dev, pos + PCI_MSI_DATA_64,
					      msg->data);
335
		} else {
B
Bjorn Helgaas 已提交
336 337
			pci_write_config_word(dev, pos + PCI_MSI_DATA_32,
					      msg->data);
338
		}
L
Linus Torvalds 已提交
339
	}
340
	entry->msg = *msg;
L
Linus Torvalds 已提交
341
}
342

343
void pci_write_msi_msg(unsigned int irq, struct msi_msg *msg)
Y
Yinghai Lu 已提交
344
{
345
	struct msi_desc *entry = irq_get_msi_desc(irq);
Y
Yinghai Lu 已提交
346

347
	__pci_write_msi_msg(entry, msg);
Y
Yinghai Lu 已提交
348
}
349
EXPORT_SYMBOL_GPL(pci_write_msi_msg);
Y
Yinghai Lu 已提交
350

351 352
static void free_msi_irqs(struct pci_dev *dev)
{
353
	struct list_head *msi_list = dev_to_msi_list(&dev->dev);
354
	struct msi_desc *entry, *tmp;
355 356
	struct attribute **msi_attrs;
	struct device_attribute *dev_attr;
357
	int i, count = 0;
358

359
	for_each_pci_msi_entry(entry, dev)
360 361 362
		if (entry->irq)
			for (i = 0; i < entry->nvec_used; i++)
				BUG_ON(irq_has_action(entry->irq + i));
363

364
	pci_msi_teardown_msi_irqs(dev);
365

366
	list_for_each_entry_safe(entry, tmp, msi_list, list) {
367
		if (entry->msi_attrib.is_msix) {
368
			if (list_is_last(&entry->list, msi_list))
369 370
				iounmap(entry->mask_base);
		}
371

372 373 374
		list_del(&entry->list);
		kfree(entry);
	}
375 376 377 378

	if (dev->msi_irq_groups) {
		sysfs_remove_groups(&dev->dev.kobj, dev->msi_irq_groups);
		msi_attrs = dev->msi_irq_groups[0]->attrs;
379
		while (msi_attrs[count]) {
380 381 382 383 384 385 386 387 388 389 390
			dev_attr = container_of(msi_attrs[count],
						struct device_attribute, attr);
			kfree(dev_attr->attr.name);
			kfree(dev_attr);
			++count;
		}
		kfree(msi_attrs);
		kfree(dev->msi_irq_groups[0]);
		kfree(dev->msi_irq_groups);
		dev->msi_irq_groups = NULL;
	}
391
}
S
Satoru Takeuchi 已提交
392

393 394 395 396 397 398
static void pci_intx_for_msi(struct pci_dev *dev, int enable)
{
	if (!(dev->dev_flags & PCI_DEV_FLAGS_MSI_INTX_DISABLE_BUG))
		pci_intx(dev, enable);
}

399
static void __pci_restore_msi_state(struct pci_dev *dev)
400 401
{
	u16 control;
402
	struct msi_desc *entry;
403

404 405 406
	if (!dev->msi_enabled)
		return;

407
	entry = irq_get_msi_desc(dev->irq);
408

409
	pci_intx_for_msi(dev, 0);
410
	pci_msi_set_enable(dev, 0);
411
	arch_restore_msi_irqs(dev);
412

413
	pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &control);
414 415
	msi_mask_irq(entry, msi_mask(entry->msi_attrib.multi_cap),
		     entry->masked);
416
	control &= ~PCI_MSI_FLAGS_QSIZE;
417
	control |= (entry->msi_attrib.multiple << 4) | PCI_MSI_FLAGS_ENABLE;
418
	pci_write_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, control);
419 420 421
}

static void __pci_restore_msix_state(struct pci_dev *dev)
422 423 424
{
	struct msi_desc *entry;

E
Eric W. Biederman 已提交
425 426
	if (!dev->msix_enabled)
		return;
427
	BUG_ON(list_empty(dev_to_msi_list(&dev->dev)));
E
Eric W. Biederman 已提交
428

429
	/* route the table */
430
	pci_intx_for_msi(dev, 0);
431
	pci_msix_clear_and_set_ctrl(dev, 0,
432
				PCI_MSIX_FLAGS_ENABLE | PCI_MSIX_FLAGS_MASKALL);
433

434
	arch_restore_msi_irqs(dev);
435
	for_each_pci_msi_entry(entry, dev)
436
		msix_mask_irq(entry, entry->masked);
437

438
	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_MASKALL, 0);
439
}
440 441 442 443 444 445

void pci_restore_msi_state(struct pci_dev *dev)
{
	__pci_restore_msi_state(dev);
	__pci_restore_msix_state(dev);
}
446
EXPORT_SYMBOL_GPL(pci_restore_msi_state);
447

448
static ssize_t msi_mode_show(struct device *dev, struct device_attribute *attr,
449 450
			     char *buf)
{
451 452 453
	struct msi_desc *entry;
	unsigned long irq;
	int retval;
454

455 456 457
	retval = kstrtoul(attr->attr.name, 10, &irq);
	if (retval)
		return retval;
458

459 460 461 462 463
	entry = irq_get_msi_desc(irq);
	if (entry)
		return sprintf(buf, "%s\n",
				entry->msi_attrib.is_msix ? "msix" : "msi");

464
	return -ENODEV;
465 466 467 468
}

static int populate_msi_sysfs(struct pci_dev *pdev)
{
469 470 471 472 473
	struct attribute **msi_attrs;
	struct attribute *msi_attr;
	struct device_attribute *msi_dev_attr;
	struct attribute_group *msi_irq_group;
	const struct attribute_group **msi_irq_groups;
474
	struct msi_desc *entry;
475 476
	int ret = -ENOMEM;
	int num_msi = 0;
477 478
	int count = 0;

479
	/* Determine how many msi entries we have */
480
	for_each_pci_msi_entry(entry, pdev)
481 482 483
		++num_msi;
	if (!num_msi)
		return 0;
484

485 486 487 488
	/* Dynamically create the MSI attributes for the PCI device */
	msi_attrs = kzalloc(sizeof(void *) * (num_msi + 1), GFP_KERNEL);
	if (!msi_attrs)
		return -ENOMEM;
489
	for_each_pci_msi_entry(entry, pdev) {
490
		msi_dev_attr = kzalloc(sizeof(*msi_dev_attr), GFP_KERNEL);
491
		if (!msi_dev_attr)
492
			goto error_attrs;
493
		msi_attrs[count] = &msi_dev_attr->attr;
494

495
		sysfs_attr_init(&msi_dev_attr->attr);
496 497 498 499
		msi_dev_attr->attr.name = kasprintf(GFP_KERNEL, "%d",
						    entry->irq);
		if (!msi_dev_attr->attr.name)
			goto error_attrs;
500 501 502
		msi_dev_attr->attr.mode = S_IRUGO;
		msi_dev_attr->show = msi_mode_show;
		++count;
503 504
	}

505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520
	msi_irq_group = kzalloc(sizeof(*msi_irq_group), GFP_KERNEL);
	if (!msi_irq_group)
		goto error_attrs;
	msi_irq_group->name = "msi_irqs";
	msi_irq_group->attrs = msi_attrs;

	msi_irq_groups = kzalloc(sizeof(void *) * 2, GFP_KERNEL);
	if (!msi_irq_groups)
		goto error_irq_group;
	msi_irq_groups[0] = msi_irq_group;

	ret = sysfs_create_groups(&pdev->dev.kobj, msi_irq_groups);
	if (ret)
		goto error_irq_groups;
	pdev->msi_irq_groups = msi_irq_groups;

521 522
	return 0;

523 524 525 526 527 528 529 530 531 532 533 534 535
error_irq_groups:
	kfree(msi_irq_groups);
error_irq_group:
	kfree(msi_irq_group);
error_attrs:
	count = 0;
	msi_attr = msi_attrs[count];
	while (msi_attr) {
		msi_dev_attr = container_of(msi_attr, struct device_attribute, attr);
		kfree(msi_attr->name);
		kfree(msi_dev_attr);
		++count;
		msi_attr = msi_attrs[count];
536
	}
537
	kfree(msi_attrs);
538 539 540
	return ret;
}

541
static struct msi_desc *msi_setup_entry(struct pci_dev *dev, int nvec)
542 543 544 545 546
{
	u16 control;
	struct msi_desc *entry;

	/* MSI Entry Initialization */
547
	entry = alloc_msi_entry(&dev->dev);
548 549 550 551 552 553 554 555 556 557 558
	if (!entry)
		return NULL;

	pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &control);

	entry->msi_attrib.is_msix	= 0;
	entry->msi_attrib.is_64		= !!(control & PCI_MSI_FLAGS_64BIT);
	entry->msi_attrib.entry_nr	= 0;
	entry->msi_attrib.maskbit	= !!(control & PCI_MSI_FLAGS_MASKBIT);
	entry->msi_attrib.default_irq	= dev->irq;	/* Save IOAPIC IRQ */
	entry->msi_attrib.multi_cap	= (control & PCI_MSI_FLAGS_QMASK) >> 1;
559 560
	entry->msi_attrib.multiple	= ilog2(__roundup_pow_of_two(nvec));
	entry->nvec_used		= nvec;
561 562 563 564 565 566 567 568 569 570 571 572 573

	if (control & PCI_MSI_FLAGS_64BIT)
		entry->mask_pos = dev->msi_cap + PCI_MSI_MASK_64;
	else
		entry->mask_pos = dev->msi_cap + PCI_MSI_MASK_32;

	/* Save the initial mask status */
	if (entry->msi_attrib.maskbit)
		pci_read_config_dword(dev, entry->mask_pos, &entry->masked);

	return entry;
}

574 575 576 577
static int msi_verify_entries(struct pci_dev *dev)
{
	struct msi_desc *entry;

578
	for_each_pci_msi_entry(entry, dev) {
579 580 581 582 583 584 585 586 587
		if (!dev->no_64bit_msi || !entry->msg.address_hi)
			continue;
		dev_err(&dev->dev, "Device has broken 64-bit MSI but arch"
			" tried to assign one above 4G\n");
		return -EIO;
	}
	return 0;
}

L
Linus Torvalds 已提交
588 589 590
/**
 * msi_capability_init - configure device's MSI capability structure
 * @dev: pointer to the pci_dev data structure of MSI device function
591
 * @nvec: number of interrupts to allocate
L
Linus Torvalds 已提交
592
 *
593 594 595 596 597 598 599
 * Setup the MSI capability structure of the device with the requested
 * number of interrupts.  A return value of zero indicates the successful
 * setup of an entry with the new MSI irq.  A negative return value indicates
 * an error, and a positive return value indicates the number of interrupts
 * which could have been allocated.
 */
static int msi_capability_init(struct pci_dev *dev, int nvec)
L
Linus Torvalds 已提交
600 601
{
	struct msi_desc *entry;
602
	int ret;
603
	unsigned mask;
L
Linus Torvalds 已提交
604

605
	pci_msi_set_enable(dev, 0);	/* Disable MSI during set up */
606

607
	entry = msi_setup_entry(dev, nvec);
608 609
	if (!entry)
		return -ENOMEM;
610

611
	/* All MSIs are unmasked by default, Mask them all */
612
	mask = msi_mask(entry->msi_attrib.multi_cap);
613 614
	msi_mask_irq(entry, mask, mask);

615
	list_add_tail(&entry->list, dev_to_msi_list(&dev->dev));
616

L
Linus Torvalds 已提交
617
	/* Configure MSI capability structure */
618
	ret = pci_msi_setup_msi_irqs(dev, nvec, PCI_CAP_ID_MSI);
619
	if (ret) {
620
		msi_mask_irq(entry, mask, ~mask);
621
		free_msi_irqs(dev);
622
		return ret;
623
	}
624

625 626 627 628 629 630 631
	ret = msi_verify_entries(dev);
	if (ret) {
		msi_mask_irq(entry, mask, ~mask);
		free_msi_irqs(dev);
		return ret;
	}

632 633 634 635 636 637 638
	ret = populate_msi_sysfs(dev);
	if (ret) {
		msi_mask_irq(entry, mask, ~mask);
		free_msi_irqs(dev);
		return ret;
	}

L
Linus Torvalds 已提交
639
	/* Set MSI enabled bits	 */
640
	pci_intx_for_msi(dev, 0);
641
	pci_msi_set_enable(dev, 1);
642
	dev->msi_enabled = 1;
L
Linus Torvalds 已提交
643

644
	pcibios_free_irq(dev);
645
	dev->irq = entry->irq;
L
Linus Torvalds 已提交
646 647 648
	return 0;
}

649
static void __iomem *msix_map_region(struct pci_dev *dev, unsigned nr_entries)
650
{
651
	resource_size_t phys_addr;
652
	u32 table_offset;
653
	unsigned long flags;
654 655
	u8 bir;

656 657
	pci_read_config_dword(dev, dev->msix_cap + PCI_MSIX_TABLE,
			      &table_offset);
658
	bir = (u8)(table_offset & PCI_MSIX_TABLE_BIR);
659 660 661 662
	flags = pci_resource_flags(dev, bir);
	if (!flags || (flags & IORESOURCE_UNSET))
		return NULL;

663
	table_offset &= PCI_MSIX_TABLE_OFFSET;
664 665 666 667 668
	phys_addr = pci_resource_start(dev, bir) + table_offset;

	return ioremap_nocache(phys_addr, nr_entries * PCI_MSIX_ENTRY_SIZE);
}

669 670
static int msix_setup_entries(struct pci_dev *dev, void __iomem *base,
			      struct msix_entry *entries, int nvec)
671 672 673 674 675
{
	struct msi_desc *entry;
	int i;

	for (i = 0; i < nvec; i++) {
676
		entry = alloc_msi_entry(&dev->dev);
677 678 679 680 681 682 683 684 685 686 687 688 689 690
		if (!entry) {
			if (!i)
				iounmap(base);
			else
				free_msi_irqs(dev);
			/* No enough memory. Don't try again */
			return -ENOMEM;
		}

		entry->msi_attrib.is_msix	= 1;
		entry->msi_attrib.is_64		= 1;
		entry->msi_attrib.entry_nr	= entries[i].entry;
		entry->msi_attrib.default_irq	= dev->irq;
		entry->mask_base		= base;
691
		entry->nvec_used		= 1;
692

693
		list_add_tail(&entry->list, dev_to_msi_list(&dev->dev));
694 695 696 697 698
	}

	return 0;
}

699
static void msix_program_entries(struct pci_dev *dev,
700
				 struct msix_entry *entries)
701 702 703 704
{
	struct msi_desc *entry;
	int i = 0;

705
	for_each_pci_msi_entry(entry, dev) {
706 707 708 709 710 711 712 713 714 715
		int offset = entries[i].entry * PCI_MSIX_ENTRY_SIZE +
						PCI_MSIX_ENTRY_VECTOR_CTRL;

		entries[i].vector = entry->irq;
		entry->masked = readl(entry->mask_base + offset);
		msix_mask_irq(entry, 1);
		i++;
	}
}

L
Linus Torvalds 已提交
716 717 718
/**
 * msix_capability_init - configure device's MSI-X capability
 * @dev: pointer to the pci_dev data structure of MSI-X device function
R
Randy Dunlap 已提交
719 720
 * @entries: pointer to an array of struct msix_entry entries
 * @nvec: number of @entries
L
Linus Torvalds 已提交
721
 *
722
 * Setup the MSI-X capability structure of device function with a
723 724
 * single MSI-X irq. A return of zero indicates the successful setup of
 * requested MSI-X entries with allocated irqs or non-zero for otherwise.
L
Linus Torvalds 已提交
725 726 727 728
 **/
static int msix_capability_init(struct pci_dev *dev,
				struct msix_entry *entries, int nvec)
{
729
	int ret;
730
	u16 control;
L
Linus Torvalds 已提交
731 732
	void __iomem *base;

733
	/* Ensure MSI-X is disabled while it is set up */
734
	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_ENABLE, 0);
735

736
	pci_read_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, &control);
L
Linus Torvalds 已提交
737
	/* Request & Map MSI-X table region */
738
	base = msix_map_region(dev, msix_table_size(control));
739
	if (!base)
L
Linus Torvalds 已提交
740 741
		return -ENOMEM;

742
	ret = msix_setup_entries(dev, base, entries, nvec);
743 744
	if (ret)
		return ret;
745

746
	ret = pci_msi_setup_msi_irqs(dev, nvec, PCI_CAP_ID_MSIX);
747
	if (ret)
748
		goto out_avail;
749

750 751 752 753 754
	/* Check if all MSI entries honor device restrictions */
	ret = msi_verify_entries(dev);
	if (ret)
		goto out_free;

755 756 757 758 759
	/*
	 * Some devices require MSI-X to be enabled before we can touch the
	 * MSI-X registers.  We need to mask all the vectors to prevent
	 * interrupts coming in before they're fully set up.
	 */
760
	pci_msix_clear_and_set_ctrl(dev, 0,
761
				PCI_MSIX_FLAGS_MASKALL | PCI_MSIX_FLAGS_ENABLE);
762

763
	msix_program_entries(dev, entries);
764

765
	ret = populate_msi_sysfs(dev);
766 767
	if (ret)
		goto out_free;
768

769
	/* Set MSI-X enabled bits and unmask the function */
770
	pci_intx_for_msi(dev, 0);
771
	dev->msix_enabled = 1;
772
	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_MASKALL, 0);
773

774
	pcibios_free_irq(dev);
L
Linus Torvalds 已提交
775
	return 0;
776

777
out_avail:
778 779 780 781 782
	if (ret < 0) {
		/*
		 * If we had some success, report the number of irqs
		 * we succeeded in setting up.
		 */
783
		struct msi_desc *entry;
784 785
		int avail = 0;

786
		for_each_pci_msi_entry(entry, dev) {
787 788 789 790 791 792 793
			if (entry->irq != 0)
				avail++;
		}
		if (avail != 0)
			ret = avail;
	}

794
out_free:
795 796 797
	free_msi_irqs(dev);

	return ret;
L
Linus Torvalds 已提交
798 799
}

800
/**
801
 * pci_msi_supported - check whether MSI may be enabled on a device
802
 * @dev: pointer to the pci_dev data structure of MSI device function
803
 * @nvec: how many MSIs have been requested ?
804
 *
805
 * Look at global flags, the device itself, and its parent buses
806
 * to determine if MSI/-X are supported for the device. If MSI/-X is
807
 * supported return 1, else return 0.
808
 **/
809
static int pci_msi_supported(struct pci_dev *dev, int nvec)
810 811 812
{
	struct pci_bus *bus;

813
	/* MSI must be globally enabled and supported by the device */
814
	if (!pci_msi_enable)
815
		return 0;
816 817

	if (!dev || dev->no_msi || dev->current_state != PCI_D0)
818
		return 0;
819

820 821 822 823 824 825
	/*
	 * You can't ask to have 0 or less MSIs configured.
	 *  a) it's stupid ..
	 *  b) the list manipulation code assumes nvec >= 1.
	 */
	if (nvec < 1)
826
		return 0;
827

H
Hidetoshi Seto 已提交
828 829 830
	/*
	 * Any bridge which does NOT route MSI transactions from its
	 * secondary bus to its primary bus must set NO_MSI flag on
831 832 833 834
	 * the secondary pci_bus.
	 * We expect only arch-specific PCI host bus controller driver
	 * or quirks for specific PCI bridges to be setting NO_MSI.
	 */
835 836
	for (bus = dev->bus; bus; bus = bus->parent)
		if (bus->bus_flags & PCI_BUS_FLAGS_NO_MSI)
837
			return 0;
838

839
	return 1;
840 841
}

842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866
/**
 * pci_msi_vec_count - Return the number of MSI vectors a device can send
 * @dev: device to report about
 *
 * This function returns the number of MSI vectors a device requested via
 * Multiple Message Capable register. It returns a negative errno if the
 * device is not capable sending MSI interrupts. Otherwise, the call succeeds
 * and returns a power of two, up to a maximum of 2^5 (32), according to the
 * MSI specification.
 **/
int pci_msi_vec_count(struct pci_dev *dev)
{
	int ret;
	u16 msgctl;

	if (!dev->msi_cap)
		return -EINVAL;

	pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &msgctl);
	ret = 1 << ((msgctl & PCI_MSI_FLAGS_QMASK) >> 1);

	return ret;
}
EXPORT_SYMBOL(pci_msi_vec_count);

867
void pci_msi_shutdown(struct pci_dev *dev)
L
Linus Torvalds 已提交
868
{
869 870
	struct msi_desc *desc;
	u32 mask;
L
Linus Torvalds 已提交
871

872
	if (!pci_msi_enable || !dev || !dev->msi_enabled)
E
Eric W. Biederman 已提交
873 874
		return;

875
	BUG_ON(list_empty(dev_to_msi_list(&dev->dev)));
876
	desc = first_pci_msi_entry(dev);
877

878
	pci_msi_set_enable(dev, 0);
879
	pci_intx_for_msi(dev, 1);
880
	dev->msi_enabled = 0;
881

882
	/* Return the device with MSI unmasked as initial states */
883
	mask = msi_mask(desc->msi_attrib.multi_cap);
884
	/* Keep cached state to be restored */
885
	__pci_msi_desc_mask_irq(desc, mask, ~mask);
886 887

	/* Restore dev->irq to its default pin-assertion irq */
888
	dev->irq = desc->msi_attrib.default_irq;
889
	pcibios_alloc_irq(dev);
890
}
891

H
Hidetoshi Seto 已提交
892
void pci_disable_msi(struct pci_dev *dev)
893 894 895 896 897
{
	if (!pci_msi_enable || !dev || !dev->msi_enabled)
		return;

	pci_msi_shutdown(dev);
898
	free_msi_irqs(dev);
L
Linus Torvalds 已提交
899
}
900
EXPORT_SYMBOL(pci_disable_msi);
L
Linus Torvalds 已提交
901

902
/**
903
 * pci_msix_vec_count - return the number of device's MSI-X table entries
904
 * @dev: pointer to the pci_dev data structure of MSI-X device function
905 906 907 908 909 910
 * This function returns the number of device's MSI-X table entries and
 * therefore the number of MSI-X vectors device is capable of sending.
 * It returns a negative errno if the device is not capable of sending MSI-X
 * interrupts.
 **/
int pci_msix_vec_count(struct pci_dev *dev)
911 912 913
{
	u16 control;

914
	if (!dev->msix_cap)
915
		return -EINVAL;
916

917
	pci_read_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, &control);
918
	return msix_table_size(control);
919
}
920
EXPORT_SYMBOL(pci_msix_vec_count);
921

L
Linus Torvalds 已提交
922 923 924
/**
 * pci_enable_msix - configure device's MSI-X capability structure
 * @dev: pointer to the pci_dev data structure of MSI-X device function
925
 * @entries: pointer to an array of MSI-X entries
926
 * @nvec: number of MSI-X irqs requested for allocation by device driver
L
Linus Torvalds 已提交
927 928
 *
 * Setup the MSI-X capability structure of device function with the number
929
 * of requested irqs upon its software driver call to request for
L
Linus Torvalds 已提交
930 931
 * MSI-X mode enabled on its hardware device function. A return of zero
 * indicates the successful configuration of MSI-X capability structure
932
 * with new allocated MSI-X irqs. A return of < 0 indicates a failure.
L
Linus Torvalds 已提交
933
 * Or a return of > 0 indicates that driver request is exceeding the number
934 935
 * of irqs or MSI-X vectors available. Driver should use the returned value to
 * re-send its request.
L
Linus Torvalds 已提交
936
 **/
H
Hidetoshi Seto 已提交
937
int pci_enable_msix(struct pci_dev *dev, struct msix_entry *entries, int nvec)
L
Linus Torvalds 已提交
938
{
939
	int nr_entries;
E
Eric W. Biederman 已提交
940
	int i, j;
L
Linus Torvalds 已提交
941

942 943
	if (!pci_msi_supported(dev, nvec))
		return -EINVAL;
944

945 946 947
	if (!entries)
		return -EINVAL;

948 949 950
	nr_entries = pci_msix_vec_count(dev);
	if (nr_entries < 0)
		return nr_entries;
L
Linus Torvalds 已提交
951
	if (nvec > nr_entries)
952
		return nr_entries;
L
Linus Torvalds 已提交
953 954 955 956 957 958 959 960 961 962

	/* Check for any invalid entries */
	for (i = 0; i < nvec; i++) {
		if (entries[i].entry >= nr_entries)
			return -EINVAL;		/* invalid entry */
		for (j = i + 1; j < nvec; j++) {
			if (entries[i].entry == entries[j].entry)
				return -EINVAL;	/* duplicate entry */
		}
	}
E
Eric W. Biederman 已提交
963
	WARN_ON(!!dev->msix_enabled);
964

965
	/* Check whether driver already requested for MSI irq */
H
Hidetoshi Seto 已提交
966
	if (dev->msi_enabled) {
967
		dev_info(&dev->dev, "can't enable MSI-X (MSI IRQ already assigned)\n");
L
Linus Torvalds 已提交
968 969
		return -EINVAL;
	}
970
	return msix_capability_init(dev, entries, nvec);
L
Linus Torvalds 已提交
971
}
972
EXPORT_SYMBOL(pci_enable_msix);
L
Linus Torvalds 已提交
973

H
Hidetoshi Seto 已提交
974
void pci_msix_shutdown(struct pci_dev *dev)
975
{
976 977
	struct msi_desc *entry;

978
	if (!pci_msi_enable || !dev || !dev->msix_enabled)
E
Eric W. Biederman 已提交
979 980
		return;

981
	/* Return the device with MSI-X masked as initial states */
982
	for_each_pci_msi_entry(entry, dev) {
983
		/* Keep cached states to be restored */
984
		__pci_msix_desc_mask_irq(entry, 1);
985 986
	}

987
	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_ENABLE, 0);
988
	pci_intx_for_msi(dev, 1);
989
	dev->msix_enabled = 0;
990
	pcibios_alloc_irq(dev);
991
}
992

H
Hidetoshi Seto 已提交
993
void pci_disable_msix(struct pci_dev *dev)
994 995 996 997 998
{
	if (!pci_msi_enable || !dev || !dev->msix_enabled)
		return;

	pci_msix_shutdown(dev);
999
	free_msi_irqs(dev);
L
Linus Torvalds 已提交
1000
}
1001
EXPORT_SYMBOL(pci_disable_msix);
L
Linus Torvalds 已提交
1002

1003 1004 1005 1006
void pci_no_msi(void)
{
	pci_msi_enable = 0;
}
1007

1008 1009 1010 1011 1012 1013 1014
/**
 * pci_msi_enabled - is MSI enabled?
 *
 * Returns true if MSI has not been disabled by the command-line option
 * pci=nomsi.
 **/
int pci_msi_enabled(void)
1015
{
1016
	return pci_msi_enable;
1017
}
1018
EXPORT_SYMBOL(pci_msi_enabled);
1019

1020
void pci_msi_init_pci_dev(struct pci_dev *dev)
1021 1022
{
}
1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037

/**
 * pci_enable_msi_range - configure device's MSI capability structure
 * @dev: device to configure
 * @minvec: minimal number of interrupts to configure
 * @maxvec: maximum number of interrupts to configure
 *
 * This function tries to allocate a maximum possible number of interrupts in a
 * range between @minvec and @maxvec. It returns a negative errno if an error
 * occurs. If it succeeds, it returns the actual number of interrupts allocated
 * and updates the @dev's irq member to the lowest new interrupt number;
 * the other interrupt numbers allocated to this device are consecutive.
 **/
int pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec)
{
1038
	int nvec;
1039 1040
	int rc;

1041 1042
	if (!pci_msi_supported(dev, minvec))
		return -EINVAL;
1043 1044 1045 1046 1047 1048 1049 1050 1051 1052

	WARN_ON(!!dev->msi_enabled);

	/* Check whether driver already requested MSI-X irqs */
	if (dev->msix_enabled) {
		dev_info(&dev->dev,
			 "can't enable MSI (MSI-X already enabled)\n");
		return -EINVAL;
	}

1053 1054 1055
	if (maxvec < minvec)
		return -ERANGE;

1056 1057 1058 1059 1060 1061 1062 1063
	nvec = pci_msi_vec_count(dev);
	if (nvec < 0)
		return nvec;
	else if (nvec < minvec)
		return -EINVAL;
	else if (nvec > maxvec)
		nvec = maxvec;

1064
	do {
1065
		rc = msi_capability_init(dev, nvec);
1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116
		if (rc < 0) {
			return rc;
		} else if (rc > 0) {
			if (rc < minvec)
				return -ENOSPC;
			nvec = rc;
		}
	} while (rc);

	return nvec;
}
EXPORT_SYMBOL(pci_enable_msi_range);

/**
 * pci_enable_msix_range - configure device's MSI-X capability structure
 * @dev: pointer to the pci_dev data structure of MSI-X device function
 * @entries: pointer to an array of MSI-X entries
 * @minvec: minimum number of MSI-X irqs requested
 * @maxvec: maximum number of MSI-X irqs requested
 *
 * Setup the MSI-X capability structure of device function with a maximum
 * possible number of interrupts in the range between @minvec and @maxvec
 * upon its software driver call to request for MSI-X mode enabled on its
 * hardware device function. It returns a negative errno if an error occurs.
 * If it succeeds, it returns the actual number of interrupts allocated and
 * indicates the successful configuration of MSI-X capability structure
 * with new allocated MSI-X interrupts.
 **/
int pci_enable_msix_range(struct pci_dev *dev, struct msix_entry *entries,
			       int minvec, int maxvec)
{
	int nvec = maxvec;
	int rc;

	if (maxvec < minvec)
		return -ERANGE;

	do {
		rc = pci_enable_msix(dev, entries, nvec);
		if (rc < 0) {
			return rc;
		} else if (rc > 0) {
			if (rc < minvec)
				return -ENOSPC;
			nvec = rc;
		}
	} while (rc);

	return nvec;
}
EXPORT_SYMBOL(pci_enable_msix_range);
1117

1118 1119 1120 1121 1122
struct pci_dev *msi_desc_to_pci_dev(struct msi_desc *desc)
{
	return to_pci_dev(desc->dev);
}

1123 1124 1125 1126 1127 1128 1129 1130
void *msi_desc_to_pci_sysdata(struct msi_desc *desc)
{
	struct pci_dev *dev = msi_desc_to_pci_dev(desc);

	return dev->bus->sysdata;
}
EXPORT_SYMBOL_GPL(msi_desc_to_pci_sysdata);

1131 1132 1133 1134 1135 1136 1137 1138
#ifdef CONFIG_PCI_MSI_IRQ_DOMAIN
/**
 * pci_msi_domain_write_msg - Helper to write MSI message to PCI config space
 * @irq_data:	Pointer to interrupt data of the MSI interrupt
 * @msg:	Pointer to the message
 */
void pci_msi_domain_write_msg(struct irq_data *irq_data, struct msi_msg *msg)
{
1139
	struct msi_desc *desc = irq_data_get_msi_desc(irq_data);
1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245

	/*
	 * For MSI-X desc->irq is always equal to irq_data->irq. For
	 * MSI only the first interrupt of MULTI MSI passes the test.
	 */
	if (desc->irq == irq_data->irq)
		__pci_write_msi_msg(desc, msg);
}

/**
 * pci_msi_domain_calc_hwirq - Generate a unique ID for an MSI source
 * @dev:	Pointer to the PCI device
 * @desc:	Pointer to the msi descriptor
 *
 * The ID number is only used within the irqdomain.
 */
irq_hw_number_t pci_msi_domain_calc_hwirq(struct pci_dev *dev,
					  struct msi_desc *desc)
{
	return (irq_hw_number_t)desc->msi_attrib.entry_nr |
		PCI_DEVID(dev->bus->number, dev->devfn) << 11 |
		(pci_domain_nr(dev->bus) & 0xFFFFFFFF) << 27;
}

static inline bool pci_msi_desc_is_multi_msi(struct msi_desc *desc)
{
	return !desc->msi_attrib.is_msix && desc->nvec_used > 1;
}

/**
 * pci_msi_domain_check_cap - Verify that @domain supports the capabilities for @dev
 * @domain:	The interrupt domain to check
 * @info:	The domain info for verification
 * @dev:	The device to check
 *
 * Returns:
 *  0 if the functionality is supported
 *  1 if Multi MSI is requested, but the domain does not support it
 *  -ENOTSUPP otherwise
 */
int pci_msi_domain_check_cap(struct irq_domain *domain,
			     struct msi_domain_info *info, struct device *dev)
{
	struct msi_desc *desc = first_pci_msi_entry(to_pci_dev(dev));

	/* Special handling to support pci_enable_msi_range() */
	if (pci_msi_desc_is_multi_msi(desc) &&
	    !(info->flags & MSI_FLAG_MULTI_PCI_MSI))
		return 1;
	else if (desc->msi_attrib.is_msix && !(info->flags & MSI_FLAG_PCI_MSIX))
		return -ENOTSUPP;

	return 0;
}

static int pci_msi_domain_handle_error(struct irq_domain *domain,
				       struct msi_desc *desc, int error)
{
	/* Special handling to support pci_enable_msi_range() */
	if (pci_msi_desc_is_multi_msi(desc) && error == -ENOSPC)
		return 1;

	return error;
}

#ifdef GENERIC_MSI_DOMAIN_OPS
static void pci_msi_domain_set_desc(msi_alloc_info_t *arg,
				    struct msi_desc *desc)
{
	arg->desc = desc;
	arg->hwirq = pci_msi_domain_calc_hwirq(msi_desc_to_pci_dev(desc),
					       desc);
}
#else
#define pci_msi_domain_set_desc		NULL
#endif

static struct msi_domain_ops pci_msi_domain_ops_default = {
	.set_desc	= pci_msi_domain_set_desc,
	.msi_check	= pci_msi_domain_check_cap,
	.handle_error	= pci_msi_domain_handle_error,
};

static void pci_msi_domain_update_dom_ops(struct msi_domain_info *info)
{
	struct msi_domain_ops *ops = info->ops;

	if (ops == NULL) {
		info->ops = &pci_msi_domain_ops_default;
	} else {
		if (ops->set_desc == NULL)
			ops->set_desc = pci_msi_domain_set_desc;
		if (ops->msi_check == NULL)
			ops->msi_check = pci_msi_domain_check_cap;
		if (ops->handle_error == NULL)
			ops->handle_error = pci_msi_domain_handle_error;
	}
}

static void pci_msi_domain_update_chip_ops(struct msi_domain_info *info)
{
	struct irq_chip *chip = info->chip;

	BUG_ON(!chip);
	if (!chip->irq_write_msi_msg)
		chip->irq_write_msi_msg = pci_msi_domain_write_msg;
1246 1247 1248 1249
	if (!chip->irq_mask)
		chip->irq_mask = pci_msi_mask_irq;
	if (!chip->irq_unmask)
		chip->irq_unmask = pci_msi_unmask_irq;
1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266
}

/**
 * pci_msi_create_irq_domain - Creat a MSI interrupt domain
 * @node:	Optional device-tree node of the interrupt controller
 * @info:	MSI domain info
 * @parent:	Parent irq domain
 *
 * Updates the domain and chip ops and creates a MSI interrupt domain.
 *
 * Returns:
 * A domain pointer or NULL in case of failure.
 */
struct irq_domain *pci_msi_create_irq_domain(struct device_node *node,
					     struct msi_domain_info *info,
					     struct irq_domain *parent)
{
1267 1268
	struct irq_domain *domain;

1269 1270 1271 1272 1273
	if (info->flags & MSI_FLAG_USE_DEF_DOM_OPS)
		pci_msi_domain_update_dom_ops(info);
	if (info->flags & MSI_FLAG_USE_DEF_CHIP_OPS)
		pci_msi_domain_update_chip_ops(info);

1274 1275 1276 1277 1278 1279
	domain = msi_create_irq_domain(node, info, parent);
	if (!domain)
		return NULL;

	domain->bus_token = DOMAIN_BUS_PCI_MSI;
	return domain;
1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306
}

/**
 * pci_msi_domain_alloc_irqs - Allocate interrupts for @dev in @domain
 * @domain:	The interrupt domain to allocate from
 * @dev:	The device for which to allocate
 * @nvec:	The number of interrupts to allocate
 * @type:	Unused to allow simpler migration from the arch_XXX interfaces
 *
 * Returns:
 * A virtual interrupt number or an error code in case of failure
 */
int pci_msi_domain_alloc_irqs(struct irq_domain *domain, struct pci_dev *dev,
			      int nvec, int type)
{
	return msi_domain_alloc_irqs(domain, &dev->dev, nvec);
}

/**
 * pci_msi_domain_free_irqs - Free interrupts for @dev in @domain
 * @domain:	The interrupt domain
 * @dev:	The device for which to free interrupts
 */
void pci_msi_domain_free_irqs(struct irq_domain *domain, struct pci_dev *dev)
{
	msi_domain_free_irqs(domain, &dev->dev);
}
1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333

/**
 * pci_msi_create_default_irq_domain - Create a default MSI interrupt domain
 * @node:	Optional device-tree node of the interrupt controller
 * @info:	MSI domain info
 * @parent:	Parent irq domain
 *
 * Returns: A domain pointer or NULL in case of failure. If successful
 * the default PCI/MSI irqdomain pointer is updated.
 */
struct irq_domain *pci_msi_create_default_irq_domain(struct device_node *node,
		struct msi_domain_info *info, struct irq_domain *parent)
{
	struct irq_domain *domain;

	mutex_lock(&pci_msi_domain_lock);
	if (pci_msi_default_domain) {
		pr_err("PCI: default irq domain for PCI MSI has already been created.\n");
		domain = NULL;
	} else {
		domain = pci_msi_create_irq_domain(node, info, parent);
		pci_msi_default_domain = domain;
	}
	mutex_unlock(&pci_msi_domain_lock);

	return domain;
}
1334
#endif /* CONFIG_PCI_MSI_IRQ_DOMAIN */