access.c 8.5 KB
Newer Older
1
#include <linux/delay.h>
L
Linus Torvalds 已提交
2 3
#include <linux/pci.h>
#include <linux/module.h>
A
Al Viro 已提交
4
#include <linux/sched.h>
L
Linus Torvalds 已提交
5
#include <linux/ioport.h>
6
#include <linux/wait.h>
L
Linus Torvalds 已提交
7

8 9
#include "pci.h"

L
Linus Torvalds 已提交
10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67
/*
 * This interrupt-safe spinlock protects all accesses to PCI
 * configuration space.
 */

static DEFINE_SPINLOCK(pci_lock);

/*
 *  Wrappers for all PCI configuration access functions.  They just check
 *  alignment, do locking and call the low-level functions pointed to
 *  by pci_dev->ops.
 */

#define PCI_byte_BAD 0
#define PCI_word_BAD (pos & 1)
#define PCI_dword_BAD (pos & 3)

#define PCI_OP_READ(size,type,len) \
int pci_bus_read_config_##size \
	(struct pci_bus *bus, unsigned int devfn, int pos, type *value)	\
{									\
	int res;							\
	unsigned long flags;						\
	u32 data = 0;							\
	if (PCI_##size##_BAD) return PCIBIOS_BAD_REGISTER_NUMBER;	\
	spin_lock_irqsave(&pci_lock, flags);				\
	res = bus->ops->read(bus, devfn, pos, len, &data);		\
	*value = (type)data;						\
	spin_unlock_irqrestore(&pci_lock, flags);			\
	return res;							\
}

#define PCI_OP_WRITE(size,type,len) \
int pci_bus_write_config_##size \
	(struct pci_bus *bus, unsigned int devfn, int pos, type value)	\
{									\
	int res;							\
	unsigned long flags;						\
	if (PCI_##size##_BAD) return PCIBIOS_BAD_REGISTER_NUMBER;	\
	spin_lock_irqsave(&pci_lock, flags);				\
	res = bus->ops->write(bus, devfn, pos, len, value);		\
	spin_unlock_irqrestore(&pci_lock, flags);			\
	return res;							\
}

PCI_OP_READ(byte, u8, 1)
PCI_OP_READ(word, u16, 2)
PCI_OP_READ(dword, u32, 4)
PCI_OP_WRITE(byte, u8, 1)
PCI_OP_WRITE(word, u16, 2)
PCI_OP_WRITE(dword, u32, 4)

EXPORT_SYMBOL(pci_bus_read_config_byte);
EXPORT_SYMBOL(pci_bus_read_config_word);
EXPORT_SYMBOL(pci_bus_read_config_dword);
EXPORT_SYMBOL(pci_bus_write_config_byte);
EXPORT_SYMBOL(pci_bus_write_config_word);
EXPORT_SYMBOL(pci_bus_write_config_dword);
68

69 70 71 72 73 74 75 76 77
/*
 * The following routines are to prevent the user from accessing PCI config
 * space when it's unsafe to do so.  Some devices require this during BIST and
 * we're required to prevent it during D-state transitions.
 *
 * We have a bit per device to indicate it's blocked and a global wait queue
 * for callers to sleep on until devices are unblocked.
 */
static DECLARE_WAIT_QUEUE_HEAD(pci_ucfg_wait);
78

79 80 81 82 83 84 85 86 87 88 89 90
static noinline void pci_wait_ucfg(struct pci_dev *dev)
{
	DECLARE_WAITQUEUE(wait, current);

	__add_wait_queue(&pci_ucfg_wait, &wait);
	do {
		set_current_state(TASK_UNINTERRUPTIBLE);
		spin_unlock_irq(&pci_lock);
		schedule();
		spin_lock_irq(&pci_lock);
	} while (dev->block_ucfg_access);
	__remove_wait_queue(&pci_ucfg_wait, &wait);
91 92 93 94 95 96 97 98 99
}

#define PCI_USER_READ_CONFIG(size,type)					\
int pci_user_read_config_##size						\
	(struct pci_dev *dev, int pos, type *val)			\
{									\
	int ret = 0;							\
	u32 data = -1;							\
	if (PCI_##size##_BAD) return PCIBIOS_BAD_REGISTER_NUMBER;	\
100 101 102
	spin_lock_irq(&pci_lock);					\
	if (unlikely(dev->block_ucfg_access)) pci_wait_ucfg(dev);	\
	ret = dev->bus->ops->read(dev->bus, dev->devfn,			\
103
					pos, sizeof(type), &data);	\
104
	spin_unlock_irq(&pci_lock);					\
105 106 107 108 109 110 111 112 113 114
	*val = (type)data;						\
	return ret;							\
}

#define PCI_USER_WRITE_CONFIG(size,type)				\
int pci_user_write_config_##size					\
	(struct pci_dev *dev, int pos, type val)			\
{									\
	int ret = -EIO;							\
	if (PCI_##size##_BAD) return PCIBIOS_BAD_REGISTER_NUMBER;	\
115 116 117
	spin_lock_irq(&pci_lock);					\
	if (unlikely(dev->block_ucfg_access)) pci_wait_ucfg(dev);	\
	ret = dev->bus->ops->write(dev->bus, dev->devfn,		\
118
					pos, sizeof(type), val);	\
119
	spin_unlock_irq(&pci_lock);					\
120 121 122 123 124 125 126 127 128 129
	return ret;							\
}

PCI_USER_READ_CONFIG(byte, u8)
PCI_USER_READ_CONFIG(word, u16)
PCI_USER_READ_CONFIG(dword, u32)
PCI_USER_WRITE_CONFIG(byte, u8)
PCI_USER_WRITE_CONFIG(word, u16)
PCI_USER_WRITE_CONFIG(dword, u32)

130 131 132 133 134 135
/* VPD access through PCI 2.2+ VPD capability */

#define PCI_VPD_PCI22_SIZE (PCI_VPD_ADDR_MASK + 1)

struct pci_vpd_pci22 {
	struct pci_vpd base;
136 137
	struct mutex lock;
	u16	flag;
138
	bool	busy;
139
	u8	cap;
140 141
};

142 143 144 145 146 147
/*
 * Wait for last operation to complete.
 * This code has to spin since there is no other notification from the PCI
 * hardware. Since the VPD is often implemented by serial attachment to an
 * EEPROM, it may take many milliseconds to complete.
 */
148 149 150 151
static int pci_vpd_pci22_wait(struct pci_dev *dev)
{
	struct pci_vpd_pci22 *vpd =
		container_of(dev->vpd, struct pci_vpd_pci22, base);
152 153
	unsigned long timeout = jiffies + HZ/20 + 2;
	u16 status;
154 155 156 157 158 159
	int ret;

	if (!vpd->busy)
		return 0;

	for (;;) {
160
		ret = pci_user_read_config_word(dev, vpd->cap + PCI_VPD_ADDR,
161
						&status);
162
		if (ret)
163
			return ret;
164 165

		if ((status & PCI_VPD_ADDR_F) == vpd->flag) {
166 167 168
			vpd->busy = false;
			return 0;
		}
169 170

		if (time_after(jiffies, timeout))
171
			return -ETIMEDOUT;
172 173 174 175
		if (fatal_signal_pending(current))
			return -EINTR;
		if (!cond_resched())
			udelay(10);
176 177 178 179 180 181 182 183 184
	}
}

static int pci_vpd_pci22_read(struct pci_dev *dev, int pos, int size,
			      char *buf)
{
	struct pci_vpd_pci22 *vpd =
		container_of(dev->vpd, struct pci_vpd_pci22, base);
	u32 val;
185
	int ret = 0;
186 187
	int begin, end, i;

188
	if (pos < 0 || pos > vpd->base.len || size > vpd->base.len  - pos)
189 190 191 192
		return -EINVAL;
	if (size == 0)
		return 0;

193 194 195
	if (mutex_lock_killable(&vpd->lock))
		return -EINTR;

196 197 198 199 200 201 202
	ret = pci_vpd_pci22_wait(dev);
	if (ret < 0)
		goto out;
	ret = pci_user_write_config_word(dev, vpd->cap + PCI_VPD_ADDR,
					 pos & ~3);
	if (ret < 0)
		goto out;
203

204
	vpd->busy = true;
205
	vpd->flag = PCI_VPD_ADDR_F;
206 207 208 209 210 211
	ret = pci_vpd_pci22_wait(dev);
	if (ret < 0)
		goto out;
	ret = pci_user_read_config_dword(dev, vpd->cap + PCI_VPD_DATA,
					 &val);
out:
212
	mutex_unlock(&vpd->lock);
213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232
	if (ret < 0)
		return ret;

	/* Convert to bytes */
	begin = pos & 3;
	end = min(4, begin + size);
	for (i = 0; i < end; ++i) {
		if (i >= begin)
			*buf++ = val;
		val >>= 8;
	}
	return end - begin;
}

static int pci_vpd_pci22_write(struct pci_dev *dev, int pos, int size,
			       const char *buf)
{
	struct pci_vpd_pci22 *vpd =
		container_of(dev->vpd, struct pci_vpd_pci22, base);
	u32 val;
233
	int ret = 0;
234

235 236
	if (pos < 0 || pos > vpd->base.len || pos & 3 ||
	    size > vpd->base.len - pos || size < 4)
237 238 239 240 241 242 243
		return -EINVAL;

	val = (u8) *buf++;
	val |= ((u8) *buf++) << 8;
	val |= ((u8) *buf++) << 16;
	val |= ((u32)(u8) *buf++) << 24;

244 245
	if (mutex_lock_killable(&vpd->lock))
		return -EINTR;
246 247 248 249 250 251 252 253 254 255 256 257 258 259 260
	ret = pci_vpd_pci22_wait(dev);
	if (ret < 0)
		goto out;
	ret = pci_user_write_config_dword(dev, vpd->cap + PCI_VPD_DATA,
					  val);
	if (ret < 0)
		goto out;
	ret = pci_user_write_config_word(dev, vpd->cap + PCI_VPD_ADDR,
					 pos | PCI_VPD_ADDR_F);
	if (ret < 0)
		goto out;
	vpd->busy = true;
	vpd->flag = 0;
	ret = pci_vpd_pci22_wait(dev);
out:
261
	mutex_unlock(&vpd->lock);
262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290
	if (ret < 0)
		return ret;

	return 4;
}

static void pci_vpd_pci22_release(struct pci_dev *dev)
{
	kfree(container_of(dev->vpd, struct pci_vpd_pci22, base));
}

static struct pci_vpd_ops pci_vpd_pci22_ops = {
	.read = pci_vpd_pci22_read,
	.write = pci_vpd_pci22_write,
	.release = pci_vpd_pci22_release,
};

int pci_vpd_pci22_init(struct pci_dev *dev)
{
	struct pci_vpd_pci22 *vpd;
	u8 cap;

	cap = pci_find_capability(dev, PCI_CAP_ID_VPD);
	if (!cap)
		return -ENODEV;
	vpd = kzalloc(sizeof(*vpd), GFP_ATOMIC);
	if (!vpd)
		return -ENOMEM;

291
	vpd->base.len = PCI_VPD_PCI22_SIZE;
292
	vpd->base.ops = &pci_vpd_pci22_ops;
293
	mutex_init(&vpd->lock);
294 295 296 297 298 299
	vpd->cap = cap;
	vpd->busy = false;
	dev->vpd = &vpd->base;
	return 0;
}

300 301 302 303
/**
 * pci_block_user_cfg_access - Block userspace PCI config reads/writes
 * @dev:	pci device struct
 *
304 305 306 307
 * When user access is blocked, any reads or writes to config space will
 * sleep until access is unblocked again.  We don't allow nesting of
 * block/unblock calls.
 */
308 309 310
void pci_block_user_cfg_access(struct pci_dev *dev)
{
	unsigned long flags;
311
	int was_blocked;
312 313

	spin_lock_irqsave(&pci_lock, flags);
314
	was_blocked = dev->block_ucfg_access;
315 316
	dev->block_ucfg_access = 1;
	spin_unlock_irqrestore(&pci_lock, flags);
317 318 319 320

	/* If we BUG() inside the pci_lock, we're guaranteed to hose
	 * the machine */
	BUG_ON(was_blocked);
321 322 323 324 325 326 327 328
}
EXPORT_SYMBOL_GPL(pci_block_user_cfg_access);

/**
 * pci_unblock_user_cfg_access - Unblock userspace PCI config reads/writes
 * @dev:	pci device struct
 *
 * This function allows userspace PCI config accesses to resume.
329
 */
330 331 332 333 334
void pci_unblock_user_cfg_access(struct pci_dev *dev)
{
	unsigned long flags;

	spin_lock_irqsave(&pci_lock, flags);
335 336 337 338 339

	/* This indicates a problem in the caller, but we don't need
	 * to kill them, unlike a double-block above. */
	WARN_ON(!dev->block_ucfg_access);

340
	dev->block_ucfg_access = 0;
341
	wake_up_all(&pci_ucfg_wait);
342 343 344
	spin_unlock_irqrestore(&pci_lock, flags);
}
EXPORT_SYMBOL_GPL(pci_unblock_user_cfg_access);